diff --git a/src/main/java/com/naaturel/ANN/implementation/training/AdalineTraining.java b/src/main/java/com/naaturel/ANN/implementation/training/AdalineTraining.java index fbd8219..f1a2166 100644 --- a/src/main/java/com/naaturel/ANN/implementation/training/AdalineTraining.java +++ b/src/main/java/com/naaturel/ANN/implementation/training/AdalineTraining.java @@ -1,19 +1,11 @@ package com.naaturel.ANN.implementation.training; import com.naaturel.ANN.domain.abstraction.Model; -import com.naaturel.ANN.domain.abstraction.Neuron; import com.naaturel.ANN.domain.abstraction.Trainer; import com.naaturel.ANN.domain.abstraction.TrainingStep; import com.naaturel.ANN.domain.model.dataset.DataSet; -import com.naaturel.ANN.domain.model.dataset.DataSetEntry; -import com.naaturel.ANN.domain.model.neuron.Input; -import com.naaturel.ANN.domain.model.neuron.Synapse; -import com.naaturel.ANN.domain.model.neuron.Weight; import com.naaturel.ANN.domain.model.training.TrainingPipeline; import com.naaturel.ANN.implementation.adaline.AdalineTrainingContext; -import com.naaturel.ANN.implementation.gradientDescent.GradientDescentCorrectionStrategy; -import com.naaturel.ANN.implementation.gradientDescent.GradientDescentErrorStrategy; -import com.naaturel.ANN.implementation.gradientDescent.GradientDescentTrainingContext; import com.naaturel.ANN.implementation.gradientDescent.SquareLossStrategy; import com.naaturel.ANN.implementation.simplePerceptron.SimpleCorrectionStrategy; import com.naaturel.ANN.implementation.simplePerceptron.SimpleDeltaStrategy; @@ -21,7 +13,6 @@ import com.naaturel.ANN.implementation.simplePerceptron.SimpleErrorRegistrationS import com.naaturel.ANN.implementation.simplePerceptron.SimplePredictionStrategy; import com.naaturel.ANN.implementation.training.steps.*; -import java.util.ArrayList; import java.util.List; @@ -48,12 +39,8 @@ public class AdalineTraining implements Trainer { new TrainingPipeline(steps) .stopCondition(ctx -> ctx.globalLoss <= 0.125F || ctx.epoch > 10000) - .beforeEpoch(ctx -> { - ctx.globalLoss = 0.0F; - }) - .afterEpoch(ctx -> { - ctx.globalLoss /= context.dataset.size(); - }) + .beforeEpoch(ctx -> ctx.globalLoss = 0.0F) + .afterEpoch(ctx -> ctx.globalLoss /= context.dataset.size()) .withVerbose(true) .withTimeMeasurement(true) .run(context);