@@ -1170,7 +1170,7 @@ public void TrainingRprop()
11701170
11711171 var loss = TrainLoop ( seq , x , y , optimizer ) ;
11721172
1173- LossIsClose ( 229.68f , loss ) ;
1173+ LossIsClose ( 77.279f , loss ) ;
11741174 }
11751175
11761176
@@ -1187,7 +1187,7 @@ public void TrainingRpropMax()
11871187
11881188 var loss = TrainLoop ( seq , x , y , optimizer , maximize : true ) ;
11891189
1190- LossIsClose ( 229.68f , - loss ) ;
1190+ LossIsClose ( 77.279f , - loss ) ;
11911191 }
11921192
11931193 [ Fact ]
@@ -1203,7 +1203,7 @@ public void TrainingRpropEtam()
12031203
12041204 var loss = TrainLoop ( seq , x , y , optimizer ) ;
12051205
1206- LossIsClose ( 201.417f , loss ) ;
1206+ LossIsClose ( 171.12f , loss ) ;
12071207 }
12081208
12091209 [ Fact ]
@@ -1219,7 +1219,7 @@ public void TrainingRpropEtap()
12191219
12201220 var loss = TrainLoop ( seq , x , y , optimizer ) ;
12211221
1222- LossIsClose ( 221.365f , loss ) ;
1222+ LossIsClose ( 65.859f , loss ) ;
12231223 }
12241224
12251225
@@ -1240,7 +1240,7 @@ public void TrainingRpropParamGroups()
12401240
12411241 var loss = TrainLoop ( seq , x , y , optimizer ) ;
12421242
1243- LossIsClose ( 78.619f , loss ) ;
1243+ LossIsClose ( 66.479f , loss ) ;
12441244 }
12451245
12461246 /// <summary>
0 commit comments