File tree Expand file tree Collapse file tree 2 files changed +4
-4
lines changed Expand file tree Collapse file tree 2 files changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -51,7 +51,7 @@ def test_convergence(self):
51
51
model = MultitaskClassifier (tasks = [task1 , task2 ])
52
52
53
53
# Train
54
- trainer = Trainer (lr = 0.001 , n_epochs = 10 , progress_bar = False )
54
+ trainer = Trainer (lr = 0.002 , n_epochs = 50 , progress_bar = False )
55
55
trainer .fit (model , dataloaders )
56
56
scores = model .score (dataloaders )
57
57
Original file line number Diff line number Diff line change @@ -19,15 +19,15 @@ def test_sce_equals_ce(self):
19
19
20
20
ce_loss = F .cross_entropy (Y_probs , Y_golds , reduction = "none" )
21
21
ces_loss = cross_entropy_with_probs (Y_probs , Y_golds_probs , reduction = "none" )
22
- np .testing .assert_equal (ce_loss .numpy (), ces_loss .numpy ())
22
+ np .testing .assert_almost_equal (ce_loss .numpy (), ces_loss .numpy ())
23
23
24
24
ce_loss = F .cross_entropy (Y_probs , Y_golds , reduction = "sum" )
25
25
ces_loss = cross_entropy_with_probs (Y_probs , Y_golds_probs , reduction = "sum" )
26
- np .testing .assert_equal (ce_loss .numpy (), ces_loss .numpy ())
26
+ np .testing .assert_almost_equal (ce_loss .numpy (), ces_loss .numpy ())
27
27
28
28
ce_loss = F .cross_entropy (Y_probs , Y_golds , reduction = "mean" )
29
29
ces_loss = cross_entropy_with_probs (Y_probs , Y_golds_probs , reduction = "mean" )
30
- np .testing .assert_equal (ce_loss .numpy (), ces_loss .numpy ())
30
+ np .testing .assert_almost_equal (ce_loss .numpy (), ces_loss .numpy ())
31
31
32
32
def test_perfect_predictions (self ):
33
33
# Does soft ce loss achieve approx. 0 loss with perfect predictions?
You can’t perform that action at this time.
0 commit comments