Hey @Patrick. Thanks for your suggestions.
I would think this should not matter over a couple of epochs.
If you look at the numbers above you see that the difference from epoch to epoch within the training loss series or within the validation loss series is tiny. This is especially true, if you compare it to the difference between validation loss and training loss for a given epoch. Agreed?
Anyway, I am feeling my way into this, so I would like to do the test you suggested to get confirmation. What did you have in mind? How would you implement this test?
Let the learner work for 20 epochs and then compare the train_loss of the last epoch with learn.validate(learn.data.train_dl)
?
Should I run something like this, but with 20 epochs, not 1?
What would we be looking for in the results? train
should then be lower than val
, correct?