Yes, you can train for longer and the accuracy should keep improving. When the accuracy starts to decrease then it will be a sign of overfitting, until then you can keep training. I believe the validation loss is lower because during the validation phase dropout is not active.
I don’t think the validation loss is lower because during the validation phase dropout is not active.
At 51m53s of the lesson 2 video, ps = 0 means turn off the dropout, but train_loss is still higher than valid_loss.
mario_jorge
(Mario jorge lopes chagas de almeida)
4
I’m having the same problem with the MNIST dataset , using default LR, trying differents arquitetures (resnet18, resnet50), image size (16,32,64), splinting the validation with more data (0.3), but everytime I’m getting the training loss bigger than the validation:
am I doing something wrong?