I am traning the model and I have tried few different learning rates but my validation loss is not decrasing. below is the learning rate finder plot:
And I have tried the learning rate of 2e-01 and 1e-01 but stil my validation loss is fluctation after few cycles (not like as Jeremy mentioned in 3 lecture it increases and then decreases). Below is my validation loss history:
According to what you’ve provide, I think that yout model is overfitted. You can notice this by seing the extrememly low training losses and the high validation losses. That is one thing …
The other, is when you see that behavior in validation losses, one can say that gradient descent is not converging (up’s and down’s … as yours) due to a large learning rate …