I worked through lesson’s 4 statefarm notebook and had an issue with in pre-computing the output of the last convolutional layer, I’m trying to understand if it’s something I did wrong or is there a mistake in the notebook.
In the notebook the batches that are used in conv_model are with shuffle=True (default), and when I used the conv_feat (the output of the conv_model) as an input to the bn_model I got training accuracy and validation accuracy of ~10%.
I tried playing with the learning rate and some other things that didn’t help which make me think that the labels are not corresponding correctly.
I changed batches to shuffle=False, re-run everything and got ~60% training accuracy after 1 epoch.
So, is there a mistake in the notebook and the batches must not be shuffled when pre-computing an output of intermediate level? or did I miss something?