That switch you see there, is after the first eval, so the first step is done.
Also I have changed the bs and see no change in mem usage, should I see a change in here?.
Obviously I get an RuntimeError: CUDA out of memory. Tried to allocate 102.00 MiB (GPU 0; 7.79 GiB total capacity; 6.44 GiB already allocated; 86.12 MiB free; 6.61 GiB reserved in total by PyTorch)
I think I changed the bs there, it doesnt seem to change something on the CUDA memory usage… but maybe Im passing it in the wrong place… or should try with bs=8 hehe…