Need advice with training on a pretty big image dataset

Hello people.
I’m presently training the food-101 dataset that classifies 101 dishes. However the dataset is massive, and every epoch takes 30 mins on Colab

Is there a way (in fastai/ logically speaking) to sample a subset of the dataset or should I do it manually?

1 Like

Yes, you can sample a random subset use_partial_data or create your own filter using the data block api:


Thank you! That was very helpful!