Fastai v2 TPU support

You’ll probably have to ask Jeremy about that. Personally, I don’t think this is something that’s strictly necessary, but it’s likely a decision that Jeremy will need to make about whether or not to include TPUs as a default. But I guess I didn’t really have a problem with that and instead was talking about the separate optimizer classes you created, which is not necessary.

Yep, my goal the is same! :slight_smile:

I see you are using my kernel (developed with the help of the PyTorch XLA team and @abhi1thakur) :wink: .
Which version of the kernel is the working one? The latest one is just a quick save, and while there is an older working one, I am not sure if that’s your final fastai version or if there’s more to it?

Exactly my thoughts! I have tried single-core TPU training with very little benefit. Hence, I have been focusing on multi-core TPU training.

Anyway, I will work on it more today and keep you guys updated in this thread!