You’ll probably have to ask Jeremy about that. Personally, I don’t think this is something that’s strictly necessary, but it’s likely a decision that Jeremy will need to make about whether or not to include TPUs as a default. But I guess I didn’t really have a problem with that and instead was talking about the separate optimizer classes you created, which is not necessary.
Yep, my goal the is same!
I see you are using my kernel (developed with the help of the PyTorch XLA team and @abhi1thakur) .
Which version of the kernel is the working one? The latest one is just a quick save, and while there is an older working one, I am not sure if that’s your final fastai version or if there’s more to it?
Exactly my thoughts! I have tried single-core TPU training with very little benefit. Hence, I have been focusing on multi-core TPU training.
Anyway, I will work on it more today and keep you guys updated in this thread!