This paper seems to have flown a bit under the radar in August. It was put out by two authors from Google Cloud and it proposes a network design based on attention for tabular data which appears to outperform tree based models AND provide interpretability.
Seems too good to be true, but worth examining.
There are tensorflow and pytorch implementations of this network out there. Given how much tabular data is out there, Iβm surprised this hasnβt gotten more attention.
Shouldnβt be too hard to try playing with in fastai. If I have time Iβll try it out later this week. That pytorch implementation is here: https://github.com/dreamquark-ai/tabnet
Yes I did, and I taught it in my course, see here:
If youβre attention focused and canβt be convinced by FI (though I argue itβs a pretty good idea if it makes sense to you), then use TabNet. It may not be as accurate as the straight fastai model and probably will take much longer to train though (in terms of epochs, but not necessarily seconds), keep that in mind.