Can anyone share if pytorch fastai does one hot encoding underneath the hood when creating an embedding of the categorical variable.
I noticed in the code for Rossman, that we tell the embedding layer how many values there are for a given categorical, but we pass in the original column, not a DF w/ the columns one hot encoded.
Unless I’m misunderstanding how categorical embeddings are created we dot a one hot vector for a given categorical value and that then gives us the weights for that one hot vector associated, which then pipes into the next fully connected later where it’s combined with the continuous variables, etc etc etc.