I saw some recommendations to use pack_padded_sequence when padding LSTM inputs to make sure the padding won’t affect the LSTM output.
Is it necessary? Is anyone using it?
Update: I built a text classifier similar to the IMDB classifier. The predictions were different when padding was added to the text. It makes the results inconsistent, e.g. predicting a batch of texts together will give different results than prediction each one separately.