How would you handle loading a pretrained word embedding that is missing some important words in your dataset?
Should I be loading the word embedding then adding additional rows for each additional word and initializing them with a random set of weights then training them?