which runs without an issue. As soon as I swap out the
SentencePieceTokenizer instead of spacy,
learn.predict results in an error:
~/fastai/fastai2/fastai2/text/core.py in __call__(self, items) 371 372 def __call__(self, items): --> 373 for t in items: yield self.tok.EncodeAsPieces(t) 374 375 # Cell AttributeError: 'NoneType' object has no attribute 'EncodeAsPieces'
Here is the gist with just one difference.
I do see that self.tok is indeed None. Has anybody come across this before?
The commit of fastcore and fastai2 that I am running: