Inference using load_learner

If I may also ask an additional question: according to this post: Load_learner on CPU throws "RuntimeError('Attempting to deserialize object on a CUDA)" there is no more a cpu argument (like the one described in the docs: https://docs.fast.ai/basic_train.html#load_learner ), so what is the appropriate way to set this flag?