I’m creating language_model_learner from the dataset which contains 18 millions short sentences (in csv it’s about 3GB).
When I start fit_one_cycle, after some time I have out of memory error on my GPU.
Is there possibility to train this model in parts which chunks? How can i do that?
@muellerzr
I’ve got 16 GB memory on my GPU available.
@soorajviraat
I’m making my databunch with this command:
bs = 64
df = TextList.from_df(df, cols=‘xxx’, processor=SPProcessor.load(path_with_sp)).split_by_rand_pct(0.01).label_for_lm().databunch(bs=bs)
Parameters such as
bptt is set inside TextLMDataBunch
max_vocab is set inside TextDataBunch
So you may or may not be able to set it just using base class data bunch