huggingface / transfer-learning-conv-ai

🦄 State-of-the-Art Conversational AI with Transfer Learning
MIT License
1.74k stars 430 forks source link

Out of memory XLM-roberta-large #91

Open hadifar opened 3 years ago

hadifar commented 3 years ago

Thanks for releasing your code. I replaced GPT2 with XLM-roberta-large, tensor dataset with LazyLoader, and did padding inside the data loader. Unfortunately, even with batch_size of 1, I got Cuda out of memory. Do you have any suggestions on how can I decrease memory usage? I have no problem with xlm-roberta-base.