Closed n-akram closed 4 years ago
What I have tried so far:
GPU and TPU always result in memory overflow error. Running on local machine results in ETA of 300+ hours.
Potential solutions:
Changed the model to xlm-base, and it is working in both google Colab and local GPU. Local GPU: 3 hours, Colab 9 hours ETA.
However, colab seems to use around 8.14 GB. Whereas local machine is using complete assigned GPU.
Another implementation seems to use parallelization.
Next task is to try that today on TPU with xlm roberta base.
The current implementation runs out of memory (with one GPU system, of memory 12 GB). The problem exists for both Local GPU and google colab.