Open sonalkum opened 9 months ago
hi there,
Thanks for the question.
No, we didn't have this error. FYI, our machine has 512G CPU RAM (not VRAM). This might be an RAM issue.
A simple Google find this: https://github.com/huggingface/datasets/issues/4782
-Yuan
Consider trimming the data a little bit. I actually believe cutting the stage 4 data into half with a random sampling and double the training epoch could train a model with reasonable performance.
Thank you for your quick response. Even in our setting we have 512G of CPU RAM, so I was curious if you had faced any similar issue.
This is bit weird, that reminds me that in LTU-AS, we have 10.6M training data, but still do not have any issue.
Have you changed any code?
Hello,
Thank you so much for sharing the code. Great work on the repo!!
I am trying to run the code for LTU openaqa, I've completed the first 3 stage of training, but I am stuck on the 4th stage. So, I was wondering if you faced any issue similar to the following:
If yes, how did you resolve it? I think this is an issue with pyarrow which is used in HuggingFace's datasets.load_dataset() function.
Thanks in advance.