Open edwardsp opened 3 months ago
I have exactly the same problem. 1 node works, but 2 node fails. I think this is a problem on huggingface side.
We run into issues like that too. We don't have a robust solution yet, but one trick we do is caching the datasets locally (or once per node or however many file systems you have) as follows and then making HF not call the hub by setting the environment variable HF_DATASETS_OFFLINE=1
.
from olmo.eval.downstream import *
tokenizer = Tokenizer.from_file("tokenizers/allenai_gpt-neox-olmo-dolma-v1_5.json")
for x in label_to_task_map.values():
kwargs = {}
if isinstance(x, tuple):
x, kwargs = x
x(tokenizer=tokenizer, **kwargs)
I recently merged https://github.com/allenai/OLMo/pull/623, which improves the HF loading situation. Some of the optimizations can be used just by merging in the fix. For the rest, you'll need to set hf_datasets_cache_dir
in config (or pass it as an argument --hf_datasets_cache_dir=<dir>
to scripts/train.py
) to a folder that can be used for caching.
❓ The question
I am trying to run training and I get this error when staring up:
I am running on 2 nodes each with 8 GPUs, using the main branch and pytorch 2.2.2+cu121.
This works with just 1 node using 8 GPUs.