bigscience-workshop / petals

🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
https://petals.dev
MIT License
8.89k stars 490 forks source link

Error when trying to launch private swarm using locally stored model #549

Open sdryluth opened 6 months ago

sdryluth commented 6 months ago

Hi,

I am trying to launch a private swarm using a model I have downloaded to a local directory, which I think in theory should work fine. However, it presents an error when I try to run the command to start the first server.

This is what I'm getting.

python -m petals.cli.run_server lzlv --new_swarm Dec 21 19:41:15.890 [INFO] Running Petals 2.3.0.dev1 Dec 21 19:41:16.231 [INFO] Make sure you follow the LLaMA's terms of use: https://bit.ly/llama2-license for LLaMA 2, >https://bit.ly/llama-license for LLaMA 1 Traceback (most recent call last): File "/opt/conda/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/opt/conda/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/opt/conda/lib/python3.10/site-packages/petals/cli/run_server.py", line 235, in main() File "/opt/conda/lib/python3.10/site-packages/petals/cli/run_server.py", line 219, in main server = Server( File "/opt/conda/lib/python3.10/site-packages/petals/server/server.py", line 119, in init assert UID_DELIMITER not in dht_prefix and CHAIN_DELIMITER not in dht_prefix, ( TypeError: argument of type 'NoneType' is not iterable

My setup using the same command works fine when I launch it using the huggingface path for the same model. The folder I reference on my machine contains the model downloaded right from huggingface including config.json.