Open TomLBZ opened 1 year ago
Having the same problem in linux attempting to connect to a private swarm.
File "/home/rcap3/anaconda3/lib/python3.11/site-packages/hivemind/dht/node.py", line 192, in create Sep 17 17:55:02 i7ubuntu python[322820]: p2p = await P2P.create(**kwargs) Sep 17 17:55:02 i7ubuntu python[322820]: ^^^^^^^^^^^^^^^^^^^^^^^^^^ Sep 17 17:55:02 i7ubuntu python[322820]: File "/home/rcap3/anaconda3/lib/python3.11/site-packages/hivemind/p2p/p2p_daemon.py", line 234, in create Sep 17 17:55:02 i7ubuntu python[322820]: await asyncio.wait_for(ready, startup_timeout) Sep 17 17:55:02 i7ubuntu python[322820]: File "/home/rcap3/anaconda3/lib/python3.11/asyncio/tasks.py", line 479, in wait_for Sep 17 17:55:02 i7ubuntu python[322820]: return fut.result() Sep 17 17:55:02 i7ubuntu python[322820]: ^^^^^^^^^^^^ Sep 17 17:55:02 i7ubuntu python[322820]: hivemind.p2p.p2p_daemon_bindings.utils.P2PDaemonError: Daemon failed to start: 2023/09/17 17:55:02 failed to connect to bootstrap peers
023-09-19 16:43:06.190015: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
Sep 19 16:43:07.354 [INFO] Running Petals 2.2.0
/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py:1006: FutureWarning: The use_auth_token
argument is deprecated and will be removed in v5 of Transformers.
warnings.warn(
Downloading (…)lve/main/config.json: 100% 610/610 [00:00<00:00, 2.74MB/s]
Sep 19 16:43:07.883 [INFO] Make sure you follow the LLaMA's terms of use: https://bit.ly/llama2-license for LLaMA 2, https://bit.ly/llama-license for LLaMA 1
Sep 19 16:43:07.884 [INFO] Using DHT prefix: Llama-2-13b-hf
/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py:485: FutureWarning: The use_auth_token
argument is deprecated and will be removed in v5 of Transformers.
warnings.warn(
Traceback (most recent call last):
File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/lib/python3.10/runpy.py", line 86, in _run_code
exec(code, run_globals)
File "/usr/local/lib/python3.10/dist-packages/petals/cli/run_server.py", line 235, in
Hi @TomLBZ @redcap3000 @edugamerplay1228,
This may be an issue with DNS/IPv6 addresses present among the default bootstrap peers. Can you please try again with this option (this uses IPv4 addresses only)?
--initial_peers /ip4/159.89.214.152/tcp/31337/p2p/QmedTaZXmULqwspJXz44SsPZyTNKxhnnFvYRajfH7MGhCY /ip4/159.203.156.48/tcp/31338/p2p/QmQGTqmM7NKjV6ggU1ZCap8zWiyKR89RViDXiqehSiCpY5
Hello @borzunov ,
I encountered a similar issue on Windows with WSL2 while attempting to connect to my own private swarm backbone. I have two hosts connected within the local area network and the error log is totally the same as this. Following your advice, I used this argument:
--initial_peers /ip4/159.89.214.152/tcp/31337/p2p/QmedTaZXmULqwspJXz44SsPZyTNKxhnnFvYRajfH7MGhCY /ip4/159.203.156.48/tcp/31338/p2p/QmQGTqmM7NKjV6ggU1ZCap8zWiyKR89RViDXiqehSiCpY5
This successfully connected the private swarm. However, I have encountered severe network throughput degradation with the private swarm backbone you provided. So I am curious about how to directly solve this problem instead of using public initial_peers.
I tried to host the docker container on truenas scale but failed with the error above. Might be a bug?