meta-llama / llama

Inference code for Llama models
Other
54.12k stars 9.32k forks source link

Torch Error #1132

Open Jufyer opened 1 week ago

Jufyer commented 1 week ago

Hi, I get an Error if I want to run the model. My input in Anaconda is the following:

torchrun --nproc_per_node 1 example_chat_completion.py \ --ckpt_dir LLamaa/ \ --tokenizer_path tokenizer.model \ --max_seq_len 512 --max_batch_size 6

This is the Command from the GitHub-Page here. I think the important part of the Error is following:

 [W socket.cpp:697] [c10d] The client socket has failed to connect to [NB-KUHNLA]:29500 (system error: 10049 - Die angeforderte Adresse ist in diesem Kontext ung³ltig.).
C:\Users\User\miniconda3\Lib\site-packages\torch\distributed\distributed_c10d.py:613: UserWarning: Attempted to get default timeout for nccl backend, but NCCL support is not compiled

But the whole Error can be fund here: https://pastebin.com/emJyPEC2

Runtime Environment

I am thankfull if you could help me