Closed sergsb closed 1 year ago
I had this error before, I updated the repo and gave 7 gpus for it.
And I had some issues with NCCL, so I use export NCCL_P2P_DISABLE=1 # 3090s do not support p2p
Other than that, it's pretty much the same:
srun python3 $FASTCHAT_DIR/fastchat/serve/model_worker.py \
--controller http://controller-server-address:21001 \
--port 31020 --worker http://$(hostname):31020 \
--num-gpus 7 \
--host 0.0.0.0 \
--model-path /$FASTCHAT_DIR/models/FreeWilly2
@surak -- As I found, only updating to the latest version from GitHub works. The pip version can not work with FreeWilly2 for a reason.
Maybe they haven't released it on pip yet.
@surak -- As I found, only updating to the latest version from GitHub works. The pip version can not work with FreeWilly2 for a reason.
@sergsb , what exactly did you update? Accelerate? Transformers? Fastchat? I'm getting the same error, so any further guidance would be much appreciated.
@surak -- As I found, only updating to the latest version from GitHub works. The pip version can not work with FreeWilly2 for a reason.
@sergsb , what exactly did you update? Accelerate? Transformers? Fastchat? I'm getting the same error, so any further guidance would be much appreciated.
I updated fastchat pip install git+https//path to repo
Thanks :)
Let's close this one and let the developers worry about other bugs, then? :-)
Dear all,
I am trying to run
stabilityai/FreeWilly2
(LLaMA 2 70B) model. When I run any request to the model I have:How can one fix it?