LucienShui / huggingface-vscode-endpoint-server

starcoder server for huggingface-vscdoe custom endpoint
Apache License 2.0
166 stars 56 forks source link

ValueError: Could not load model bigcode/starcoder with any of the following classes #2

Open 24mlight opened 1 year ago

24mlight commented 1 year ago

pip freeze:

accelerate==0.19.0 anyio==3.6.2 boto3==1.26.135 botocore==1.29.135 certifi==2023.5.7 charset-normalizer==3.1.0 click==8.1.3 fastapi==0.95.2 filelock==3.12.0 fsspec==2023.5.0 h11==0.14.0 huggingface-hub==0.14.1 idna==3.4 jmespath==1.0.1 numpy==1.24.3 nvidia-cublas-cu11==11.10.3.66 nvidia-cuda-nvrtc-cu11==11.7.99 nvidia-cuda-runtime-cu11==11.7.99 nvidia-cudnn-cu11==8.5.0.96 packaging==23.1 psutil==5.9.5 pydantic==1.10.7 python-dateutil==2.8.2 pytorch-pretrained-bert==0.6.2 PyYAML==6.0 regex==2023.5.5 requests==2.30.0 s3transfer==0.6.1 six==1.16.0 sniffio==1.3.0 starlette==0.27.0 tokenizers==0.13.3 torch==1.13.1 tqdm==4.65.0 transformers==4.28.1 typing_extensions==4.5.0 urllib3==1.26.15 uvicorn==0.22.0 `

python main.py --host=0.0.0.0 --port=7088 --pretrained="bigcode/starcoder" Traceback (most recent call last): File "/data/sd-webui/7099/main.py", line 38, in main() File "/data/sd-webui/7099/main.py", line 33, in main generator = StarCoder(args.pretrained, device_map='auto') File "/data/sd-webui/7099/generators.py", line 17, in init self.pipe: Pipeline = pipeline( File "/data/sd-webui/7099/venv/lib/python3.10/site-packages/transformers/pipelines/init.py", line 779, in pipeline framework, model = infer_framework_load_model( File "/data/sd-webui/7099/venv/lib/python3.10/site-packages/transformers/pipelines/base.py", line 271, in infer_framework_load_model raise ValueError(f"Could not load model {model} with any of the following classes: {class_tuple}.") ValueError: Could not load model bigcode/starcoder with any of the following classes: (<class 'transformers.models.auto.modeling_auto.AutoModelForCausalLM'>, <class 'transformers.models.gpt_bigcode.modeling_gpt_bigcode.GPTBigCodeForCausalLM'>).

LucienShui commented 1 year ago

Have you checked your model's hash? It looks like something goes wrong when loading config or weights. Perhaps problem from system maybe.

24mlight commented 1 year ago

Have you checked your model's hash? It looks like something goes wrong when loading config or weights. Perhaps problem from system maybe.

image yes, how to check whether the hash is right or not?

LucienShui commented 1 year ago

What does your git status command display? If there is any diff, that means there is something wrong maybe.

Chapoly1305 commented 1 year ago

Are you using an AMD GPU?

If yes, continue reading, if not, just ignore my comment.

I have a 6800XT and managed to make it work, but eventually it ran out of VRAM after loading the model files. I suggest just give up now before wasting more time, unless you have a AMD GPU which have at least 32GB of VRAM.

Screenshot from 2023-06-04 18-14-15

And: Traceback of TorchScript (most recent call last): RuntimeError: HIP out of memory. Tried to allocate 224.00 MiB (GPU 0; 15.98 GiB total capacity; 15.36 GiB already allocated; 156.00 MiB free; 15.80 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_HIP_ALLOC_CONF

You need to use ROCM supported PyTorch. Use pip list |grep rocm to determine if you have correct pytorch installed.

Expected: Screenshot from 2023-06-04 18-01-03

If you have a similar result, run this simple test command in python to determine your torch worked with rocm.

import torch
torch.cuda.is_available()

The result must be True. If not your amd driver install steps is incomplete. Screenshot from 2023-06-04 18-02-43

fr3deric commented 1 year ago

I have the same problem (on windows) and was trying to get starcode to run using different ways, i.e. downloading the model seperatly and also let the endpoint-server download the model. It wasn't succesful. So I tried a small model (to save some time) python main.py --pretrained="bigcode/tiny_starcoder_py" and that seems to work fine.

So, I'm am wildly guessing, that there is something wrong with the bigcode/starcoder in conjunction with the endpoint-server so that it wouldn't find the model. Also, maybe it's only a windows problem?

masterchop commented 9 months ago

I have the same problem (on windows) and was trying to get starcode to run using different ways, i.e. downloading the model seperatly and also let the endpoint-server download the model. It wasn't succesful. So I tried a small model (to save some time) python main.py --pretrained="bigcode/tiny_starcoder_py" and that seems to work fine.

So, I'm am wildly guessing, that there is something wrong with the bigcode/starcoder in conjunction with the endpoint-server so that it wouldn't find the model. Also, maybe it's only a windows problem?

Thanks, this worked for me, do you have the information for other ones? how to add your own? like Llama2?

iamjameswalters commented 6 months ago

Also, maybe it's only a windows problem?

Not a Windows only problem, I just ran into it on Ubuntu 22.04