Closed thiner closed 7 months ago
After tons of googling, this may be the root cause: https://github.com/QwenLM/Qwen-VL/issues/106#issuecomment-1751955889
can we expose the trust-remote-code
parameter to command line in backend/python/autogptq/autogptq.py
?
LocalAI version: Docker image:
localai/localai:v2.9.0-cublas-cuda12-core
with extra backendautogptq
Environment, CPU architecture, OS, and Version:
Describe the bug Trying to start Qwen-VL-Chat-int4 model, but failed due to autogptq can't find the
config.json
in the model folder.To Reproduce
RUN apt-get update -y && apt-get install -y curl gcc libxml2 libxml2-dev RUN apt install -y wget git && \ apt clean && \ rm -rf /var/lib/apt/lists/ /tmp/ /var/tmp/*
ENV PATH="/root/miniconda3/bin:${PATH}" ARG PATH="/root/miniconda3/bin:${PATH}"
RUN wget \ https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh \ && mkdir .conda \ && bash Miniconda3-latest-Linux-x86_64.sh -b \ && rm -f Miniconda3-latest-Linux-x86_64.sh RUN conda init bash
RUN PATH=$PATH:/opt/conda/bin make -C backend/python/autogptq ENV EXTERNAL_GRPC_BACKENDS="autogptq:/build/backend/python/autogptq/run.sh" ENV BUILD_TYPE="cublas"
Run the model:
docker run -p 8080:8080 -v $PWD/models:/opt/models -e MODELS_PATH=/opt/models localai:v2.9.0-autogptq --config-file /opt/models/qwen-vl.yaml
Call the API
Expected behavior Respond with answers.
Logs