xorbitsai / inference

Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
https://inference.readthedocs.io
Apache License 2.0
5.32k stars 430 forks source link

RuntimeError: Failed to launch model, detail: No such file or directory #2106

Closed Yue-Rain closed 2 months ago

Yue-Rain commented 2 months ago

System Info / 系統信息

Package Version


accelerate 0.33.0 aiobotocore 2.7.0 aiofiles 23.2.1 aiohappyeyeballs 2.3.2 aiohttp 3.10.0 aioitertools 0.11.0 aioprometheus 23.12.0 aiosignal 1.3.1 altair 5.3.0 annotated-types 0.7.0 anyio 4.4.0 async-timeout 4.0.3 attrdict 2.0.1 attrs 23.2.0 auto_gptq 0.7.1 autoawq 0.2.6 autoawq_kernels 0.0.7 bcrypt 4.2.0 bitsandbytes 0.42.0 botocore 1.31.64 Brotli 1.0.9 certifi 2024.7.4 cffi 1.16.0 charset-normalizer 3.3.2 click 8.1.7 cloudpickle 3.0.0 cmake 3.30.1 colorama 0.4.6 coloredlogs 15.0.1 contourpy 1.2.1 cryptography 43.0.0 cycler 0.12.1 datasets 2.20.0 dill 0.3.8 diskcache 5.6.3 distro 1.9.0 dnspython 2.6.1 ecdsa 0.19.0 einops 0.8.0 email_validator 2.2.0 exceptiongroup 1.2.2 fastapi 0.110.3 fastapi-cli 0.0.4 ffmpy 0.4.0 filelock 3.13.1 fonttools 4.53.1 frozenlist 1.4.1 fsspec 2023.10.0 gekko 1.2.1 gmpy2 2.1.2 gradio 4.26.0 gradio_client 0.15.1 h11 0.14.0 httpcore 1.0.5 httptools 0.6.1 httpx 0.27.0 huggingface-hub 0.24.1 humanfriendly 10.0 idna 3.7 importlib_resources 6.4.0 interegular 0.3.3 Jinja2 3.1.4 jmespath 1.0.1 joblib 1.4.2 jsonschema 4.23.0 jsonschema-specifications 2023.12.1 kiwisolver 1.4.5 lark 1.1.9 llvmlite 0.43.0 lm-format-enforcer 0.9.8 markdown-it-py 3.0.0 MarkupSafe 2.1.3 matplotlib 3.9.1 mdurl 0.1.2 mkl-fft 1.3.8 mkl-random 1.2.4 mkl-service 2.4.0 modelscope 1.16.1 mpmath 1.3.0 msgpack 1.0.8 multidict 6.0.5 multiprocess 0.70.16 nest-asyncio 1.6.0 networkx 3.3 ninja 1.11.1.1 numba 0.60.0 numpy 1.26.4 nvidia-cublas-cu12 12.1.3.1 nvidia-cuda-cupti-cu12 12.1.105 nvidia-cuda-nvrtc-cu12 12.1.105 nvidia-cuda-runtime-cu12 12.1.105 nvidia-cudnn-cu12 8.9.2.26 nvidia-cufft-cu12 11.0.2.54 nvidia-curand-cu12 10.3.2.106 nvidia-cusolver-cu12 11.4.5.107 nvidia-cusparse-cu12 12.1.0.106 nvidia-ml-py 12.555.43 nvidia-nccl-cu11 2.19.3 nvidia-nccl-cu12 2.19.3 nvidia-nvjitlink-cu12 12.5.82 nvidia-nvtx-cu12 12.1.105 openai 1.37.1 opencv-contrib-python 4.10.0.84 optimum 1.21.2 orjson 3.10.6 outlines 0.0.34 packaging 24.1 pandas 2.2.2 passlib 1.7.4 peft 0.12.0 pillow 10.4.0 pip 24.0 prometheus_client 0.20.0 protobuf 5.27.2 psutil 6.0.0 py-cpuinfo 9.0.0 pyarrow 17.0.0 pyarrow-hotfix 0.6 pyasn1 0.6.0 pycparser 2.22 pydantic 2.8.2 pydantic_core 2.20.1 pydub 0.25.1 Pygments 2.18.0 pynvml 11.5.0 pyparsing 3.1.2 PySocks 1.7.1 python-dateutil 2.9.0.post0 python-dotenv 1.0.1 python-jose 3.3.0 python-multipart 0.0.9 pytz 2024.1 PyYAML 6.0.1 quantile-python 1.1 ray 2.32.0 referencing 0.35.1 regex 2024.5.15 requests 2.32.3 rich 13.7.1 rouge 1.0.1 rpds-py 0.19.0 rsa 4.9 ruff 0.5.5 s3fs 2023.10.0 safetensors 0.4.3 scikit-learn 1.5.1 scipy 1.14.0 semantic-version 2.10.0 sentence-transformers 3.0.1 sentencepiece 0.2.0 setuptools 69.5.1 shellingham 1.5.4 six 1.16.0 sniffio 1.3.1 sse-starlette 2.1.2 starlette 0.37.2 sympy 1.12 tabulate 0.9.0 tblib 3.0.0 threadpoolctl 3.5.0 tiktoken 0.6.0 timm 1.0.8 tokenizers 0.19.1 tomlkit 0.12.0 toolz 0.12.1 torch 2.2.1 torchaudio 2.2.0 torchvision 0.17.0 tqdm 4.66.4 transformers 4.41.0 transformers-stream-generator 0.0.5 triton 2.2.0 typer 0.11.1 typing_extensions 4.11.0 tzdata 2024.1 urllib3 2.0.7 uvicorn 0.30.3 uvloop 0.19.0 vllm 0.4.1 vllm_nccl_cu12 2.18.1.0.4.0 watchfiles 0.22.0 websockets 11.0.3 wheel 0.43.0 wrapt 1.16.0 xformers 0.0.25 xinference 0.14.1 xoscar 0.3.2 xxhash 3.4.1 yarl 1.9.4 zstandard 0.23.0

Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?

Version info / 版本信息

xinference 0.14.1

The command used to start Xinference / 用以启动 xinference 的命令

xinference-local --host 0.0.0.0 --port 9997

the result of command "xinference list" is expected:

UID Type Name Dimensions


bge-large-zh-v1.5 embedding bge-large-zh-v1.5 1024

Reproduction / 复现过程

xinference launch --model-name bge-reranker-v2-m3 --model-type rerank

Expected behavior / 期待表现

bge-reranker-v2-m3 work

Yue-Rain commented 2 months ago

操作系统是 centos7;可以注册模型 但是xinference launch 失败 RuntimeError: Failed to launch model, detail: [address=0.0.0.0:46393, pid=98690] [Errno 2] No such file or directory

github-actions[bot] commented 2 months ago

This issue is stale because it has been open for 7 days with no activity.

github-actions[bot] commented 2 months ago

This issue was closed because it has been inactive for 5 days since being marked as stale.

lijinrui6 commented 1 week ago

请问您解决了吗,我遇到一样的问题了