-
### Validations
- [X] I believe this is a way to improve. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [X] I'm not able to find an [open issue](https://githu…
-
### Model Series
Qwen2.5
### What are the models used?
Qwen2.5-32B-Instruct
### What is the scenario where the problem happened?
Xinference
### Is this a known issue?
- [X] I have followed [the…
-
### Is there an existing issue for the same bug?
- [X] I have checked the existing issues.
### Branch name
v0.12.0
### Commit ID
na
### Other environment information
```Markdown
Hardware parame…
-
can release a xinference demo since not all model can be found by Ollama?
or fastchat, text-generation-webui or....
i can not found the model i used in ollama library...
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a…
-
### System Info / 系統信息
Driver Version: 535.171.04 CUDA Version: 12.2
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [X] docker / docker
- [ ] pip install / 通过 pip insta…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
(base) tskj@tskj:~$ sudo docker pull dhub.kubesre.xyz/xprobe/xinference:v0.16.1
Error response from daemon: toomanyrequests: You have reached your pull rate limit. You may increase the limit by authe…
-
### System Info / 系統信息
cuda:12.04
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [ ] docker / docker
- [ ] pip install / 通过 pip install 安装
- [X] installation from source / 从源码安装…
-
### System Info / 系統信息
CUDA12.4
win10系统
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [X] docker / docker
- [ ] pip install / 通过 pip install 安装
- [ ] installation from sourc…