Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
通过docker启动xinference ,开启大模型后,在chat2db中调用xinference model api ,显示报错
Fast Chat AI error:{"detail":"Method Not Allowed"}
在本地打开postman,post发送http://localhost:9997/,显示{
"detail": "Method Not Allowed"
}
如何正确调用api地址?
Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
[X] docker / docker
[ ] pip install / 通过 pip install 安装
[ ] installation from source / 从源码安装
Version info / 版本信息
.
The command used to start Xinference / 用以启动 xinference 的命令
System Info / 系統信息
通过docker启动xinference ,开启大模型后,在chat2db中调用xinference model api ,显示报错 Fast Chat AI error:{"detail":"Method Not Allowed"} 在本地打开postman,post发送http://localhost:9997/,显示{ "detail": "Method Not Allowed" } 如何正确调用api地址?
Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
Version info / 版本信息
.
The command used to start Xinference / 用以启动 xinference 的命令
.
Reproduction / 复现过程
.
Expected behavior / 期待表现
.