Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
System Info / 系統信息
Python 3.10.6 centos 7.9
Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
Version info / 版本信息
0.13.3
The command used to start Xinference / 用以启动 xinference 的命令
curl -X 'POST' 'http://192.168.1.88:9997/v1/audio/speech' -H 'accept: application/json' -H 'Content-Type: application/json' -d '{ "model": "ChatTTS", "text": "我是中国人,我深深的热爱着我的祖国和人民", "voice": "echo", "stream": False, "response_format": "mp3" }' -v
Reproduction / 复现过程
curl -X 'POST' 'http://192.168.1.88:9997/v1/audio/speech' -H 'accept: application/json' -H 'Content-Type: application/json' -d '{ "model": "ChatTTS", "text": "我是中国人,我深深的热爱着我的祖国和人民", "voice": "echo", "stream": False, "response_format": "mp3" }' -v
Expected behavior / 期待表现
正常请求到结果