[BUG]: I need help,When requesting/api/v1/openai/chat/completions for anything-llm 1.6.0 desktop version, the max_token does not take effect, and the connected model is a local model(xinference) #2023
When requesting/api/v1/openai/chat/completions for anything-llm 1.6.0 desktop version, the max_token does not take effect, and the connected model is a local model(xinference)
How are you running AnythingLLM?
AnythingLLM desktop app
What happened?
When requesting/api/v1/openai/chat/completions for anything-llm 1.6.0 desktop version, the max_token does not take effect, and the connected model is a local model(xinference)
Are there known steps to reproduce?
this is xinference: this is anything-llm desktop: