vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
29.05k stars 4.33k forks source link

/v1/assistants HTTP/1.1" 404 Not Found #1590

Closed xxm1668 closed 7 months ago

simon-mo commented 11 months ago

Hi @xxm1668 which tools are you using? vLLM currently do not support OpenAI assistant or threads API, which does require higher level orchestration. I would to learn more about your feature request.

xxm1668 commented 11 months ago
image

I want start this demo

hmellor commented 7 months ago

This endpoint isn't supported, see https://github.com/vllm-project/vllm/blob/main/vllm/entrypoints/openai/api_server.py for the available endpoints

simon-mo commented 7 months ago

Please checkout other projects like MemGPT where the assistants API is fully implemented.

Cognitus-Stuti commented 1 day ago

Has this been updated?