-
I can confirm that `async` completion works via LiteLLM:
```python
from litellm import acompletion
import asyncio
async def test_get_response():
user_message = "Hello, who are you?"
…
-
### Version
Command-line (Python) version
### Operating System
Ubuntu Linux
### What happened?
The addition of Claude 3 support has been great.
The biggest advantage of Claude 3 over current ope…
-
### What do you need?
this is awesome project, but it needs ollama support.
the OpenAI api is the ease way out.
please add support for local LLMs too.
thank you
-
The pattern I need to figure out for many of the tests I write is how to safely use this with an authenticated API endpoint. Basically I want to:
- Have code that talks to a protected API, e.g. the…
-
I am trying out the streaming example provided in your documentation [here](https://langchain-ai.github.io/langgraphjs/how-tos/stream-tokens/#call-streamevents) with v2 schema. I got it to work based …
-
### 您希望的更新和改进是什么 | Update or Improve
---
- 希望「临时翻译页面」能对更多翻译服务(其他 LLM 模型、DeepLX、自定义 AI 等)进行适配
---
![image](https://github.com/user-attachments/assets/3d6bece7-ae09-47f7-bc27-1585daad65f3)
…
-
### Bug Description
Claude模型的请求地址改了吗?原来是BASE_URL/v1/chat/completions,现在Claude模型的请求地址改为了BASE_URL/v1/messages,于是报错了
### Steps to Reproduce
使用ChatGPT next web,部署方式是vercel
返回信息如下:
```json
{
"er…
-
### What happened?
When using the AsyncOpenAI client with an Anthropic model on litellm, any new completion request will kill previous requests, causing an httpx.ReadError exception (see log output)
…
-
Hi,
For my use case the number of functions and instructions exceeds the max tokens for gpt3 so i am force to use gpt4 which is painfully slow and expensive and after a couple of days i gave up.
I…
-
I'd really appreciate that feature if it's not too much trouble