-
- [ ] AI21StudioModel
- [ ] AnthropicModel
- [ ] Anthropic Tool Model
- [ ] Cohere Model
- [ ] Cohere Tool Model
- [ ] DeepInfra Model
- [ ] DeepSeek Model
- [ ] GeminiPro Model
- [ ] Gemini T…
-
`DEEPSEEK_FREE_URL=https://api.deepseek.com/chat/completions,`
应该注意要有引号
`DEEPSEEK_FREE_URL='https://api.deepseek.com/chat/completions',`
-
Hey, it would be awesome if https://github.com/deepseek-ai/DeepSeek-V2 would be supported if it's not too much work, I'd really like to quantize and publish it. (Also, is multi-GPU quantization possib…
-
### Your current environment
```text
Collecting environment information...
PyTorch version: 2.3.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
…
-
-
### Motivation
VLLM has announced their support for running llama3.1-405b-fp8 on 8xA100. This is the [blog](https://blog.vllm.ai/2024/07/23/llama31.html)
Does sglang support running DeepSeek-Coder…
-
我的代码如下
```
from langchain_openai import ChatOpenAI, OpenAIEmbeddings
llm = ChatOpenAI(
model='deepseek-chat',
openai_api_key='xxx',
openai_api_base='https://api.deepsee…
-
I am using the newest AMI image from yesterday, with optimum-neuronx 0.0.17 (https://aws.amazon.com/marketplace/pp/prodview-gr3e6yiscria2) I have not tried using another image yet.
I am trying to …
-
Hi!
Does trtllm support https://huggingface.co/deepseek-ai/deepseek-moe-16b-base ? Do you have any plans to support?
-
## problem description:
Hi! While training the original code using deepseek-ai/deepseek-coder-7b-base-v1.5 as base model on 8 A100 GPUs, I encountered the OOM issues below:
`Traceback (most recent c…