-
I noticed that currently only a few series of models, including **Qwen, ChatGLM, and GPT**, support **IFB**. The lack of support for other models has severely impacted the practicality of the TRT-LLM …
-
### Your current environment
Edit 1
```text
Collecting environment information...
PyTorch version: 2.2.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorc…
-
### Issue
*Environment*
```
Windows 11
conda 23.7.2
Python 3.11.9
aider 0.57.1
```
Export env.variables:
`(aider) D:\SD> set OPENAI_API_KEY=`
`(aider) D:\SD> set OPENAI_API_BASE=https:/…
-
# Trending repositories for C#
1. [**ihm-tswow / Links-Awakening-DX-HD**](https://github.com/ihm-tswow/Links-Awakening-DX-HD)
__Unaffiliated fork off https://linksawakeningdxhd.it…
-
### Which API Provider are you using?
Ollama
### Which Model are you using?
Qwen 2.5 7B & Ollama 3.1 8B
### What happened?
When I try to tag or use a specific file from my codebase it starts inde…
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [X] 2. The bug has not been fixed in the latest version.
### Describe the bug
发现一个问题, 在rtx4090 * 8 环境, 针对q…
-
### Which API Provider are you using?
OpenAI Compatible
### Which Model are you using?
QWEN 2.5 72B INSTRUCT
### What happened?
When Claude DEV tries to connect to the HYPERVOLIC provider’s API, …
-
### Describe the bug
I followed the document and video as well, installed in docker and configured the .env file as well. However, I am getting errors. PFA pf the pictures for your reference
![a11]…
-
### Proposal to improve performance
_No response_
### Report of performance regression
At the begining there is a high throughput when processing 8 parallel requests:
![image](https://github.com/v…
-
本地环境:
cuda: 11.8
python: 3.11
transformers: 4.39.3
lmdeploy: 0.4.0
torch: 2.2.1
----------------------------------------------------------------
问题描述:
我使用以下命令启动部署了CodeQwen1.5-7b模…