-
*This is neither a feature request nor a bug but hopefully others may find it useful.*
I wanted to experiment with code refactoring using local models but still using the *awesome* chatgpt-shell. H…
-
deepspeed --num_gpus 3 dbgpt_hub/train/sft_train.py \
--deepspeed dbgpt_hub/configs/ds_config.json \
--quantization_bit 4 \
--model_name_or_path models/CodeLlama-13b-Instruct-hf \
…
-
![image](https://github.com/dbpunk-labs/octogen/assets/11504716/dddcbdca-3876-4009-9f95-cf4d15875a88)
-
**例行检查**
[//]: # (方框内删除已有的空格,填 x 号)
+ [ ] 我已确认目前没有类似 issue
+ [ ] 我已确认我已升级到最新版本
+ [ ] 我已完整查看过项目 README,已确定现有版本无法满足需求
+ [ ] 我理解并愿意跟进此 issue,协助测试和提供反馈
+ [ ] 我理解并认可上述内容,并理解项目维护者精力有限,**不遵循规则的 issue…
ixxmu updated
1 month ago
-
Hey! This is a super cool project.
I noticed your comment on https://news.ycombinator.com/item?id=38285191 about supporting open source models. I'm from the Anyscale team and we offer an OpenAI com…
-
I use tutorials/Quick_Deploy/vLLM to deploy codeLlama 7B, then I call metrics API, and a part of the metrics info is:
`
nv_inference_request_summary_us_count{model="triton-vllm-code-llama-model",v…
-
expected JSON body:
```json
{
"title":"Found a bug",
"body":"I'\''m having a problem with this.",
"assignees":["octocat"],
"milestone":1,
"state":"open",
"labels":["bug"]
}
```
expected…
-
我看到你们推出了两个版本的XAgentLlama,34B和7B对GPU要求是多少,它是要下载到本地调用对吧,
-
### What happened?
When is use the proxy with debugging the requests are nicely logged, like:
```
POST Request Sent from LiteLLM:
curl -X POST \
https://api-inference.huggingface.co/models/code…
-
### sqlcoder
目前sqlcoder在text2sql领域表现不错,在很多数据应用开发中被广泛使用,希望官方可以提供推理支持