Closed ganler closed 6 months ago
It seems vLLM has a very restrictive context size limit for Llama-3 (trained on 8k max) that anything beyond 8k is rejected. DS series are fine and its ctx size can be extended as is shown in the CodeQwen report.
*CodeQwen TypeScript results are missing. will catch that up soon.
CodeQwen results updated.
Running databricks/dbrx-instruct
as well.
databricks/dbrx-instruct
produces empty output all the time. I think I will skip it then.
Added bigcode/starcoder2-instruct-15b-v0.1
.
Got rate limited by Gemini Pro and Claude....
OSS model 🤗
CodeLlama
🤔 marks models with 8~16k context trained. May need to modify the
config.json
.DeepSeekCoder
Llama 3
CodeQwen
Qwen1.5
CodeGemma
Mistral
Starcoder2
Private model 💲💰💸