-
### Feature request
Thank you for this great repo. I have recently integrated OpenLLM with Microsoft Word through my local Word Add-in ([GPTLocalhost](https://gptlocalhost.com/demo#OpenLLM)). Would i…
-
Hi, Thanks for integrating the Qserve.
I’ve been interested in Qserve and really wanted to use it on TensorRT-LLM for speed.
This week, your team integrated it. I tested Llama-3-8B-Instruct + Qserve…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
-
## Background
We are curious to know whether ontology score correlates to performance on downstream tasks.
We could evaluate performance on downstream tasks ourselves, but as a first approximation,…
-
### Bug Description
HuggingFaceEmbedding 依赖于llama-index-core-0.10.68.post1。OpenLLM 依赖于llama-index-core-0.11.9。
运行OpenLLm成功,则HuggingFaceEmbedding报错:
ile [~/anaconda3/envs/llama_factory/lib/python3.1…
-
what the difference is between Math Lvl 5 in the ZeroEval evaluation and the Eleuther eval harness (OpenLLM leaderboard v2)?
For 3.1-70B, ZeroEval is showing 43.3 and HF leaderboard is showing 28%
…
-
https://github.com/bentoml/OpenLLM
-
how can i use openllm for local lora model?
-
https://note.com/npaka/n/n1d99253ae2cf?sub_rt=share_pw
-
### Describe the bug
I want to run Qwen0.5b on a k8s cluster without GPU, but the service startup has failed so far. Is there any way to support CPU machines
![Uploading 屏幕截图 2024-09-09 164657.jpg…]…