-
/kind bug
**Describe the solution you'd like**
Current huggingfaceserver requirements [set in pyproject toml](https://github.com/kserve/kserve/blob/master/python/huggingfaceserver/pyproject.toml#L…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
-
### What is the issue?
`Error: llama runner process has terminated: signal: segmentation fault (core dumped)`. It occurs while loading larger models, that are still within the VRAM capacity. Here I…
-
I see this in the readme
`Supports EXL2, GPTQ and FP16 models`
but no links to the models themselves?
Can you give me the HF URLs for those recommended models? Or the models you think are "best" f…
-
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [X] 2. The bug has not been fixed in the latest version.
- [x] 3. Please note that if the bug-related iss…
wlwqq updated
1 month ago
-
In the documentation the Bionic GPT it is mentioned that it works with ollama and OpenAPI compatible backends and it is demonstrated running a local gemma model. I could not find information on how to…
-
*This is neither a feature request nor a bug but hopefully others may find it useful.*
I wanted to experiment with code refactoring using local models but still using the *awesome* chatgpt-shell. H…
-
''ipex-llm[cpp]==2.5.0b20240527 is consistent with [v0.1.34] of ollama.
Our current version is consistent with [v0.1.39] of ollama.''
Is it possible to update supported ollama version to 0.3.x?
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ ] I am running the latest code. Development is very rapid so there are no tagged versions as of…