-
# Steps to reproduce
1. install ollama in WSL `ollama version is 0.1.29`
1a. get vim version: ```VIM - Vi IMproved 9.1 (2024 Jan 02, compiled Jan 01 1980 00:00:00)
Included patches: 1-75
Compiled…
-
### Anything you want to discuss about vllm.
I run into the below error when using meta-llama/CodeLlama-7b-Instruct-hf with `vllm==0.4.0, torch==2.1.2`, the code works perfectly with` vllm==0.2.1`, b…
-
### Description
Issue: an unrequested call to the secondary diagram model is made every time the chat button is clicked. resulting in extra charges
Step 1. Configure Model settings with OpenRout…
-
Currently on 3 chat templates is present: https://github.com/TanvirOnGH/vscode-ollama-modelfile/blob/dev/snippets/modelfile.json#L37-L104.
## TODO Templates
- [x] ChatML
- [ ] LLaMa2
- [x] LLaMa…
-
请问每一列分别代表什么?我看到写的overall是0.789但是首页标注着一列是median
CodeLlama-13b-Instruct-hf_lora 0.789 sft train by our this project,only used spider train dataset, the same eval way in this project with lora SFT. The …
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a…
-
GPU: 2 ARC CARD
running following example,
[inference-ipex-llm](https://github.com/intel-analytics/ipex-llm/tree/main/python/llm/example/GPU/Pipeline-Parallel-Inference)
**for mistral and codell…
-
I hope this message finds you well. I recently had the opportunity to experiment with the Codellama-7b-Instruct model from GitHub repository and was pleased to observe its promising performance. Encou…
-
**Is your feature request related to a problem? Please describe.**
NAN
**Describe the solution you'd like**
NAN
**Describe alternatives you've considered**
NAN
**Additional context**
NAN
…
-
Hi team,
I am trying to build llama engine files using `nvcr.io/nvidia/tritonserver:24.05-trtllm-python-py3` container and was getting the below error:
```
Traceback (most recent call last):
…