-
Hey there, @benjaminshafii! 👋 I'm running into some challenges getting the Ollama connection working in File Organizer 2000. I've tried several configurations, but I'm still encountering issues. Here …
-
Error using llama3.2-vision:11b
- Model is downloaded
![Screenshot 2024-11-27 114721](https://github.com/user-attachments/assets/ddfa14bf-cc3d-4fa9-99e7-7ca6446b8c8b)
-
### What is the issue?
When running a query with qwen2.5 (32b) it only uses the CPU for some reason. I can switch to another model (llama, phi, gemma) and they all utilize the GPU.
Reproduce:
1. …
-
I have downloaded LLAMA 3.2 1B Model from Hugging face with optimum-cli
optimum-cli export openvino --model meta-llama/Llama-3.2-1B-Instruct llama3.2-1b/1
Below are files downloaded
!…
-
When trying to run `Llama3.2-3B-Instruct-QLORA_INT4_EO8`, I'm getting the error:
`OSError: meta-llama/Llama-3.2-3B-Instruct-QLORA_INT4_EO8 does not appear to have a file named pytorch_model.bin, mo…
-
# Language Model Overview
## OpenAI
| | gpt-4o | gpt-4o-mini …
-
May I ask what would the format be for multi turn method? Our method need 2 round of conversation, and used llama3 70b as the judge.
-
**Describe the bug**
When running the latest drop (v0) of Llama3.1-70B on LoudBox with firmware [80.13.1.0](https://github.com/tenstorrent/tt-firmware/blob/main/fw_pack-80.13.1.0.fwbundle) we experien…
-
In this repo the Llama3 tokenizer sets the `` special token to `128011` https://github.com/meta-llama/llama-models/blob/ec6b56330258f6c544a6ca95c52a2aee09d8e3ca/models/llama3/api/tokenizer.py#L79-L101…
-
I'm using :
- MacOS Ventura 13.2.1
- MacBook Air M1
When I execute the command :
```python setup_env.py --hf-repo HF1BitLLM/Llama3-8B-1.58-100B-tokens -q i2_s```
I got the message:
```
INFO:root:C…