-
### 🚀 The feature, motivation and pitch
ollama vision is new:
https://ollama.com/x/llama3.2-vision
providers:
inference:
- provider_id: remote::ollama
provider_type: remote::ollama
…
-
### What is the issue?
I tried to import finetuned llama-3.2-11b-vision, but I got "Error: unsupported architecture."
In order to make sure my model is not the problem, I downloaded [meta-llama/Ll…
-
Hi @lea-33 ,
how about introducing another LLM endpoint: [ollama](https://ollama.com/)? There were recently new vision-models published, namely [llama3.2-vision](https://ollama.com/library/llama3.2…
-
### Your current environment
The output of `python collect_env.py`
```text
Your output of `python collect_env.py` here
```
### Model Input Dumps
_No response_
### 🐛 Describe the bug
…
-
### Command:
**llama stack run Llama3.2-11B-Vision-Instruct --port 5000**
**Output:**
```
Using config `/Users/mac/.llama/builds/conda/Llama3.2-11B-Vision-Instruct-run.yaml`
Resolved 4 prov…
-
-
### What is the issue?
when I use llama3.2-vision:90b ,the model always response so slow. how can I do?
And the GPU is not fully used. CPU used very high.
it's run on V100*4 = 64G GPU
Is anyone ca…
-
In this repo the Llama3 tokenizer sets the `` special token to `128011` https://github.com/meta-llama/llama-models/blob/ec6b56330258f6c544a6ca95c52a2aee09d8e3ca/models/llama3/api/tokenizer.py#L79-L101…
-
## Issue
I keep getting `nan` loss when training Llama-3.2-vision
I tried:
- gradient clipping
- lower learning rate
- higher batch size, lora rank and alpha
But with no success.
## …
-
Dear @xiaoachen98,
Thank you very much for releasing the code. I am running your fine-tuning script to replicate your LLaMA 3 results. However, I am getting zero loss after the first iteration. Hav…