-
### What happened?
I'm trying to add an ollama via the ui. I choose Add Model, provider ollama, public model name "llama3.2", litellm model name "llama3.2", provide base URL (same as env below), then…
-
I am trying to finetune llama3.2 Vision Instruct, and I am using the distributed recipe and example (lora) config as a starting point. Eventually, I am looking to use a custom dataset, but first, I am…
-
below evaluation is vllm settings for llama3.2 evaluation
```
lm_eval --model vllm \
--model_args pretrained=/home/jovyan/data-vol-1/models/meta-llama__Llama3.2-1B-Instruct,dtype=auto,gpu_m…
-
**Is your feature request related to a problem? Please describe.**
I would like to use vllm server with streaming support. they say that tools output is openai compatible, see: https://github.com/ver…
-
### Right Case
When I follow the doc : https://github.com/pytorch/executorch/blob/main/examples/models/llama/README.md#enablement,
I export the Llama3.2-1B-Instruct:int4-spinquant-eo8 model to xnnpa…
-
### Disappointing Performance
It's really strange that I have tried many **LLMs with vision** in `GGUF` format, listed in the official website, such as `Llama3.2-vision`, `llava`, `llava-llama3`, `ll…
-
### 🐛 Describe the bug
llama3.2 1B model run on QNN backend produce wrong result
### Versions
llama3.2 1B model run on QNN backend produce wrong result
-
Hello, I find that theres no lm_head weight in model checkpoints(.safetensors).
How does model load weight for the Linear Layer of lm_head ?
-
Hey there, @benjaminshafii! 👋 I'm running into some challenges getting the Ollama connection working in File Organizer 2000. I've tried several configurations, but I'm still encountering issues. Here …
-
Error using llama3.2-vision:11b
- Model is downloaded
![Screenshot 2024-11-27 114721](https://github.com/user-attachments/assets/ddfa14bf-cc3d-4fa9-99e7-7ca6446b8c8b)