-
### What is the issue?
I had ollama compiled from source and it worked fine. Recently I rebuild it to the last version, and it seems to not use my GPU anymore (it uses a lot of CPU processes, and it …
-
Any configs in which `checkpoint_files` is a list of files > 4, use FormattedFiles utility to shrink the size of the file.
Example from [llama3/70B_lora](https://github.com/pytorch/torchtune/blob/…
-
### Contact Details
ksilverstein@mozilla.com
### What happened?
Summary: Using the llamafiler `/tokenize` endpoint does not seem to add special tokens when the corresponding flag is set to true, at…
-
I am using a locally loaded model, Llama-3.2-11B-Vision-Instruct.
![image](https://github.com/user-attachments/assets/7d517485-670c-49c8-aa4c-31d54675efc9)
![image](https://github.com/user-attac…
-
I got the following error when running model Imported from GGUF which is generated from the model fine-tuned with LoRA.
Error: llama runner process has terminated: GGML_ASSERT(src1t == GGML_TYPE_F…
-
Thanks for adding support to VLM.
I was using [this](https://github.com/stanfordnlp/dspy/blob/main/examples/vlm/mmmu.ipynb) notebook.Tried with the `Qwen2-VL-7B-Instruct` and `Llama-3.2-11B-Vision-…
-
Attempting to load vision model with `mllama` https://huggingface.co/mlx-community/Llama-3.2-11B-Vision-Instruct-8bit
MLX architecture
Is this meant to be supported on my machine/version? I can't …
-
When building like this:
```
jetson-containers build llama-vision
```
```
-- L4T_VERSION=36.4.0
-- JETPACK_VERSION=6.1
-- CUDA_VERSION=12.6
-- PYTHON_VERSION=3.10
-- LSB_RELEASE=22.04 (ja…
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [ ] 2. The bug has not been fixed in the latest version.
- [ ] 3. Please note that if the bug-related issue y…
-
# Language Model Overview
## OpenAI
| | gpt-4o | gpt-4o-mini …