-
https://github.com/karpathy/nanoGPT/blob/master/data/shakespeare/prepare.py
../data/allamo_1B_dataset/
```
input.txt
train.bin
val.bin
```
The idea for this project is great, thank you.…
-
### System Info
TGI version latest;single NVIDIA GeForce RTX 3090;
### Information
- [X] Docker
- [ ] The CLI directly
### Tasks
- [X] An officially supported command
- [ ] My own modifications
…
-
This might be a silly question, but when using the Llama3.1 base model I can effortlessly pass in tools when running it in Ollama.
```
response = ollama.chat(
model='llama3.1'…
-
Great work!
I am trying to run pyramidinfer with a Llama3-8B-Instruct model, but it seems that the version of "transformers" is too old to load the weight of Llama3-8B model.
I ran this command …
-
AssertionError: Rank 11: found NaN in local grad norm in backward pass before data-parallel communication collective. Device: 3
![error1](https://github.com/user-attachments/assets/97788e23-684c-…
-
Hello,
I am tryoing to SFT train llama3.2 11B vision instruct model. on a dataset that answer a question on an image using a context (could be more than one image). My code is:
```
def format_dat…
-
As title says. I can upload images to llava but not llama3.2-vision
llama3.2-vision:11b-instruct-q8_0
Nayar updated
2 weeks ago
-
### What is the issue?
If I try to run the `llama3.2-vision` model using `ollama run llama3.2-vision` on my Arch Linux machine, I get this error:
```
Error: llama runner process has terminated: GG…
-
It looks like `seed` is not working when using it in `chat()`, I get no consistent responses when setting it. I also ran the same test with `ollamar` and did receive consistent results:
``` r
ollama…
-
I am follwing the [instructions in the Llama2 README](https://github.com/pytorch/executorch/blob/d9aeca556566104c2594ec482a673b9ec5b11390/examples/models/llama2/README.md#instructions) to test llama m…