-
great work! can you do a mistral 1b tinyllama?
mistral ai is good.
-
I am currently running the container on unraid. I have used the docker compose file as well as maually creating the container and changing storage mounts. I am able to download the models from hf and …
-
### Discussed in https://github.com/ggerganov/whisper.cpp/discussions/1656
Originally posted by **Sing303** December 19, 2023
Now when I try to use quantization models with FULL GPU CUBLAS, in…
-
I'm currently trying out the ollama app on my iMac (i7/Vega64) and I can't seem to get it to use my GPU.
I have tried running it with num_gpu 1 but that generated the warnings below.
`
2023/11/…
-
# ❓ Questions and Help
Thanks for the great project. Recently torch 2.1.0 was released as stable. Are there any build plans or development releases for this?
-
First thanks for all of the efforts you have done for the TinyLlama project - it's awesome!
Recently I found a spurious problem. When there is only one card, the training gradient would disappear. …
-
The TinyLlama project aims to pretrain a 1.1B Llama on 3T tokens. So that model should be an ideal draft model for speculative inference.
https://github.com/jzhang38/TinyLlama
https://huggingfac…
-
If the integrity check fails, there is no feedback for this code in the tinyllama.py:
def train(fabric, state, train_dataloader, val_dataloader, monitor, resume):
model = state["model"]
opt…
-
First, I want to express my gratitude about this project. I think TinyLlama has a lot of potential and we're just starting to see it. Cudos!
I'm pretty new to this exciting field and this is the fi…
-
The `ollama serve` command runs as normally with the detection of my GPU:
```
2024/01/09 14:37:45 gpu.go:34: Detecting GPU type
ama 2024/01/09 14:37:45 gpu.go:53: Nvidia GPU detected
ggml_in…