-
Hello !
i'm trying to setup ollama to run in a docker container, in order to have it run in runpod serverless function and to do so i'd like to pull a model file in my container image (embed the mo…
-
### Describe the bug
I am on the `dev` branch right now! Very important to note.
I loaded `mistral-7b-instruct-v0.1.Q5_K_M.gguf` and `mixtral-8x7b-instruct-v0.1.Q5_K_M.gguf` using llama.cpp and …
-
It seems to me that [PY007/TinyLlama-1.1B-Chat-v0.3](https://huggingface.co/PY007/TinyLlama-1.1B-Chat-v0.3/tree/main) model weights were updated after the the 1.5T weights we issued.
I just want to…
-
## 2024-01-09T12:30/12:55+09:00 🖐Smart Maps Meetup Weekly
Smart Maps Meetup Weekly の第51回です。
- 第1回 #27
- 第2回 #28
- 第3回 #29
- 第4回 #33
- 第5回 #34
- 第6回 #43
- 第7回 #51
- 第8回 #56
- 第9回 #60
- 第1…
-
Hi,
I'm running the latest version of llama.cpp (cloned yesterday from the Git repo) on macOS Sonoma 14.0 on a M1 MacBook Pro.
I tried to finetune a Llama model, and the training worked, however it …
-
Hi nice work on the paper. I pulled your code and model locally (LLaMA-7B-PoSE-YaRN-128k).
I run it against the LongChat evaluation "topic" retrieval but model does very very poorly > 4096 tokens
…
-
## ⚙️ Request New Models
- Link to an existing implementation (e.g. Hugging Face/Github):
https://huggingface.co/PY007/TinyLlama-1.1B-intermediate-step-480k-1T
https://huggingface.co/PY007/Tiny…
Fax updated
10 months ago
-
The script is tied to Alpaca prompt style. I can understand that basically all the prepare scripts enforces Alpaca style, however `generate/base` is agnostic (for good!)
[ prompt = generate_pro…
-
# ❓ Questions and Help
Thanks for the great project. Recently torch 2.1.0 was released as stable. Are there any build plans or development releases for this?
-
```
-- Quantizing...
-- Layer: model.layers.0 (Attention)
-- Linear: model.layers.0.self_attn.q_proj -> 0.25:4b_32g/0.75:2b_32g s4, 2.63 bpw
-- Linear: model.layers.0.self_attn.k_proj -> 0.05:…