-
### System Info
Collecting environment information...
PyTorch version: 2.2.2
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: macOS 13.6.6 (x86_64)
G…
-
Use case: Users have pre-provisioned PVs that contain models on them and support ReadManyOnly. The user would be responsible for ensuring a compatible model is stored on the PV and creating a PVC.
…
-
### Background Description
Ref: https://github.com/ggerganov/llama.cpp/pull/7553 , required for supporting future vision models (https://github.com/ggerganov/llama.cpp/issues/8010)
I initially pla…
-
### System Info
1. Apple M2 Pro
2. mac OS Sequoia v15.0.1
3. Google Chrome v130.0.6723.116
4. @huggingface/transformers v3.0.2
### Environment/Platform
- [X] Website/web-app
- [ ] Browser…
-
To get this to work, first you have to get an external AMD GPU working on Pi OS. The most up-to-date instructions are currently on my website: [Get an AMD Radeon 6000/7000-series GPU running on Pi 5](…
-
Any plan to support the latest Llama-3.2-11B-Vision-Instruct and Molmo?
-
When I run the fp16 Llama-3 8b model using the evaluation framework provided I am unable to reproduce the perplexities reported in the PV-tuning paper. I get a perplexity of **5.7559** for Wikitext2 (…
-
Hello, I tried to train Llama3.2 3B. It's a full finetune, not a lora, but Unsloth always crashes under varying conditions when the model should be saved. Hardware was runpod in all cases, different c…
-
### Describe the bug
webui crashes after sending prompt
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
1) `./start_linux.sh`
2) load a model
…
-
We want to deploy https://huggingface.co/unsloth/Llama-3.2-1B-Instruct-bnb-4bit which is 4-bit quantized version of llama-3.2-1B model. It is quantized using bitsandbytes. Can we deploy this using ten…