-
### Expected Behavior
-
### Actual Behavior
![image](https://github.com/user-attachments/assets/1f9608dc-4631-41c3-bd2a-bfe506d39104)
SD15 and Flux work fine, the problem is only with SDXL
Co…
-
### System Info
- `transformers` version: 4.46.1
- Platform: macOS-15.1-arm64-arm-64bit
- Python version: 3.11.10
- Huggingface_hub version: 0.26.2
- Safetensors version: 0.4.5
- Accelerate ve…
-
### 🐛 Describe the bug
Getting `ValueError: Pointer argument (at 0) cannot be accessed from Triton (cpu tensor?)` when doing inference using HF `from_pretrained()` with `device_map="auto"`.
### …
-
### Your question
Error while loading checkpoint:!!! Exception during processing !!! 'tokenizers.AddedToken' object has no attribute 'special'
### Logs
```powershell
got prompt
model weight dtype…
npubg updated
2 weeks ago
-
With the latest torch (2.4) and iree-turbine, we are seeing this MLIR verification failure come up for a lot of our models during the export stage (aot.export).
Instructions to reproduce this error…
-
When calculating the total steps, shouldn't we use `number of batches * epoch size` ? In this case, it would be `self.total_steps = (len(train_loader.dataset) // tb_size) * ab_size` instead of `self.t…
-
``` python
# sagemaker config
instance_type = "ml.g4dn.xlarge"
number_of_gpu = 1
health_check_timeout = 300
config = {
"HF_MODEL_ID": "unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit", # mode…
-
Distributed environment: FSDP Backend: nccl
Num processes: 4
Process index: 3
Local process index: 3
Device: cuda:0
Mixed precision type: bf16
Distributed environment: FSDP Backend: nccl
…
-
### System Info
The current Transformers framework doesn't support the gguf quantized model files from deepseek2. Can you please advise when this support might be added? @SunMarc @MekkCyber
###…
-
### Problem Description
Hi,
When doing text generation with Mistral 7b with Hugginface transformers on a MI100 GPU, I can see in the collected torch trace that a lot of time is wasted due a hipMem…
Epliz updated
3 weeks ago