-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
When I want to reproduce the original model results of mistral-7b-v0.2 without `flash-attn` I got the error:
```
Traceback (most recent call last):
File "/home/yuanye/long_llm/InfLLM/benchmark/pr…
-
I want to lora finetune Qwen2.5-32B-Instruct-AWQ model(4bit quant already) through llama-factory, but occured an error.
`
[INFO|configuration_utils.py:677] 2024-11-21 19:44:25,957 >> loading confi…
-
I'm trying to run a DeepSeek-V2.5 model.
Command used: ```python -m ktransformers.local_chat --model_path ./DeepSeek-V2.5/ --gguf_path ../ ```
```
Chat: hi
Traceback (most recent call last):
Fi…
-
# Export Error Summary Dashboard ##
- This report is generated from branch of https://github.com/huggingface/optimum/pull/1712
- Produced by `RUN_SLOW=1 pytest tests/exporters/onnx -k "test_export…
-
### System Info
- `transformers` version: 4.47.0.dev0
- Platform: Linux-5.15.0-1052-oracle-x86_64-with-glibc2.35
- Python version: 3.10.12
- Huggingface_hub version: 0.25.2
- Safetensors version:…
-
I'm now trying to train llama3.1 with GRIT pipeline.
At first I directly change ``--model_name_or_path`` and run the training code (the training script I used is as follows)
```
#!/bin/bash
#SB…
-
I'm currently working with the code and I'm having some trouble OVERRIDE forward where the self.module.image_newline attribute is set or initialized in the model.
I've traced the model through the …
-
hey, while running on 4bit quantized model from https://huggingface.co/ThetaCursed/Ovis1.6-Gemma2-9B-bnb-4bit i am getting the following error
```
{
"name": "RuntimeError",
"message": "self an…
-
### Is there an existing issue for this problem?
- [X] I have searched the existing issues
### Operating system
Linux
### GPU vendor
AMD (ROCm)
### GPU model
AMD Radeon RX 7800 XT…