-
Hi,i finetune MGM-2B on coco, but i got the warning that:
`{'loss': 6.9221, 'grad_norm': tensor(18.7422, device='cuda:0', dtype=torch.float64), 'learning_rate': 9.203084832904885e-06, 'epoch': 0.01}…
-
### Feature request
Generalize the functionality in [processing_llava.py](https://github.com/huggingface/transformers/blob/main/src/transformers/models/llava/processing_llava.py) to include other t…
-
### Enhancement Request - Support for Additional LLM Types
#### Description:
After reviewing the [Mediapipe documentation](https://developers.google.com/mediapipe/solutions/genai/llm_inference) an…
-
the instruction code for mpt-7b works fine when using older version 20240123, but when updating to the latest branch, using the new code, always have OOM error with multiple gpus, even when using 8*A1…
-
C/C++: CMake Error: CMake was unable to find a build program corresponding to "Ninja". CMAKE_MAKE_PROGRAM is not set. You probably need to select a different build tool.
-
### System Info
Name: transformers
Version: 4.45.0.dev0
Name: trl
Version: 0.8.6
### Information
- [ ] The official example scripts
- [X] My own modified scripts
### Tasks
- [ ] A…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
windows 11
python 3.12
今日最新源码安装 LLaMA-Factory
### Reproduction
```yaml
### model
model_na…
neavo updated
3 months ago
-
### What is the issue?
I see this issue has been partially reported, but none of the previous reports seem to be extensive in their tests of possible methods to set this option.
The problem:
Ol…
-
### Your current environment
```text
The output of `python collect_env.py`
```
### 🐛 Describe the bug
### On the Tesla T4 the model "hangs" after loading the model (the vram usage spikes normal…
-
### URL
https://python.langchain.com/docs/tutorials/sql_qa/
### Checklist
- [x] I added a very descriptive title to this issue.
- [X] I included a link to the documentation page I am referring to (…