-
I would like another masc version and 2 fem versions.
![Image](https://github.com/untamed-team/project-untamed/assets/106409863/9c97eb95-5826-4768-ac14-f9b5ad0db765)
-
Full model fine-tuning code is given below. How can i modify the code to train Qlora based model.
```import sys
import os
current_directory = os.path.dirname(os.path.abspath(__file__))
sys.path.…
-
would it be possible to add the ability to change trainer info (name, gender, gametime, etc) to this program? i think it would be nice
-
Hey Everyone,
I was trying to finetune gemma-2-2b-it with my local PC which has got A3000 GPU inside. I have followed conda install method.
This is my trainer:
``trainer = SFTTrainer(
mo…
-
Quick question regarding the compatibility of ProGen2 with the HuggingFace Trainer - I would like to implement distributed strategies, deepspeed etc, and the Trainer class with accelerate is able to h…
-
Trying to run train.py and it doesn't work. Looks like neural network architecture is incompatible with some constrains? Is a specific version of torch library required?
```
[2024-10-04 17:46:10,588…
-
### Feature request
Adding generation configurations to the parameters that can be tuned in a `Trainer`.
### Motivation
When defining the Optuna hyper-parameter space, I would like to invest…
-
# Overview
Race condition? leading to a crash when multiple GPUs (processes) are used and the output directory doesn't exist.
## Steps to reproduce
Run a multiple GPU job with `torchrun` and …
-
### What happened?
Training checkpoints for large models (num_channels equal or greater than 912) become unreadable by pytorch and hence can't be used to resume or fork runs.
Note - for now thi…
-
### System Info
- `transformers` version: 4.44.2
- Platform: macOS-14.4-arm64-arm-64bit
- Python version: 3.12.2
- Huggingface_hub version: 0.24.5
- Safetensors version: 0.4.3
- Accelerate versi…