-
I use one NVIDIA L40S(48GB VRAM) to train a Lora for Flux, and here is my training script:
`./sd-scripts/flux_train_network.py --pretrained_model_name_or_path ./model/flux1-dev.safetensors --clip_l .…
-
### System Info
trl official DPO examples. Finetune llama3.1 with lora.
params:
lora_rank: 32
lora_target: all
pref_beta: 0.2
pref_loss: sigmoid
### dataset
dataset: train_data
template:…
-
Hello!
Thanks for your great repo!
Our work on multi-task learning named "Multi-Task Dense Prediction via Mixture of Low-Rank Experts (CVPR 2024)" has been released recently. Would you like to tak…
-
#### Describe the workflow you want to enable
If the target is discrete multiclass, but ordinal (ordered) in nature (e.g. Likert scale, user ratings, preference levels), as opposed to nominal, wo…
-
- PyTorch-Forecasting version: 0.8.4
- PyTorch version: 1.8.0
- Python version: 3.8.8
- Operating System: CentOS
### Expected behavior
I'm working through the _Demand forecasting with the T…
-
**Describe the bug**
I am trying to fine-tune DeepSeek-Coder-V2-Lite-Instruct (16B) on a system with 8 MI300X GPUs. Running on any number of GPUs less than 8 works as expected and runs to completion. …
-
I was using DistributedDataParallel to train a model on single machine 8 gpus. I thought by using DistributedDataParallel, memory on each gpu should be approximately the same, however, there is one gp…
-
At the moment the first worksheets that are shown are those submitted last. Later (when there is data and more worksheets) it might be better to rank worksheets with a Machine learning model.
-
I've been looking for an auto-ml framework that solves learning to rank problem, but I haven't found one.Does autogluon support?Thanks.
-
## Environment
AWS Deep Learning Machine with 8xA100 and CUDA 11.8
## To reproduce
Steps to reproduce the behavior:
1. Use StreamingDataset to load ImageNet from a local SSD using DDP.
…