-
llama3-sft.yaml
```yaml
### model
model_name_or_path: models/llama-3-8b-Instruct
### method
stage: sft
do_train: true
finetuning_type: full
deepspeed: examples/deepspeed/ds_z3_config.json
…
-
### System Info
- Platform: Linux-5.15.0-86-generic-x86_64-with-glibc2.35
- Python version: 3.10.14
- PyTorch version: 2.4.1
- CUDA device: NVIDIA A100-SXM4-80GB
- Transformers version: 4.45.0.…
-
I think there is a need for a priority that's higher than "user-blocking" for certain DOM rendering cases.
This priority would be synchronous wrt the outermost `scheduler.postTask()` call, but nest…
-
i want to finetune llama2-7b-hf using example finetune script [https://github.com/microsoft/Megatron-DeepSpeed/blob/main/examples_deepspeed/finetune_hf_llama/finetune_llama.sh](url)
when i run…
-
CUDA_VISIBLE_DEVICES=0 python /home/ubuntu/TextToSQL/DB-GPT-Hub/src/dbgpt-hub-sql/dbgpt_hub_sql/train/sft_train.py\
--model_name_or_path /home/ubuntu/.cache/modelscope/hub/qwen/Qwen2___5-Coder-7B…
-
I have a workflow that tries to insert 1 million rows into a Postgres db in batches. I tried it with batches of 10, 100, 1000 and 10000. They are all failing with errors that Workflow history size/cou…
-
I have modified the script as follows:
(damo) [root@bgmm01 custom]# cat finetune.sh
#!/bin/bash
# Environment Variables
export CUDA_VISIBLE_DEVICES=4,5
WORLD_SIZE=${1:-1}
NPROC_PER_NODE=${2…
-
Hi there
so SD Training on 1 GPU Works just fine
but as soon as i enable multi gpu with 2 GPUs i get this error:
![Clipboard_08-18-2024_01](https://github.com/user-attachments/assets/8dc2bd36-ddc…
-
### Issue Content:
**Description:**
It usually happens during lora training for some time.
I encountered a `subprocess.CalledProcessError` when running the `train_network.py` script using the …
-
### Please check that this issue hasn't been reported before.
- [X] I searched previous [Bug Reports](https://github.com/axolotl-ai-cloud/axolotl/labels/bug) didn't find any similar reports.
### Exp…