Open maziyarpanahi opened 5 months ago
I have the same issue did you find a fix ?
I have the same issue did you find a fix ?
Unfortunately not. I am waiting for @winglian to chime in
looks like fastchat
needs to be updated to include this commit adding support for llama3 https://github.com/lm-sys/FastChat/commit/27a05b04a35510afb1d767ae7e5990cbd278f8fe
I'm getting a similar error for FFT with latest repo, and updating fastchat
. Hasn't the commit @benredmond raised been merged? Not sure why this error still crops up then.
@maziyarpanahi did you try doing a pip uninstall fschat
first and then installing fastchat again?
I'm using the latest docker image for axolotl and with this YAML and training starts and I let it run a few steps with any errors.
@maziyarpanahi did you try doing a
pip uninstall fschat
first and then installing fastchat again?
This is interesting! I usually follow the README to install Axolotl. I will give this a shot and get back to you.
I'm using the latest docker image for axolotl and with this YAML and training starts and I let it run a few steps with any errors.
This config is using chatml
, did you change both chatml
to llama3
? (so far I tried DPO and ORPO with llama3
and got this error)
UPDATE: I just tried uninstall/force-reinstal of fastchat with main, but it doesn't work. In fact, even for SFT I get:
AttributeError: LLAMA3. Did you mean: 'LLAMA2'?
@winglian I am actually going to test the Docker, could you please share which image did you use? (is it main-latest
?)
Hey, I' m getting the same issue trying to finetune Mistral AttributeError: LLAMA3. Did you mean: 'LLAMA2'?
with this yaml
base_model: mistralai/Mistral-7B-v0.1 model_type: MistralForCausalLM tokenizer_type: LlamaTokenizer
load_in_8bit: false load_in_4bit: true strict: false datasets:
adapter: qlora lora_model_dir:
sequence_len: 8192 sample_packing: false pad_to_sequence_len: true
lora_r: 32 lora_alpha: 16 lora_dropout: 0.05 lora_target_linear: true lora_fan_in_fan_out: lora_target_modules:
wandb_project: Mistral_Jobmaker wandb_entity: wandb_watch: wandb_name: wandb_log_model:
gradient_accumulation_steps: 1 micro_batch_size: 10 num_epochs: 1 optimizer: adamw_bnb_8bit lr_scheduler: cosine learning_rate: 0.0002
train_on_inputs: false group_by_length: false bf16: true fp16: false tf32: false
gradient_checkpointing: true early_stopping_patience: resume_from_checkpoint: local_rank: logging_steps: 1 xformers_attention: flash_attention: true
loss_watchdog_threshold: 5.0 loss_watchdog_patience: 3
warmup_steps: 10
eval_steps: 0.05
eval_table_size:
eval_table_max_new_tokens: 128
save_steps:
debug:
deepspeed:
weight_decay: 0.0
fsdp:
fsdp_config:
special_tokens:
bos_token: ""
eos_token: ""
unk_token: "
Here's work around while waiting for the fix: git checkout fff06af8d0514d41a62ce7a15ac17353b3b39fce
and pip install -e .
again.
@winglian I am actually going to test the Docker, could you please share which image did you use? (is it
main-latest
?)
yeah, main-latest
of this image https://hub.docker.com/r/winglian/axolotl-cloud/tags
@maziyarpanahi @IlonaBrst @nampdn it sounds like a broken virtual environment if you're installing fastchat from main and it's still raising an AttributeError for LLAMA3.
Hi! It's definitely something with the fastchat. I've tried my older environment with latest axolotl version and it works for me.
So, my findings so far and sorry if it's not complete:
main
was broken with LLAMA2 error, however, the suggestion to upgrade to the latest fastchat
didn't work, but this worked:pip install --force-reinstall git+https://github.com/lm-sys/FastChat.git@27a05b04a35510afb1d767ae7e5990cbd278f8fe
I will test this with llama3
template for DPO/SFT just to be sure and will report back.
Hello, colleagues and Wing Lian! I am encountering the same error. I am fixing the error by replacing the file src/axolotl/cli/train.py. In doing so, I am not using the DPO/KTO/ORPO strategies and so on.
I noticed the error a week ago. Before that, I was using the latest container from the docker hub and everything was working.
I am encountering the same error, too. @sklyar61 which version of src/axolotl/cli/train.py
are you using?
I am encountering the same error, too. @sklyar61 which version of
src/axolotl/cli/train.py
are you using?
version=“0.4.0”
Version has been 0.4.0
for a while now. So I would say the commit hash would be more accurate.
Please check that this issue hasn't been reported before.
Expected Behavior
Should be able to fine-tune using DPO with any chat template and type
Current behaviour
regardless of dataset type or chat_template, the training fails with:
Steps to reproduce
1- pull the latest from main and install it 2- run any DPO training with any datasets even with chatml template/type
Config yaml
Possible solution
No response
Which Operating Systems are you using?
Python Version
3.10
axolotl branch-commit
main
Acknowledgements