FlagOpen / FlagEmbedding

Retrieval and Retrieval-augmented LLMs
MIT License
7.04k stars 514 forks source link

how to adjust hyperparameter for finetune llm embed #743

Open QuangTQV opened 5 months ago

QuangTQV commented 5 months ago

llm embed has the following training script. I don't know how to adjust hyperparameters like train_batch_size, learning rate, warmup_ratio, ...

torchrun --nproc_per_node=8 run_dense.py \ --output_dir data/outputs/tool \ --train_data llm-embedder:tool/toolbench/train.json \ --eval_data llm-embedder:tool/toolbench/test.json \ --corpus llm-embedder:tool/toolbench/corpus.json \ --key_template {text} \ --metrics ndcg \ --eval_steps 2000 \ --save_steps 2000 \ --max_steps 2000 \ --data_root /data/llm-embedder

staoxiao commented 5 months ago

This script uses the huggingface trainer to do fine-tuning, so you can use the hyper-arguments on this page: https://huggingface.co/docs/transformers/main_classes/trainer#transformers.TrainingArguments

QuangTQV commented 5 months ago

This script uses the huggingface trainer to do fine-tuning, so you can use the hyper-arguments on this page: https://huggingface.co/docs/transformers/main_classes/trainer#transformers.TrainingArguments

How can I create a new task like llm embedded task (with separate instructions for query and separate instructions for key) ?