Closed cocaer closed 4 months ago
Hi,
Thank you for your interest! ALMA should be generalized across all domains. If you are unsure about the correct usage of ALMA, a recommended approach is to run the following command:
accelerate launch --config_file configs/deepspeed_eval_config_bf16.yaml \
run_llmmt.py \
--model_name_or_path haoranxu/ALMA-13B-R \
--text_test_file $YOUR_RAW_TEXT \
--do_predict \
--low_cpu_mem_usage \
--language_pairs en-cs(your single direction) \
--mmt_data_path ./human_written_data/ \
--per_device_eval_batch_size 4 \
--output_dir ./your_output_dir/ \
--predict_with_generate \
--max_new_tokens 512 \
--max_source_length 512 \
--bf16 \
--seed 42 \
--num_beams 5 \
--overwrite_cache \
--overwrite_output_dir
where text_test_file
is a single raw test data file in text format (one line by one sentence you want to translate), this will override themmt_data_path
.
It would also be helpful if you could share your detailed results to assist with debugging.
I get the good results after following your instruction. It seems i used the broken model file. Thanks a lot!
Hi Haoran,
I've been experimenting with the ALMA and ALMA-R models to translate the multi-domain dataset available here. Unfortunately, I'm observing subpar results, particularly with the ALMA-R model, in terms of both BLEU and COMET scores.
Below is the code snippet I've been using:
Could there be any incorrect operations in my approach, or is this typical performance given that the ALMA models were trained primarily on the WMT and Flores datasets? Any insights or recommendations would be greatly appreciated.
Baijun