Open Cheungki opened 12 months ago
can you provide the execution command and with model name, batch size, n_samples..?
Sorry for the delay. I just use greedy sampling and run with the following command:
accelerate launch --mixed_precision bf16 ./bigcode-evaluation-harness/main.py
--model /mnt/models/codellama-7b-python
--tasks humaneval
--max_length_generation 512
--batch_size 1
--do_sample False
--precision bf16
--max_memory_per_gpu 'auto'
--allow_code_execution
--trust_remote_code
--save_generations
--use_auth_token
--metric_output_path ./bigcode-evaluation-harness/output/humaneval_codellama.json
--save_generations_path ./bigcode-evaluation-harness/output/generations_humaneval_codellama.json
Thx for your great work! I'm conducting evaluation under a multi-gpu scenario (2 A100x80G). But it's even slower than that with one single gpu.
Here is the output of
accelerate env
:BTW, the evaluation script is copied from the README file.