Open yushengsu-thu opened 2 weeks ago
Hi! I can't reproduce this (1 GPU). Are you using the latest transformers
?
hf (pretrained=meta-llama/Llama-3.2-3B), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 256 |
Tasks | Version | Filter | n-shot | Metric | Value | Stderr | ||
---|---|---|---|---|---|---|---|---|---|
gsm8k_cot | 3 | flexible-extract | 8 | exact_match | ↑ | 0.2980 | ± | 0.0126 | |
strict-match | 8 | exact_match | ↑ | 0.2828 | ± | 0.0124 |
Hello, I use the latest and v0.4.3 version of lm_eval and I find the weird phenomena on
llama-3.2-3B
The following is my script:[llama-3.2-3B] Nodes=1, GPUs=8, llama
accelerate 1.0.1