When I run the code "CUDA_VISIBLE_DEVICES=3 TRANSFORMERS_OFFLINE=1 lm_eval --model hf --model_args pretrained=/public/MountData/yaolu/LLM_pretrained/LLAMA2_7B/,trust_remote_code=True --tasks mmlu,cmmlu --device cuda:0 --batch_size auto --num_fewshot 0" for llama2-7b, it reports "RecursionError: maximum recursion depth exceeded while calling a Python object". However, it works well on llama3-8b and baichuan2-7b? So how can I fix this problem.
When I run the code "CUDA_VISIBLE_DEVICES=3 TRANSFORMERS_OFFLINE=1 lm_eval --model hf --model_args pretrained=/public/MountData/yaolu/LLM_pretrained/LLAMA2_7B/,trust_remote_code=True --tasks mmlu,cmmlu --device cuda:0 --batch_size auto --num_fewshot 0" for llama2-7b, it reports "RecursionError: maximum recursion depth exceeded while calling a Python object". However, it works well on llama3-8b and baichuan2-7b? So how can I fix this problem.