bigcode-project / bigcode-evaluation-harness

A framework for the evaluation of autoregressive code generation language models.
Apache License 2.0
702 stars 180 forks source link

Potentially extra slow inference when using LoRA adapter #192

Open sadaisystems opened 5 months ago

sadaisystems commented 5 months ago

Hello, everybody. Tried HumanEval benchmark on my custom Mistral tune today, but getting weird warning:

UserWarning: Input type into Linear4bit is torch.float16, but bnb_4bit_compute_dtype=torch.float32 (default). This will lead to slow inference or training speed.
  warnings.warn(f'Input type into Linear4bit is torch.float16, but bnb_4bit_compute_dtype=torch.float32 (default). This will lead to slow inference or training speed.')

Dont know how to fix this, any ideas?

My command to run the benchmark:

accelerate launch  main.py \
  --model {model_name} \
  --peft_model {peft_model_path} \
  --load_in_4bit \
  --max_length_generation 512 \
  --tasks humaneval \
  --temperature 0.2 \
  --precision bf16 \
  --n_samples 200 \
  --batch_size 32 \
  --allow_code_execution \
  --limit 25 
sadaisystems commented 5 months ago

Seems to only occur when --load_in_4bit passed.