Open kevalshah90 opened 2 months ago
Hey @kevalshah90, sorry to hear that you're experiencing issues! Did you install from source (if yes, which commit/branch) or from pip (which version?)?
@Titus-von-Koeller I am importing BitsAndBytesConfig
from Transformers version 4.41.2
System Info
I am using
bitsandbytes
quantization to loadmistral-7b
onNVIDIA T4
gpu. I loaded the model with the quantized configuration, however, I keep getting an runtime error related to device. I am ensured that the model and inputs are oncuda
.Reproduction
Error:
Expected behavior
Output logits.