NetEase-FuXi / EETQ

Easy and Efficient Quantization for Transformers
Apache License 2.0
157 stars 12 forks source link

Qlora with eetq is quite slow #17

Open hjh0119 opened 1 month ago

hjh0119 commented 1 month ago

The training process is quite slow, whereas using 8-bit hqq speeds it up by more than tenfold. Is this normal? Or have I missed any code?

import torch
from transformers import EetqConfig, AutoModelForCausalLM

config = EetqConfig("int8")

# from transformers import HqqConfig
# HqqConfig(nbits=8)

model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1", quantization_config=config)

# train...
dtlzhuangz commented 1 month ago

Sorry for your trouble. The backward propagation process of EETQ has not been fully optimized.

hjh0119 commented 1 month ago

got it. Is the optimization in schedule?

dtlzhuangz commented 1 month ago

Not yet.