Closed sayakpaul closed 3 months ago
@jerryzh168 could you advice?
quantize_
optionally takes in a filter_fn https://github.com/pytorch/ao/blob/main/torchao/quantization/quant_api.py#L267
this does look weird because we already filter linear by default, let me try to repro
@sayakpaul python3 benchmark_pixart.py --compile --quantization fp6
runs for me, can you paste the repro command?
Strange python3 benchmark_pixart.py --compile --quantization fp6
is failing for me with the same error. LMK try with the latest torchao
.
Working with with the latest torchao
.
Error:
The PixArt model has an
nn.Parameter
: https://github.com/huggingface/diffusers/blob/a57a7af45cbef004c38e2a294a6457f7f3574e5d/src/diffusers/models/transformers/pixart_transformer_2d.py#L174Is there way to provide some filtering to
fp6_llm_weight_only()
so that it doesn't pick upnn.Parameter
s?