Closed AmosLewis closed 1 month ago
I thought that linalg.quantized_batch_matmul
should all be decomposed to a sequence of linalg op? @pashu123 could you take a look?
@hanhanW @pdhirajkumarprasad I don't see this issue on the latest branch.
@AmosLewis Could you double check on the main branch?
@pashu123 the issue gone with the latest branch.
What happened?
When compiler the public onnxstorage quantized onnx/models/ConvNeXt_vaiq_int8 model, it failed with error
./dispatch/module_torch_jit_dispatch_13.mlir:9:6: error: 'func.func' op exceeded stack allocation limit of 32768 bytes for function. Got 401408 bytes
.I dumped the failed dispatch by
iree-compile --iree-input-demote-i64-to-i32 --iree-hal-target-backends=llvm-cpu ConvNeXt_vaiq_int8.default.onnx.linalg.mlir > ConvNeXt_vaiq_int8.default.vmfb --iree-hal-dump-executable-sources-to=./dispatch
:module_torch_jit_dispatch_13.mlir:
and more failed dispatch: module_torch_jit_dispatch_15.mlir, module_torch_jit_dispatch_23.mlir, module_torch_jit_dispatch_25.mlir, module_torch_jit_dispatch_33.mlir.
Steps to reproduce your issue
iree-compile --iree-input-demote-i64-to-i32 --iree-hal-target-backends=llvm-cpu ./dispatch/module_torch_jit_dispatch_13.mlir > o.vmfb
What component(s) does this issue relate to?
Compiler
Version information
candidate-20240809.980
Additional context
No response