Open pdhirajkumarprasad opened 2 weeks ago
We really need to figure out how to untangle pads from torch exported models.
This issue is fixed by https://github.com/llvm/llvm-project/pull/113551.
Testing the failing models with pad sizes folding patch alone: https://github.com/llvm/torch-mlir/pull/3813
TOTAL TESTS = 41 | Stage | # Passing | % of Total | % of Attempted |
---|---|---|---|---|
Setup | 41 | 100.0% | 100.0% | |
IREE Compilation | 23 | 56.1% | 56.1% | |
Gold Inference | 23 | 56.1% | 100.0% | |
IREE Inference Invocation | 22 | 53.7% | 95.7% | |
Inference Comparison (PASS) | 17 | 41.5% | 77.3% |
TOTAL TESTS = 41 | Stage | # Failed at Stage | % of Total |
---|---|---|---|
Setup | 0 | 0.0% | |
IREE Compilation | 18 | 43.9% | |
Gold Inference | 0 | 0.0% | |
IREE Inference Invocation | 1 | 2.4% | |
Inference Comparison | 5 | 12.2% |
Test was run with the following arguments: Namespace(device='local-task', backend='llvm-cpu', iree_compile_args=None, mode='cl-onnx-iree', torchtolinalg=True, stages=None, skip_stages=None, benchmark=False, load_inputs=False, groups='all', test_filter=None, testsfile='sample.txt', tolerance=None, verbose=True, rundirectory='test-run', no_artifacts=False, cleanup='0', report=True, report_file='sample.md', get_metadata=True)
Test | Exit Status | Mean Benchmark Time (ms) | Notes |
---|---|---|---|
edgenext_base | Numerics | None | |
edgenext_small | Numerics | None | |
edgenext_small_rw | PASS | None | |
edgenext_x_small | Numerics | None | |
edgenext_xx_small | Numerics | None | |
maxvit_base_tf_224.in1k | compilation | None | |
maxvit_base_tf_384.in1k | compilation | None | |
maxvit_base_tf_384.in21k_ft_in1k | compilation | None | |
maxvit_base_tf_512.in1k | compilation | None | |
maxvit_base_tf_512.in21k_ft_in1k | compilation | None | |
maxvit_large_tf_224.in1k | compilation | None | |
maxvit_large_tf_384.in1k | compilation | None | |
maxvit_large_tf_384.in21k_ft_in1k | compilation | None | |
maxvit_large_tf_512.in1k | compilation | None | |
maxvit_large_tf_512.in21k_ft_in1k | compilation | None | |
maxvit_small_tf_224.in1k | compilation | None | |
maxvit_small_tf_384.in1k | compilation | None | |
maxvit_small_tf_512.in1k | compilation | None | |
maxvit_tiny_tf_224.in1k | compilation | None | |
maxvit_tiny_tf_384.in1k | compilation | None | |
maxvit_tiny_tf_512.in1k | compilation | None | |
maxvit_xlarge_tf_384.in21k_ft_in1k | compilation | None | |
maxvit_xlarge_tf_512.in21k_ft_in1k | compilation | None | |
model--codegen-350M-mono--Salesforce | Numerics | None | |
model--CodeGen-350M-Multi--xhyi | compiled_inference | None | |
model--deberta-italian-question-answering--osiria | PASS | None | |
model--deberta-v3-base-qa-en--LLukas22 | PASS | None | |
model--deberta-v3-base-squad2--deepset | PASS | None | |
model--deberta-v3-base-squad2--navteca | PASS | None | |
model--deberta-v3-basesst2all-train--SetFit | PASS | None | |
model--deberta-v3-large-squad2--deepset | PASS | None | |
model--deberta-v3-large-squad2--sjrhuschlee | PASS | None | |
model--deberta-v3-xsmall-squad2--nlpconnect | PASS | None | |
model--deberta_squadnewsqa--sophiebottani | PASS | None | |
model--mdeberta-v3-base-squad2--sjrhuschlee | PASS | None | |
model--microsoft-deberta-v3-large_ner_conll2003--Gladiator | PASS | None | |
model--microsoft_deberta-base_squad--Palak | PASS | None | |
model--microsoft_deberta-large_squad--Palak | PASS | None | |
model--outputs--ankitkupadhyay | PASS | None | |
model--reward-model-deberta-v3-large-v2--OpenAssistant | PASS | None | |
tnt_s_patch16_224 | PASS | None |
These are the models with redundant tests removed:
edgenext_base edgenext_small edgenext_small_rw edgenext_x_small edgenext_xx_small tnt_s_patch16_224 model--CodeGen-350M-Multi--xhyi model--deberta-italian-question-answering--osiria model--deberta-v3-base-qa-en--LLukas22 model--deberta-v3-basesst2all-train--SetFit model--deberta-v3-large-squad2--deepset model--deberta-v3-xsmall-squad2--nlpconnect model--mdeberta-v3-base-squad2--sjrhuschlee model--microsoft-deberta-v3-large_ner_conll2003--Gladiator model--microsoft_deberta-base_squad--Palak model--microsoft_deberta-large_squad--Palak model--outputs--ankitkupadhyay model--reward-model-deberta-v3-large-v2--OpenAssistant maxvit_base_tf_224.in1k maxvit_base_tf_384.in1k maxvit_base_tf_512.in1k maxvit_large_tf_224.in1k maxvit_large_tf_384.in1k maxvit_large_tf_512.in1k maxvit_small_tf_224.in1k maxvit_small_tf_384.in1k maxvit_small_tf_512.in1k maxvit_tiny_tf_224.in1k maxvit_tiny_tf_384.in1k maxvit_tiny_tf_512.in1k maxvit_xlarge_tf_384.in21k_ft_in1k maxvit_xlarge_tf_512.in21k_ft_in1k
for given IR
command : iree-compile --iree-hal-target-backends=llvm-cpu model.torch_onnx.mlir