This issue is for the purpose of tracking all the ONNX Frontend Requirements, categorized into following parts (listed in the highest to lowest priority order):
Models:
1.Vision Models : 76 int8 quantized CNN Models used for vision
2.P0/P1 CNN Models : 487 int8 IPU CNN Models
3.Protected Models : 23+ int8 Models in the protected area (ISV or benchmark protected models)
4.Hugging Face CNN Models : 900 fp32 hugging face CNN models
5.Hugging Face non-CNN Models : Around 110 hugging face non-CNN models
IREE EP Support:
Run a sampled set of tests (20 tests) from above by loading in IREE Execution Provider, IREE EP on Windows 11 and Ubuntu 22.04.
OP Support and Tests:
OP Suport ONNX OP support to lower to linalg from ONNX
Note:
Just a gentle request that if you're working on fixing any of these failures then please go and check the other kind of failures related to the issue you're working on. Your fix for one kind of failure might fix the other kind of failure too.
For Models, if iree-compile passes, count as pass. We will require inference to pass once @saienduri has done his work.
Owners, kindly add clear steps to reproduce failures and allow ability for contributors to take up a unique issue and work on fix to have more folks join in for this quality push. Great start! Let's do it.
This issue is for the purpose of tracking all the ONNX Frontend Requirements, categorized into following parts (listed in the highest to lowest priority order):
Models: 1.Vision Models : 76 int8 quantized CNN Models used for vision 2.P0/P1 CNN Models : 487 int8 IPU CNN Models 3.Protected Models : 23+ int8 Models in the protected area (ISV or benchmark protected models) 4.Hugging Face CNN Models : 900 fp32 hugging face CNN models 5.Hugging Face non-CNN Models : Around 110 hugging face non-CNN models
IREE EP Support:
OP Support and Tests:
Latest Status (Passing/Total)
Note: Just a gentle request that if you're working on fixing any of these failures then please go and check the other kind of failures related to the issue you're working on. Your fix for one kind of failure might fix the other kind of failure too. For Models, if iree-compile passes, count as pass. We will require inference to pass once @saienduri has done his work.
Lit Tests of torch-mlir (ONNX to linalg lowering)
Owner: @renxida
The Onnx lowering Lit Tests are tracked through https://github.com/nod-ai/SHARK-Turbine/issues/450.
Steps to reproduce:
View the op name from the tracker and then take out the lit test corresponding to that op in a seperate file, and run:
Torch Op E2E Tests of torch-mlir
Owner: @vivekkhandelwal1 @zjgarvey
The Torch-MLIR Op E2E Tests are tracked through https://github.com/nod-ai/SHARK-Turbine/issues/549
Steps to reproduce:
Take out the E2E test from the tracker and run:
ONNX Op Shark_TestSuite/iree_tests
Owner: @jinchen62 @Shukla-Gaurav
The Shark_TestSuite IREE Tests - ONNX Op tests are tracked by Compile time Tests - https://github.com/nod-ai/SHARK-Turbine/issues/563 Runtime Tests - https://github.com/nod-ai/SHARK-Turbine/issues/583
Steps to reproduce:
To run the test, please follow: build venv following here and run
Models Shark_TestSuite/e2eshark
Owner: @PhaneeshB @saienduri @AmosLewis
The E2EShark Model Tests are tracked through https://github.com/nod-ai/SHARK-Turbine/issues/566
Steps to reproduce:
First, follow setup instructions at https://github.com/nod-ai/SHARK-TestSuite/tree/main/e2eshark. No need to do the Turbine setup part as we are looking at onnx mode. Then, run this command (HF_TOKEN needed for llama, gemma model):