-
Hello,
Nvidia MLPerf suggests to use [TensorRT](https://github.com/NVIDIA/TensorRT) framework for a performant inference deployment. For DLRM (DL based Recommendation Systems) inference on GPU, I h…
-
I try to achive the result in Nvidia Jetson Xavier NX. After setup up the environment, I got the fllowing error message:
```
Makefile:236: *** MLPerf Inference v1.1 code requires NVIDIA Driver Ve…
zhr01 updated
2 years ago
-
When looking at Inference/Edge/Open Division results, e.g.
https://mlperf.org/inference-results-0-7/
It seems unlikely that accuracy was maintained across the different line items (e.g. going fro…
-
Have references to Python2 and python2.7 which runs into multiple issues.
Once I resolve that, I ran into famous issue of cannot find cublas_v2.h file although the file exists in 4 different loca…
-
For reproducibility we would request NVIDIA to switch the `BASE_IMAGE` repo links from internal Gitlab links to publicly available repo links. In the `Makefile.docker`, the `BASE_IMAGE` URL is from an…
-
Hi there,
I have ran retinanet on **jetson Orin** and found the latency is greater than the official one.
My result is **118112979.00** ns which is 0.16 of the official results **19378310.00** ns.…
-
In the MLPerf Inference v2.1 round, Qualcomm and their partners submitted a number of RetinaNet results, which the Review Committee eventually accepted to the Open division under the Preview category.…
-
Hi, the rules show that min duration is 600 for all workloads (I was looking at datacenter) while it shall be 60 for most of them.
https://github.com/mlcommons/inference_policies/blob/master/inferenc…
-
-
I'm trying to load a resnet50 model with quantize_int8 using calibration data, but getting the following error: `LLVM ERROR: Expected to find GEMM, convolution, or attention op, and didn't`
The erro…