bit-bots / YOEO

YouOnlyEncodeOnce - A CNN for Embedded Object Detection and Semantic Segmentation
GNU General Public License v3.0
22 stars 4 forks source link

Bump onnxruntime-gpu from 1.9.0 to 1.10.0 #22

Closed dependabot[bot] closed 2 years ago

dependabot[bot] commented 2 years ago

Bumps onnxruntime-gpu from 1.9.0 to 1.10.0.

Release notes

Sourced from onnxruntime-gpu's releases.

ONNX Runtime v1.10.0

Announcements

  • As noted in the deprecation notice in ORT 1.9, InferenceSession now requires the providers parameters to be set when enabling Execution Providers other than default CPUExecutionProvider. e.g. InferenceSession('model.onnx', providers=['CUDAExecutionProvider'])
  • Python 3.6 support removed for Mac builds. Since 3.6 is end-of-life in December 2021, it will no longer be supported from next release (ORT 1.11) onwards
  • Removed dependency on optional-lite
  • Removed experimental Featurizers code

General

  • Support for plug-in custom thread creation and join functions to enable usage of external threads
  • Optional type support from opset15

Performance

  • Introduced indirect Convolution method for QLinearConv which has symmetrically quantized filter, i.e., filter type is int8 and zero point of filter is 0. The method leverages in-direct buffer instead of memcpy'ing the original data and doesn’t need to compute the sum of each pixel of output image for quantized Conv.
    • X64: new kernels - including avx2, avxvnni, avx512 and avx 512 vnni, for general and depthwise quantized Conv.
    • ARM64: new kernels for depthwise quantized Conv.
  • Tensor shape optimization to avoid allocating heap memory in most cases - #9542
  • Added transpose optimizer to push and cancel transpose ops, significantly improving perf for models requiring layout transformation

API

  • Python
    • Following through on the deprecation notice in ORT 1.9, InferenceSession now requires the providers parameters to be set when enabling Execution Providers other than default CPUExecutionProvider. e.g. InferenceSession('model.onnx', providers=['CUDAExecutionProvider'])
  • C/C++
    • New API to query CUDA stream to launch a custom kernel for scenarios where custom ops compiled into shared libraries need implicit synchronization with ORT CUDA kernels - #9141
    • Updated Invalid -> OrtInvalidAllocator
    • Updated every item in OrtCudnnConvAlgoSearch to a safer global name
  • WinML
    • New APIs to create OrtValues from Windows platform specific ID3D12Resources by exposing DirectML Execution Provider specific APIs. These APIs allow DML to extend the C-API and provide EP specific extensions.
      • OrtSessionOptionsAppendExecutionProviderEx_DML
      • DmlCreateGPUAllocationFromD3DResource
      • DmlFreeGPUAllocation
      • DmlGetD3D12ResourceFromAllocation
    • Bug fix: LearningModel::LoadFromFilePath in UWP apps

Packages

  • Added Mac M1 Universal2 build support for a single binary that runs natively on both Apple silicon and Intel-based Macs. These are included in the official packages and can also be built using "-arch arm64 -arch x86_64"
  • Windows C API Symbols are now uploaded to Microsoft symbol server
  • Nuget
    • Support for ARM64 Linux C#
  • Python GPU package now includes both TensorRT and CUDA EPs. Note: EPs need to be explicitly registered to ensure the correct provider is used. e.g. InferenceSession('model.onnx', providers=['TensorrtExecutionProvider', 'CUDAExecutionProvider']). Please also ensure you have appropriate TensorRT dependencies and CUDA dependencies installed.

Execution Providers

  • TensorRT EP
    • Python GPU release packages now include support for TensorRT 8.0. Enable TensorrtExecutionProvider by explicitly setting providers parameter when creating an InferenceSession. e.g. InferenceSession('model.onnx', providers=['TensorrtExecutionProvider', 'CUDAExecutionProvider'])
    • Published quantized BERT model example
  • OpenVINO EP
    • Add support for OpenVINO 2021.4.x
    • Auto Plugin support

... (truncated)

Commits


Dependabot compatibility score

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
dependabot[bot] commented 2 years ago

Looks like onnxruntime-gpu is up-to-date now, so this is no longer needed.