Klazkin / player-zero

1 stars 0 forks source link

CUDA accelerated model inference #71

Closed Klazkin closed 3 months ago

Klazkin commented 4 months ago

Implement GPU-accelerated model inference using the CUDA build of ORT.

The goal

Implement cuda and validate its impact on performance

Time tracking

Time Estimate: 2 hours 0 minutes Time spent: 1 hours 30 minutes

Resources

DirectML Example https://shalvamist.github.io/onnxruntime/docs/execution-providers/DirectML-ExecutionProvider.html CUDA Example https://github.com/leimao/ONNX-Runtime-Inference/blob/main/src/inference.cpp ORT Execution Providers https://shalvamist.github.io/onnxruntime/docs/execution-providers/