Open natke opened 5 months ago
From @ashD
@yufenglee I tried calling Utils.GetCurrentGpuDeviceId() in C# using the Microsoft.ML.OnnxRuntimeGenAI.DirectML 0.2.0-rc6 nuget
and it threw a Microsoft.ML.OnnxRuntimeGenAI.OnnxRuntimeGenAIException: 'CUDA and/or ROCM execution provider is either not enabled or not available.'
Same error on Utils.SetCurrentGpuDeviceId(1)
@natke is this still an issue?
Just ran the same code with DirectML nuget 0.3.0-rc2. Same error.
Reproduced with 0.3.0
This is somewhat expected with DML package. The implementation in ORT only supports CUDA and ROCm:
If we want to add DML support, we need to add it inside ORT itself.
See also: #566
Dear team, any update when this bug will be fixed? This prevents anyone with multiple GPU's from selecting the best one for inference.
@baijumeswani Any update on this issue?
@PatriceVignola is it possible to set the gpu device id for dml? Can we expose it through onnxruntime-genai?
From the last question in #430