-
RT,没有安装CUDA的机器,设置USE_CUDA=OFF, BUILD_ALL_EXAMPLES = OFF,因为仍然还是会编example,编译时报找不到头文件错误,具体:
```
In file included from /root/installs/libdeepvac/examples/src/test_resnet_benchmark.cpp:11:
/usr/libtorch…
-
**Description**
When running Triton container on Mac M3, calling DALI model using Python BLS with async results in CUDA runtime error, but all models are running on cpu only. The error is as follows:…
-
Hi, I'm trying to build the pybind11 extension mentioned under onemkl_gemv example DPCTL build with CUDA:
https://github.com/IntelPython/dpctl/tree/master/examples/pybind11/onemkl_gemv
Example men…
-
### Describe the issue
Hello developers:
I followed guide `docs/ORT_Use_Triton_Kernel.md` and wanted to use triton kernel in ONNX runtime. But I encountered a error.
**test script:**
```bash
…
-
### Describe the issue
MatMul is expected to produce a valid result when it is multiplying matrices with inner dimension equal to zero.
For example, operands of shapes {16,0} x {0, 16} should produc…
-
### Describe the issue
When I'm using Stable Diffusion Auto1111 WebUI with ControlNet IP-Adapter and ip-adapter-faceid-plus v2 created by h94 on hugginface, I keep getting the following error message…
-
### Checklist
- [ ] 1. I have searched related issues but cannot get the expected help.
- [ ] 2. The bug has not been fixed in the latest version.
- [ ] 3. Please note that if the bug-related issue y…
-
### Describe the issue
Below is an error I'm getting trying to use the exported model in onnx format,
1 - Training and infering on GPU using pytorch works fine,
2 - After getting this error I ens…
-
While experimenting with PyTorch's DTensor functionality in my single node "cpu" environment, I've encountered an unexpected limitation. Specifically, in the provided code example, I'm attempting to u…
-
### Describe the issue
We have converted the translation LLM 7B model to ONNX format using Optimum Hugging Face and then quantized it to 8-bit quantization with Dynamic quantization technique. Ho…