triton-inference-server / onnxruntime_backend

The Triton backend for the ONNX Runtime.
BSD 3-Clause "New" or "Revised" License
123 stars 54 forks source link

Cannot build r22.03 onnxruntime_backend with tensorrt #125

Open ZJU-lishuang opened 2 years ago

ZJU-lishuang commented 2 years ago

Description I was unable to build the onnxruntime_backend with OpenVino for Triton Inference Server r22.03 using compatible ONNXRuntime and tensorrt versions (from Triton Inference Server compatibility matrix).

Triton Information r22.03

To Reproduce follow the readme in onnxruntime_backend

cmake \
-DCMAKE_INSTALL_PREFIX:PATH=`pwd`/install \
-DTRITON_BUILD_CUDNN_HOME='/usr/lib/x86_64-linux-gnu/' \
-DTRITON_BUILD_ONNXRUNTIME_VERSION=1.10.0 -DTRITON_BUILD_CONTAINER_VERSION=22.03 -DTRITON_ENABLE_ONNXRUNTIME_TENSORRT=ON \
-DTRITON_BACKEND_REPO_TAG=r22.03 -DTRITON_CORE_REPO_TAG=r22.03 -DTRITON_COMMON_REPO_TAG=r22.03 ..

And will see the error. /usr/include/x86_64-linux-gnu/NvInferRuntimeCommon.h:56:10: fatal error: cuda_runtime_api.h: No such file or directory yidongbangong20220705111957

yidongbangong20220705112023

Expected behavior ONNX Backend Should build and run without issues.

ZJU-lishuang commented 2 years ago
-- 
-- ******** Summary ********
--   CMake version             : 3.21.0
--   CMake command             : /workspace/cmake-3.21.0-linux-x86_64/bin/cmake
--   System                    : Linux
--   C++ compiler              : /usr/bin/c++
--   C++ compiler version      : 9.4.0
--   CXX flags                 :  -ffunction-sections -fdata-sections -DCPUINFO_SUPPORTED -Wnon-virtual-dtor
--   Build type                : Release
--   Compile definitions       : EIGEN_MPL2_ONLY;PLATFORM_POSIX;__STDC_FORMAT_MACROS
--   CMAKE_PREFIX_PATH         : 
--   CMAKE_INSTALL_PREFIX      : /usr/local
--   CMAKE_MODULE_PATH         : /workspace/onnxruntime/cmake/external
-- 
--   ONNX version              : 1.10.1
--   ONNX NAMESPACE            : onnx
--   ONNX_USE_LITE_PROTO       : OFF
--   USE_PROTOBUF_SHARED_LIBS  : OFF
--   Protobuf_USE_STATIC_LIBS  : ON
--   ONNX_DISABLE_EXCEPTIONS   : OFF
--   ONNX_WERROR               : OFF
--   ONNX_BUILD_TESTS          : OFF
--   ONNX_BUILD_BENCHMARKS     : OFF
--   ONNXIFI_DUMMY_BACKEND     : OFF
--   ONNXIFI_ENABLE_EXT        : OFF
-- 
--   Protobuf compiler         : 
--   Protobuf includes         : 
--   Protobuf libraries        : 
--   BUILD_ONNX_PYTHON         : OFF
-- Looking for strtof_l
-- Looking for strtof_l - found
-- Looking for strtoull_l
-- Looking for strtoull_l - found
-- The CUDA compiler identification is NVIDIA 11.6.112
-- Detecting CUDA compiler ABI info
-- Detecting CUDA compiler ABI info - done
-- Check for working CUDA compiler: /usr/local/cuda/bin/nvcc - skipped
-- Detecting CUDA compile features
-- Detecting CUDA compile features - done
-- CMAKE_CUDA_COMPILER_VERSION: 11.6.112
-- Looking for clock_gettime in rt
-- Looking for clock_gettime in rt - found
-- Found TensorRT headers at /usr/include/x86_64-linux-gnu
-- Find TensorRT libs at /usr/lib/x86_64-linux-gnu/libnvinfer.so;/usr/lib/x86_64-linux-gnu/libnvinfer_plugin.so
-- Found TENSORRT: /usr/include/x86_64-linux-gnu  
-- The ASM compiler identification is GNU
-- Found assembler: /usr/bin/cc
-- Configuring done
-- Generating done
ZJU-lishuang commented 2 years ago

onnxruntime