NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
pytorch-quantization development has been discontinued in favor of Model Optimizer since TRT 10.2. Please try that if possible.
cc @nzmora-nvidia for opinion on the raised issue.
Description
trt10.5 pytorch-quantization has compile bug.
https://github.com/NVIDIA/TensorRT/blob/release/10.5/tools/pytorch-quantization/src/tensor_quant_gpu.cu#L28-L37 define two macro
AT_DISPATCH_CASE_FLOATING_TYPES
andAT_DISPATCH_FLOATING_TYPES
but in https://github.com/NVIDIA/TensorRT/blob/release/10.5/tools/pytorch-quantization/src/tensor_quant_gpu.cu#L18
#include <ATen/ATen.h> --> #include <ATen/Dispatch.h>
--> has already defined these two macros.I check torch1.13 and torch2.4.1, both the same case.
two macros duplicate definition. @moraxu need use
before
#define
in tensor_quant_gpu.cuEnvironment
TensorRT Version:10.5
NVIDIA GPU:rtx2000
NVIDIA Driver Version:
CUDA Version:11.8
CUDNN Version:9.1
Operating System:
Python Version (if applicable):3.8
PyTorch Version (if applicable):1.13 or 2.4.1