DerryHub / BEVFormer_tensorrt

BEVFormer inference on TensorRT, including INT8 Quantization and Custom TensorRT Plugins (float/half/half2/int8).
Apache License 2.0
425 stars 69 forks source link

support Flash Multi-Head Attention Plugin (FP32/FP16/INT8) #60

Closed DerryHub closed 1 year ago