Closed dattv closed 1 month ago
Because DLA takes chw16 inputs, which means we need to pad the input to NC/16HW16 first.
Because DLA takes chw16 inputs, which means we need to pad the input to NC/16HW16 first.
@zerollzeng Thank you for your feedback!, however there are some extra information I want to inquires,
No, the 16 means the C channel is padded to 16, So you actually have a 1x16x672x672 tensor, while only the first 3 channels represent the image.
closing since no activity for several month, thanks all!
Hi all.
I'm wondering that input resolution of yolov5 is 3x672x672, but why in https://github.com/NVIDIA-AI-IOT/cuDLA-samples/blob/a2d645b61920fead0cf70c79506518b0a159463c/src/matx_reformat/matx_reformat.cu#L146, the shape of mInput1 tensor was allocated as (1x16x672x672). (in case of using fp16) anybody known why??? @zerollzeng, @mchi-zg