Open Chay16 opened 5 years ago
You are right. Here the output form is consistent with the KITTI label files (KITTI rectified camera coordinate ). Please refer to the introduction of KITTI training labels for details.
OK thanks for your answer ! Another point, is there a way to disable the calibration step easily to train PointRCNN on other annotated point clouds ?
You need to check the codes manually to update the codes where the calibrations are used. Most of them are in the dataset related codes, maybe you should write a new dataloader for your own point clouds.
I will write a new dataloader, it will be more appropriate I think. Another point, why are you subsampling the point cloud into 16,384 points as mentioned in the paper? Is it only for a memory point of view ?
The PointNet++ needs a fixed number of input points and then I subsample 16384 points as the input.
this post has helped to clarify my doubts too. thanks.
Hi sshoashuai, Thaanks you for releasing your code and the pretrained model. I'm a bit confuse when trying to plot the 3D predicted bounding boxes. I'm guessing that the output format (the first line of the prediction for 000001.txt) encodes the bounding box as follows:
Car -1 -1 -1.7723 753.6551 163.8756 814.0421 204.0525 1.5462 (height) 1.6426 (width) 3.9845 (length) 7.0502 (x) 1.2010 (y) 29.7814 (z) -1.5398 (theta) 1.6346
with y representing the elevation and the XZ plane representing the ground plane.
Am i right ? Regards,