sshaoshuai / PointRCNN

PointRCNN: 3D Object Proposal Generation and Detection from Point Cloud, CVPR 2019.
MIT License
1.7k stars 425 forks source link

Output Format #27

Open Chay16 opened 5 years ago

Chay16 commented 5 years ago

Hi sshoashuai, Thaanks you for releasing your code and the pretrained model. I'm a bit confuse when trying to plot the 3D predicted bounding boxes. I'm guessing that the output format (the first line of the prediction for 000001.txt) encodes the bounding box as follows:

Car -1 -1 -1.7723 753.6551 163.8756 814.0421 204.0525 1.5462 (height) 1.6426 (width) 3.9845 (length) 7.0502 (x) 1.2010 (y) 29.7814 (z) -1.5398 (theta) 1.6346

with y representing the elevation and the XZ plane representing the ground plane.

Am i right ? Regards,

sshaoshuai commented 5 years ago

You are right. Here the output form is consistent with the KITTI label files (KITTI rectified camera coordinate ). Please refer to the introduction of KITTI training labels for details.

Chay16 commented 5 years ago

OK thanks for your answer ! Another point, is there a way to disable the calibration step easily to train PointRCNN on other annotated point clouds ?

sshaoshuai commented 5 years ago

You need to check the codes manually to update the codes where the calibrations are used. Most of them are in the dataset related codes, maybe you should write a new dataloader for your own point clouds.

Chay16 commented 5 years ago

I will write a new dataloader, it will be more appropriate I think. Another point, why are you subsampling the point cloud into 16,384 points as mentioned in the paper? Is it only for a memory point of view ?

sshaoshuai commented 5 years ago

The PointNet++ needs a fixed number of input points and then I subsample 16384 points as the input.

uranuscc commented 3 years ago

this post has helped to clarify my doubts too. thanks.