Open UmangBDhanani opened 1 year ago
Hello @xingyizhou. I found that the 3d ground truth bounding box label estimation during the training gives a little bit different answers for the location and rotation of the 3d bounding box compared to the ground truth label. And this is the reason for the false projection of the bounding box in the image. Why it is giving the false estimation of location and rotation. Could you point out the potential issue ? Thank you.
Hello @xingyizhou. Thank you for your wonderful work and I am following it for my fusion project. However I am stuck in a issue. I was trying to train a network and used debug=2 for the initial visualization of the data just to see if the data was feed properly, as I am using my custom kitti format data. However there is a error in the 3d projection as seen in the figure
I also tried the debug mode in the convert kitti to coco.py file and there it ran perfectly with well aligned boxes in the image. Also I noticed that for the projection of 3d bounding box from convert_kitti_to_coco.py file, the location, dimension and rotation of the annotation are directly taken into calculation. But in the training file, the ground truth data (gt_det) are saved in the metadata and is post processed before projection into image. How is this organized or why the ground truth data is postprocessed as ground truth projection during training is faulty as seen in the image above. Thank you and hopefully you got the point what I am trying to convey.