utiasSTARS / pykitti

Python tools for working with KITTI data.
MIT License
1.14k stars 239 forks source link

point projection gives different results comparing to kitti depth dataset #50

Open liruikangyk opened 5 years ago

liruikangyk commented 5 years ago

In my project I have to downsample the KITTI velodyne lidar data to 16 lines (KITTI depth dataset provides original 64-line projected png files of cam02 and cam03) so I need to (1) downsample the 64-line bin lidar data to 16-line (this has already been done) (2) use the calib parameters to project the point from bin files to 2d pictures of cam02 and cam03 however, I met a problem in step 2: I find that points in my projected pngs look a little bit shifted so I tried to test my projection procedure : project the original 64-line lidar data to cam02 png, and merge my projected image with the KITTI-provided projected image of cam02. (image from dataset is used as green and blue channel, and my own projected image is used as red channel) I can see that most of the points are at the same pixel location, but the points in the very right and very left of the image are shifted from the correct location (especially points of the objects that are near from the camera, as you can see in the picture below, the correctly projected points shows in white color)

20190724112440

my code is here: 20190724113053

trns1997 commented 4 years ago

Hi @liruikangyk may I ask how did you go about downsampling the 64-line bin lidar data from the kitti dataset to 16-line?

hulsmeier commented 3 years ago

Hi @liruikangyk may I ask how did you go about downsampling the 64-line bin lidar data from the kitti dataset to 16-line?

did you ever figure out a way?

EthanZhangYi commented 3 years ago

@liruikangyk Hi, I also meet the same problem. I tried several coordinate transformation methods but the generated depth map is still different from those in proj_depth/velodyne_raw of KITTI Depth Dataset.

Do you figure out the reason for this mismatching? I am also interested in the method you used to down-sample the 64-beam point cloud. Could share some information on it?

Thanks~

minghaohsu410168 commented 3 years ago

@liruikangyk ,

How do you downsample the lidar data from 64 line to 16 lines? Please give me some advice, thank you.