Open tzuchiehTT opened 5 years ago
For cameras with different intrinsic parameters than the KITTI cameras, there is a slightly more complicated image resizing process than simple cropping. In particular, we need to resize the image in a way such that the resized image has the same effective intrinsics as KITTI.
Hi Fangchang, I have the same question. Why can't we just use our own intrinsic parameters instead of resizing images? Is it related to some hyperparameters needed to tune?
@XiaotaoGuo It might or might not work well with a different set of intrinsics, but there is simply no guarantee that the trained network would transfer directly to this new set of intrinsics (and image sizes). My suggestion is to keep the test images as close to the train images as possible.
Thanks! What if we use our own dataset to train the network and test with it?
What if we use our own dataset to train the network and test with it?
Then there is no need for any resizing
Hi @MelvinTT, Did you manage to work this out? I am fighting similar problems with my own dataset, generated with a VLP32 and a Zed camera.
What if we use our own dataset to train the network and test with it?
Then there is no need for any resizing
Hi, When we use 'resize' operation, how to deal with the sparse depth? Could it be correct to just resize the sparse depth input like the rgb images, although it is sparse? Thanks!
Hi Fangchang,
Our lab is currently working on a project which requires generating depth maps from our vlp-16 lidar and camera setting. Your work looks great as the depth map solution. Since we got different size images as input, I think what we need to do to use this network is (1) read in our own calibration information (K) and (2) crop input images as (width, high) both multiples of 16 (since we got errors when going through decode layers with some other sizes), is that right?
We've tested with a rather small dataset (only ~700 frames) and got results like the figure showing below. We are wondering if the dataset is too small or the depth info from vlp-16 is too sparse since the results remain clear projected lines. It would be great if you have any suggestions, thanks!