Brummi / MonoRec

Official implementation of the paper: MonoRec: Semi-Supervised Dense Reconstruction in Dynamic Environments from a Single Moving Camera (CVPR 2021)
MIT License
587 stars 85 forks source link

Inference on custom data #4

Closed aditdoshi333 closed 3 years ago

aditdoshi333 commented 3 years ago

Hello,

Thank you for sharing your amazing work. I am just trying to inference on custom RGB data. But I checked in the example that the model is using ['keyframe', 'keyframe_pose', 'keyframe_intrinsics', 'frames', 'poses', 'intrinsics', 'sequence', 'image_id'] for inference. I am just curious to know is it possible to run the inference pipeline just for RGB data? Thank you

nynyg commented 3 years ago

Hi, thanks for your interest in our work!

To answer your question more precisely, may I ask what do you mean by "RGB data" exactly? Do you mean images without poses and intrisincs?

aditdoshi333 commented 3 years ago

Question 1: Simply RGB data (no poses no intrinsics)

Is it possible to run the inference on such data?

Question 2: I want to collect real-world data for inference can you please tell me what all data I need to collect along with RGB frames? Does it need lidar data also? If possible can you brief what you mean by intrinsic, keyframe_pose, and poses?

Brummi commented 3 years ago

Q1: It is possible to run inference on such data as long as you know the intrinsics for the camera (Tutorial by OpenCV). However, you will have to run an odometry system to obtain the pose information (camera rotation and position). We recommend to use the open-source implementation of DSO (we tested this) to compute the poses (this will give the same format for the poses that is used in the implementation of the KITTI Odometry dataset).

Q2: Neither for training, nor for inference you need lidar data. We only use it to evaluate the predicted depth maps.