TRI-ML / vidar

Other
573 stars 67 forks source link

How to learn camera intrinsics for DepthFormer? #34

Closed tommiekerssies closed 1 year ago

tommiekerssies commented 1 year ago

I want to train on videos from different cameras (e.g. random YouTube videos). How do I get the camera intrinsics? The paper states they can be learned. Is there code to do this? Can I use the self-calibration code for that perhaps? Or will performance still be reasonable without learning the intrinsics?

1005452649 commented 1 year ago

hello,can I have a look at your code? @tommiekerssies

tommiekerssies commented 1 year ago

I don't have any code yet

1005452649 commented 1 year ago

How do I get the camera intrinsics? @VitorGuizilini-TRI @IgorVasiljevic-TRI

VitorGuizilini-TRI commented 1 year ago

For DepthFormer we do not learn camera intrinsics, so you need to provide those alongside the images. We have some other works that include camera self-calibration, you could perhaps run those first, get the predicted intrinsics, and feed them to DepthFormer. That would be an interesting thing to try!

1005452649 commented 1 year ago

Where is the code for camera self-calibration? @VitorGuizilini-TRI

VitorGuizilini-TRI commented 1 year ago

You can self-calibrate on KITTI using this script, let me know if it works out for you: https://github.com/TRI-ML/vidar/blob/main/configs/papers/selfcalib/ucm_kitti.yaml