Closed ugoleone closed 3 weeks ago
The function get_rgbd_point_cloud is the recommended and friendly interface for backprojecting rgbd data through a camera.
By looking at the file https://github.com/facebookresearch/pytorch3d/blob/main/pytorch3d/implicitron/dataset/llff_dataset_map_provider.py and the functions it calls you should have an example of how to load the llff cameras. You may well be doing this correctly anyway.
I would like to render new views of a scene, starting from a single image, its depth map, and camera poses.
I have an image taken from LLFF dataset, the metric depth map relative to that image (in meters) and 5 camera poses provided by the LLFF dataset in the
poses_bounds.npy
file (including the pose of the camera from which the image was taken). All files are attached.Poses, as documented, are stored in the
poses_bounds.npy
file in this way:What I'm trying to do
poses_bounds.npy
)What I got
The images I'm obtaining are upside down, I suppose this is due to a wrong RT matrix conversion/multiplication
My code
I read here and here methods to convert poses for LLFF dataset, but using them I got worse results.
I believe it is just a matter of the pose matrices and the use of back projection functions. I would appreciate any help on this, as I have not found a tutorial in the documentation for such a case.