Hi there, just arriving at kinfu_remake Party from learning (late) about Kinect Fusion-> PCL kinFu-> here
As I understand these processes rely on RGB feature detection and ICP for pose estimation. is that correct? I'm very interested in the overall approach used to reconstruct a fused meshes in realitime, but (have; and) would like to use an external tracking system to provide an absolute device pose to this process.
Is doing this feasible? Or is the code too tightly written to consider providing a (known) submillimeter accurate pose truth.
Hi there, just arriving at kinfu_remake Party from learning (late) about Kinect Fusion-> PCL kinFu-> here
As I understand these processes rely on RGB feature detection and ICP for pose estimation. is that correct? I'm very interested in the overall approach used to reconstruct a fused meshes in realitime, but (have; and) would like to use an external tracking system to provide an absolute device pose to this process.
Is doing this feasible? Or is the code too tightly written to consider providing a (known) submillimeter accurate pose truth.