Open h8c2 opened 4 months ago
I have found out that the floaters are introduced by improper parameters of the occupancy grid, however, I think the background is still not satisfactory.
Hi, thank you for your question.
I think this may result from (1) a lack of point cloud data for objects in the background scene box and (2) a limited background capacity.
In the latter case, you may set a larger value for pipeline.model.bg-color-grid-max-res
or pipeline.model.bg-color-log2-hashmap-size
to enrich the background's capacity. You may also try fixing the number of importance samples along a ray, i.e., set pipeline.model.pdf-samples-fixed-ratio
to 1
.
For the first case, as the model relies on lidar initialization for efficient reconstruction, regions, where density grids are not properly initialized (i.e., some objects without lidar observation), may be skipped during occupancy sampling, leading to less satisfactory results. We augment the background point cloud (i.e., $P_{bg}$ in our paper) to alleviate this issue, but there are still some cases of failure. If you could obtain the point cloud data for background objects, the results would be better. It would also be helpful if you could visualize the foreground (red) scene box and the point cloud data, in a 3D viewer like this:
Then, you may enlarge the foreground box accordingly to include some regions initially located in the background portion for better visual performance.
I have applied your method to my dataset. From what I understand, you scale the pose in advance, whereas I use the original scale of the pose. This means I need to set the scene_box.aabb correctly (e.g., aabb in real scale and set scale=1.0) and center the pose. Is there something else I might have missed? I noticed that the background region's result is poor, while the foreground region looks good.
I'm wondering if there might be an issue with my data processing or if the background's capacity is limiting the performance..