Open thehasnism opened 3 years ago
Hi @thehasnism, Thank you for the questions.
The attr
from cam_rgb_points
are [lidar_intensity, r, g, b]
, the lidar intensity
are a part of lidar data so we keep them in the vertex features. As for the cam_rgb_points.xyz
, they are lidar points in the camera coordinate frame. We train the model and predict the labels in the camera coordinate frame.
The global coordinates distinguish each individual point. When the object translates for distance, the global coordinates of its points changes while the relative coordinates can remain invariant. So if we use global coordinates, the network needs to handle objects at all possible locations. But if we use relative coordinates, the network sees the same object regardless of the object's location. Hope it helps. Thank you.
Hi! First of all, I'm really impressed by your work and I have a question about your code.
input_v = cam_rgb_points.attr[:, [0]]
this was in the training file so what is the camera point and why we are using this as input initial vertex features?
Another question is, what happens if we use global coordinates instead of relative coordinates for the edge feature extraction?
Thank you