argoverse / av2-api

Argoverse 2: Next generation datasets for self-driving perception and forecasting.
https://argoverse.github.io/user-guide/
MIT License
317 stars 75 forks source link

Annotations for High Definition Map #214

Open EchoQiHeng opened 1 year ago

EchoQiHeng commented 1 year ago

Hello, I am currently working on High Definition Map generation based on the AV2 dataset, and my confusion lies in the annotation of High Definition Map. As shown in the following figure, map elements within the sensor range (point cloud map area) can be easily labeled based on the movement trajectory of the vehicle, while the problem is how to label other areas (not observed by LiDAR). I want to do some High Definition Map automatic generation tasks. Do I need to manually select point and line elements within the range of the point cloud map? Figure_7 图片

James-Hays commented 1 year ago

Hi!

The HD map elements paired with each scenario are intentionally sampled from a wide radius. Some of those HD map elements are probably not observable from lidar or from camera. So it isn't a realistic task to try and generate the entire HD map for a scenario given only its sensor data. Argo generated the maps by looking at lots of data that isn't part of this scenario.

If you're trying to establish a benchmark for map generation, you'd probably want some rule such as "only train or evaluate on map elements within 25 meters of the ego vehicle" or maybe even include some explicit notion of visibility e.g. "only train or evaluate on map elements within 50 meters of the ego vehicle and that have a least 1 lidar return within 1 meter" or something like that.

We haven't established an HD map synthesis benchmark. We did, however, establish an HD map verification benchmark with the Trust, but Verify dataset: https://www.argoverse.org/av2.html#mapchange-link

EchoQiHeng commented 1 year ago

Hi!

The HD map elements paired with each scenario are intentionally sampled from a wide radius. Some of those HD map elements are probably not observable from lidar or from camera. So it isn't a realistic task to try and generate the entire HD map for a scenario given only its sensor data. Argo generated the maps by looking at lots of data that isn't part of this scenario.

If you're trying to establish a benchmark for map generation, you'd probably want some rule such as "only train or evaluate on map elements within 25 meters of the ego vehicle" or maybe even include some explicit notion of visibility e.g. "only train or evaluate on map elements within 50 meters of the ego vehicle and that have a least 1 lidar return within 1 meter" or something like that.

We haven't established an HD map synthesis benchmark. We did, however, establish an HD map verification benchmark with the Trust, but Verify dataset: https://www.argoverse.org/av2.html#mapchange-link

Thank you very much for your reply. As you mentioned, I am committed to building high definition maps (online/offline). I will follow your suggestions and manually set some rules to establish map benchmarks.

I am also interested in the TbV dataset, but when I visualized the map, I noticed some annotation issues. As shown in the following figure, for an intersection in the scene, the purple area represents a pedestrian crossing, but the pedestrian crossing of the left road is not included in the GT. Therefore, I visualized the point cloud (reflection intensity) and RGB point cloud, , and it is clear that the left road includes a pedestrian crossing.

May I ask if this is an omission in the labeling or my understanding is incorrect? Looking forward to your reply!

Figure_1 intensity rgb