avishkarsaha / translating-images-into-maps

Official PyTorch code for 'Translating Images Into Maps' ICRA 2022 (Outstanding Paper Award)
Other
400 stars 49 forks source link

Question about dataset configuration #33

Open lithium770 opened 1 year ago

lithium770 commented 1 year ago

Hi, thank you for sharing code! I'm very insterested in this great work. When I read the code, I am confused about dataset configuration:

  1. It seems you adopt data only from CAM_FRONT when training with mini nuscenes dataset, is it the same when using full nuscenes dataset?
  2. In code of https://github.com/tom-roddick/mono-semantic-maps, they contain all 6 cameras to generate monocular data. Do you implement their code in your dataset when doing comparation?

Looking forward to your reply, thanks in advance!

QixingJiang commented 1 year ago

The related articles are not aligned, such as "Understanding Bird’s-Eye View of Road Semantics using an Onboard Camera". Although the training data used in the article is front view, the results of the training and validation are compared with the results of the surround view training in the PON paper, which is a serious academic misconduct. This should be clearly stated in the paper.