brade31919 / radar_depth

Source code of the IROS 2020 paper "Depth Estimation from Monocular Images and Sparse Radar Data"
MIT License
87 stars 13 forks source link

Release date #2

Closed pjckoch closed 3 years ago

pjckoch commented 3 years ago

Hi @brade31919,

thanks for your contribution and interesting work. I've read your paper and aim to solve a similar task. It's great that you are willing to share your code. Do you already know when you are going to publish it?

Thanks in advance!

Best, Patrick

hz658832 commented 3 years ago

Hello Patrick @pjckoch, me and my student from RWTH Aachen are also working on the same task, developing a low cost radar / camera sensor fusion platform for depth estimation / object detection.

If you are interested, we could share and discuss together. Also with 沅 and Dr. Dai.

Beste Grüße aus Aachen

Haoming

brade31919 commented 3 years ago

Hi @pjckoch and @hz658832,

Sorry I was busy with CVPR 2021. I think I can try to release the code around the mid or end of December 2020.

Thank you @hz658832! I think Dr. Dengxin Dai will definitely be interested in some discussions or possible collaborations.

Sincerely, Juan-Ting Lin

pjckoch commented 3 years ago

Hi @brade31919 ,

thank you for your quick response. Mid or end of December sounds perfect, thank you very much.

@hz658832 I would be happy to discuss the topic and exchange some ideas. I will talk to my supervisor and get back to you.

Best, Patrick

hz658832 commented 3 years ago

Hey Juan-Ting @brade31919,

would you like to release the code? :)

Thx? LG Haoming

brade31919 commented 3 years ago

Hi @hz658832,

I am working on it

Sincerely, Juan-Ting Lin

pjckoch commented 3 years ago

Hi @brade31919,

Sorry for bothering you again. How are you progressing with the code release? Can we expect the code to be released by end of January?

Thanks a lot and best regards, Patrick

brade31919 commented 3 years ago

Hi @pjckoch,

Sorry for the delay on the release date! The whole process takes more time than expected. Currently, I have finished some code re-organization (cleaned some experimental code not related to the main results). I am still discussing with Dr. Dai about the release of the processed data (need to find some storage). Due to some issues of the ETH clusters, some of my backed up checkpoints were deleted. I am now trying to re-train some of them for the release.

We are also considering to release the code to extract processed data from the original nuScenes dataset, but we need some time to check the compatibility. They have some updates in 2020.

Regarding the release date, I think at least the code and trained models can be released by the end of January, but I am not sure about the processed dataset.

Sincerely, Juan-Ting Lin

pjckoch commented 3 years ago

Hi @brade31919,

Ok, I totally understand. Thank you for getting back to me so quickly!

I look forward to the release of your code. In the meantime, would you mind specifying which data augmentation you used in your training? I can't find any information on that in your paper, except for the fact that you downscaled nuScenes images to 450x800 pixels. Was your data augmentation pipeline like that of the Sparse-to-Dense paper? They describe their data augmentation as follows:

C. Data Augmentation We augment the training data in an online manner with random transformations, including

  • Scale: color images are scaled by a random number s∈[1,1.5], and depths are divided by s.
  • Rotation: color and depths are both rotated with a random degree r∈[−5,5].
  • Color Jitter: the brightness, contrast, and saturation of color images are each scaled by ki∈[0.6,1.4].
  • Color Normalization: RGB is normalized through mean subtraction and division by standard deviation.
  • Flips: color and depths are both horizontally flipped with a 50% chance.
brade31919 commented 3 years ago

Hi @pjckoch

Thank you for the understanding!

Regarding the data augmentation info, I used (almost the same as sparse-to-dense):

I used "bilinear interpolation" to resize the images and "nearest neighbor" to resize the depth maps.

pjckoch commented 3 years ago

Hi @brade31919,

this helps a lot, thank you very much!

Best, Patrick

Three last things: 1.)

The parameters I used are (brightness, contrast, saturation, hue) = (0.2, 0.2, 0.2, 0.).

Is it supposed to be a 0.2 for hue as well?

2.): So, no normalization for the input radar, correct?

3.):

Crop: the scaled images and depth maps are cropped back to [450, 800] (this resolution is for nuscenes).

Are you using a center crop or random crop?

brade31919 commented 3 years ago

Hi @pjckoch

  1. It's 0. in my case 😂😂, but yeah I think you should also use maybe 0.2. It's always better if we can have more diverse augmentations.
  2. Nope. However, I think you can try to normalize the depth maps. Most of time, the learning procedure is more stable on normalized data.
  3. Random cropping in training time and no cropping in testing time.
pjckoch commented 3 years ago

Hi @brade31919 ,

Great, thanks for answering all my questions!

brade31919 commented 3 years ago

Hi @pjckoch and @hz658832 ,

I just updated the repo. I tested the installation and training procedures on the cluster I can access, but there might still be some bugs. Let me know if you guys encounter any problem using the code.

Sorry for the delayed release.