ifnspaml / SGDepth

[ECCV 2020] Self-Supervised Monocular Depth Estimation: Solving the Dynamic Object Problem by Semantic Guidance
MIT License
200 stars 26 forks source link

question about valid #28

Open xuchen-dev opened 2 years ago

xuchen-dev commented 2 years ago

i think train the model successful,i dont change anything in the code the result looks good image it is about 10 epoch result

but the a1 a2 a3 rmse and so on is far from the paper privde image

is there something wrong in my depth gt i just use the code to get the gt depth from lidar point image image the gt depth look like this image

is there something wrong in this way ,very hopeful for your helping i am just a green hand students

xuchen-dev commented 2 years ago

image and my rmse

xuchen-dev commented 2 years ago

i download you provide sgddepth full trained model and eval it the result like this 1632883741(1) there must something wrong with my depth gt , could you help me,i will be very grateful

klingner commented 2 years ago

Hi,

up to now, I have never observed such an error. How did you prepare the data? Did the download_kitti.py script run successfully? Or did you prepare the data differently in any way?

xuchen-dev commented 2 years ago

thank for your replying ,i didnt use the all the download_kitti.py, because i have downloaded it i prepare my data like this

image

klingner commented 2 years ago

Hi, so what I observe in the output of the depth evaluation is that your scaling ratio is rather unusual (0.023). I usuall have scaling ratios in the order of 20-30 (deviation of a factor of $10^3$), so I would agree that something regarding you depth gt is wrong. You can download the depth gt for the image kitti/Depth/2011_09_26/2011_09_26_drive_0001_sync/image_02/data/0000000000.png at this link:

https://drive.google.com/file/d/1QRudmxq5ENl4tfXuO2A_n381_VYRDtTX/view?usp=sharing

Maybe you can compare it to your depth gt and see, if they match up.

xuchen-dev commented 2 years ago

i wonder the ratio is ratio = (median_gt / median_pred).item() i wonder the lower the ratio the closer the value between gt and pred why normal vlaue is 20 - 30 , i just use the generate_depth_from_velo in the kitti_download to get depth gt , i dont know i got this problem image image

yes t'here is something wrong in my depth gt , i just use the generate_depth_from_velo in the kitti_download to get depth gt , i dont know why i got this problem

xuchen-dev commented 2 years ago

and i wonder why your gt is png in the kitti_download code is

image is it the way i get the depth gt is false ?

klingner commented 2 years ago

Hi, so I just set up the repository from scratch again and also tried out the download script. I can still exactly reproduce the results. What I also noted is that you store the images as jpg and not as png. In the unchanged download script the images are stored as png. I would recommend to try out the unchanged download script and then run the evaluation again.

huhaoyue commented 2 years ago

Hi, I want to run this program now, but I don't know how to arrange the data set. I don't know if I can get your guidance

klingner commented 2 years ago

Hi, to what extent did you follow the provided instructions in the README?

huhaoyue commented 2 years ago

Hi, to what extent did you follow the provided instructions in the README?

I downloaded the dataset from the official website, and then followed the the latter part of download_kitti.py you gave to arrange the structure of the dataset, but the following error is reported when running tranin.py. 1640160339(1)

After debugging, I found that some of its samples are empty. Is this a problem with my dataset? 1640160379(1)

huhaoyue commented 2 years ago

Hi, to what extent did you follow the provided instructions in the README?

1640161588(1) When I debug to this place, I can see that it is Nonetype. I don't quite understand where the problem is. If it is a dataset problem, how can I check where it appears? 1640161617(1)

klingner commented 2 years ago

So for me it appears as if you are trying to load an image that does not exist at the specified path location, so likely the arrangement of your dataset is not yet correct. I would check in the file dataloader/pt_data_loader/basedataset.py in the function read_image_file(self, filepath), which filepath is loaded and if this path actually exist on your machine. Maybe this will give you a hint, if something has gone wrong when arranging the dataset.

huhaoyue commented 2 years ago

This error occurs in the middle of the loop. Not all the data loaded is None, so I don't think it's a path problem. This is the amount of data I loaded. Is it correct? 1640165639(1)

klingner commented 2 years ago

Yes, the amount is correct, but it is not calculated from the data on disk but from the downloaded json-files, so I wouldn't expect this to be wrong. The None-error is typical for a non-existing image where the path is wrong.

huhaoyue commented 2 years ago

是的,这个数字是正确的,但它不是根据磁盘上的数据计算的,而是从下载的json文件中计算出来的,所以我不会认为这是错误的。对于路径错误的不存在的图像,"无"错误是典型的。

I checked that the image is missing here. Does the color means the RGB image corresponding to segmentation map? 1640185407(1)

Does the None of this mean that the images captured by the camera in leftImg8bit corresponding to the segmentation maps in gtFine are not loaded?

The structure of my cityscapes dataset is as follows: --data --cityscapes --gtFine --train --aachen --bochum --bremen --cologne .... --test --val --leftImag8bit_trainvaltest --train --val --test --kitti --Depth --Raw_data --kitti_kitti_split --.json --kitti_zhou_split --.json

huhaoyue commented 2 years ago

Thank you very much for your reply. I found the problem. The reason for the error is that I found above. As long as the name of the dataset is changed, just change the leftimage8bit_ trainvaltest in cityscapes to leftimage8bit. This program can run normally.