med-air / EndoNeRF

Neural Rendering for Stereo 3D Reconstruction of Deformable Tissues in Robotic Surgery
https://med-air.github.io/EndoNeRF/
196 stars 18 forks source link

About the dataset depth. #26

Open huahangc opened 1 year ago

huahangc commented 1 year ago

How did you get the groundtruth of depth in the dataset?

yuehaowang commented 1 year ago

Rigorously speaking, there is no ground truth of depth. As mentioned in the paper, those depths in the dataset are estimated via STTR-light.

huahangc commented 1 year ago

Thanks for your replying.

darthandvader commented 9 months ago

Can I compare the rendered depth map with thoses in the dataset? How can I obtain the rendered depth map?

huahangc commented 9 months ago

@darthandvader https://github.com/med-air/EndoNeRF/blob/2d4546f58970b7cb3bb2465daee6c36c4f68f3cb/run_endonerf.py#L421 this variable means the 1/depth, i recommend use the 1/disp rather use the depth directly.

junzastar commented 5 months ago

Rigorously speaking, there is no ground truth of depth. As mentioned in the paper, those depths in the dataset are estimated via STTR-light.

Hi, I have a simple question about the GT depth map. Why can't you get the GT depth map directly with binocular video? Is it because the error is too large or something else? OR is it feasible to use this way to obtain GT depth map in this task? Thank you

yuehaowang commented 5 months ago

Rigorously speaking, there is no ground truth of depth. As mentioned in the paper, those depths in the dataset are estimated via STTR-light.

Hi, I have a simple question about the GT depth map. Why can't you get the GT depth map directly with binocular video? Is it because the error is too large or something else? OR is it feasible to use this way to obtain GT depth map in this task? Thank you

I don't fully understand your question but wonder if there is a way to obtain the real "GT depth" from binocular videos? I mean the real depths, not the estimated ones. IMO, the only way is to use depth sensors, which are not equipped with most endoscopes.

junzastar commented 5 months ago

Rigorously speaking, there is no ground truth of depth. As mentioned in the paper, those depths in the dataset are estimated via STTR-light.

Hi, I have a simple question about the GT depth map. Why can't you get the GT depth map directly with binocular video? Is it because the error is too large or something else? OR is it feasible to use this way to obtain GT depth map in this task? Thank you

I don't fully understand your question but wonder if there is a way to obtain the real "GT depth" from binocular videos? I mean the real depths, not the estimated ones. IMO, the only way is to use depth sensors, which are not equipped with most endoscopes.

Thank you for your reply. Yes, the best way is to use the depth sensors. But it is impossible in such scenarios. So, for the binocular video, I mean we can obtain the real depth by stereo matching if we have the camera parameters, right?

yuehaowang commented 5 months ago

Rigorously speaking, there is no ground truth of depth. As mentioned in the paper, those depths in the dataset are estimated via STTR-light.

Hi, I have a simple question about the GT depth map. Why can't you get the GT depth map directly with binocular video? Is it because the error is too large or something else? OR is it feasible to use this way to obtain GT depth map in this task? Thank you

I don't fully understand your question but wonder if there is a way to obtain the real "GT depth" from binocular videos? I mean the real depths, not the estimated ones. IMO, the only way is to use depth sensors, which are not equipped with most endoscopes.

Thank you for your reply. Yes, the best way is to use the depth sensors. But it is impossible in such scenarios. So, for the binocular video, I mean we can obtain the real depth by stereo matching if we have the camera parameters, right?

Stereo matching is still a way to estimate the depth. And it requires correspondence information on the image pairs, which is not available in our case.

junzastar commented 5 months ago

Rigorously speaking, there is no ground truth of depth. As mentioned in the paper, those depths in the dataset are estimated via STTR-light.

Hi, I have a simple question about the GT depth map. Why can't you get the GT depth map directly with binocular video? Is it because the error is too large or something else? OR is it feasible to use this way to obtain GT depth map in this task? Thank you

I don't fully understand your question but wonder if there is a way to obtain the real "GT depth" from binocular videos? I mean the real depths, not the estimated ones. IMO, the only way is to use depth sensors, which are not equipped with most endoscopes.

Thank you for your reply. Yes, the best way is to use the depth sensors. But it is impossible in such scenarios. So, for the binocular video, I mean we can obtain the real depth by stereo matching if we have the camera parameters, right?

Stereo matching is still a way to estimate the depth. And it requires correspondence information on the image pairs, which is not available in our case.

I see, thank you very much for your reply.