hisfog / SfMNeXt-Impl

[AAAI 2024] Official implementation of "SQLdepth: Generalizable Self-Supervised Fine-Structured Monocular Depth Estimation", and more.
MIT License
91 stars 12 forks source link

The training results differ from the paper by 33% #36

Open bizheyandebing opened 7 months ago

bizheyandebing commented 7 months ago

I have completely trained your model, and the difference with the model you provided is close to 33%. Could you please help me check my training files.

image

image Below is the content of my training file:

--data_path /home/ubuntu/ubuntu_jixie/temp/kitti-raw --dataset kitti --model_name res_088 --backbone resnet_lite --height 192 --width 640 --batch_size 16 --num_epochs 25 --scheduler_step_size 15 --num_layers 50 --num_features 256 --model_dim 32 --patch_size 16 --dim_out 64 --query_nums 64 --min_depth 0.001 --max_depth 80.0 --eval_mono --post_process

bizheyandebing commented 7 months ago

Yes, I use res50 as the backbone network, If I set --load_weight_floder, how should its parameters be set?

bizheyandebing commented 7 months ago

I think this setting should have no effect, because I found that the relevant code of the train.py file (lines 137-138h) has been logged out.

Moreover, there is no mention of additional pre-training in the paper.

hisfog commented 7 months ago

Apologies for the delayed response, For reproducing results on KITTI,please DO NOT use the latest code release. Instead, you can kindly utilize the following version by

git checkout 6a1e997f97caef8de080bb2873f71cfbad9a8047

which is consistent with the implementation of paper SQLdepth, without any additional modifications. You can refer to https://github.com/hisfog/SfMNeXt-Impl/issues/26#issuecomment-2006322584, Thanks!