TQTQliu / ET-MVSNet

[ICCV 2023] When Epipolar Constraint Meets Non-local Operators in Multi-View Stereo
MIT License
60 stars 1 forks source link

Quantitative results of TNT dataset #23

Open 05063112lcs opened 1 month ago

05063112lcs commented 1 month ago

Hello author, this is the result of my reproduction, it turns out to be my problem, but these final indicators are still a little bit behind your paper, what is the reason for this, is there a way to improve it image

TQTQliu commented 1 month ago

Hello, it is strange that the metric you get is still not good enough, the inermediate part should be around 65.5, advanced part should be around 40.5. I will recently reuse the pre-training weights I provided to run the results and upload the corresponding point cloud.

I'm working on a project right now, so it may take a few days to give you feedback.

05063112lcs commented 1 month ago

Hello, it is strange that the metric you get is still not good enough, the inermediate part should be around 65.5, advanced part should be around 40.5. I will recently reuse the pre-training weights I provided to run the results and upload the corresponding point cloud.

I tested it with my own model, so I guess it's the effect of the virtual environment。

TQTQliu commented 1 month ago

Using our provided pre-trained model, the inermediate part should be around 65.5 and advanced part should be around 40.5. If the results are produced by your own model, it may be that the trained model is not good enough, or it may be an environmental problem.

05063112lcs commented 1 month ago

Using our provided pre-trained model, the inermediate part should be around 65.5 and advanced part should be around 40.5. If the results are produced by your own model, it may be that the trained model is not good enough, or it may be an environmental problem. If it's because the training model isn't good enough, how can I make the results close to yours, I haven't made any changes to the code

TQTQliu commented 1 month ago

First trained on the DTU dataset, then fine-tuned on the BlendedMVS dataset, we choose the model that has trained after 13 or 14 epoches as the final model to be tested.

TQTQliu commented 1 month ago

In addition, for Horse scene, you can use DTU-trained models for testing here, and for other scenes you can use fine-tuned models on blendedmvs.

05063112lcs commented 1 month ago

In addition, for Horse scene, you can use DTU-trained models for testing here, and for other scenes you can use fine-tuned models on blendedmvs.

How many rounds of the model to train as the final model, is this determined by the final training loss, I see that your source code says that it is 15 rounds of training, so can I train a few more rounds to get better results