yhw-yhw / PVAMVSNet

This official repository is established for Pyramid Multi-view Stereo Net with Self-adaptive View Aggregation (ECCV2020)
77 stars 11 forks source link

where is sigmoid? #3

Open kwea123 opened 4 years ago

kwea123 commented 4 years ago

you use volumegatelight here https://github.com/yhw-yhw/PVAMVSNet/blob/05084abac2512a367a81558785fb8d7e83b4ed0f/models/vamvsnet.py#L455-L459

which is defined as https://github.com/yhw-yhw/PVAMVSNet/blob/05084abac2512a367a81558785fb8d7e83b4ed0f/models/submodule.py#L75-L80

but there is no sigmoid. Generally the attention is something between 0 and 1, and in your paper you also have sigmoid function in the supplementary material table 2. However I cannot find any sigmoid in the code; if the attention is without sigmoid in my opinion it is strange. Does it also work without sigmoid? Or did I miss anything?

yhw-yhw commented 4 years ago

Yes, it did work, we also did the experiment with using sigmoid, but replacing it with ReLU as in our code and our provided model, the performance is slightly better than the model with using sigmoid. We guess that in our Voxel-wise view aggregation, we consider that only using weight map to enhance the feature map is better than highlight and suppress. By the way, we will provide the 2D Pixel-wise view aggregation in the future and update our code base with new results.

kwea123 commented 4 years ago

Yes, the current code misses pixelVA; but it seems that it doesn't contribute much? There is + VoxelVA + CTF + MMP in Table 2, but no + VoxelVA + PixelVA + CTF + MMP.

Also in Figure 9., you have the same color for Baseline and +VoxelVA, if you can fix that it'll be good.

yhw-yhw commented 4 years ago

Our VoxelVA and PixelVA are two different self-adaptive view aggregation for the same target, in our experiments, VoxelVA is better than PixelVA in 3DCNN but PixelVA still have improvements on the performance, it is useful for the situation that you do not have 3D feature volume like in R-MVSNet.