Closed UditSinghParihar closed 4 years ago
Hello. The format that you are describing seems correct to me. You can check if the ground-truth correspondences are correct by passing the argument --plot
to the training script - this will plot the images with dense correspondences as well as the detection scores.
--plot
. It seems my training data format is correct as I can see the diagonal type correspondences which are expected in opposite view images. I also added a small code to visualize the correspondences in your '--plot' code block and the output looks like this for training data. Hi Sir,
0.0005
, which might be implying that learning is taking place, but during inference on training dataset itself, I am getting wrong correspondences. Dense ground-truth correspondences, inference on training dataset image, and loss log can be seen here. d2_ots.pth
weights with a learning rate of 0.0003
for 10 epochs. skimage.feature.match_descriptors
without RANSAC, with RANSAC I getting very few correspondences (7-8 correspondences). Thanks
The loss value is way too low (the least I was able to get on real-world datasets was 0.3-0.4 if I recall correctly). I suspect that the training data is too hard and the network converges to a trivial solution (such as not detecting anything or having the same descriptor across the board). Moreover, I don't think that convolutions are adequate for opposite viewpoints due to the lack of strong rotation invariance.
Hello Sir,
pose1
: World wrt to camera 1.d2_tf.pth
(have to add 1e-5 at 5 places to avoid NaN values for loss). But correspondences obtained are incorrect.