Closed demmlert closed 3 years ago
I have the same problem with you. And my result is not better as yours, what is the size of you pretrained model of the YCB dataset, thank you!
The size of the trained models is 449MB
Have you found the reason, my results show that it seems overfit.
I have achieved some better results (not as good as the paper claims). But that were just some lucky exceptions. My best guess is that there is a fundamental problem with the loss function on symmetric objects. Or even with the keypoint approach since keypoints are not well defined on symmetric objects. In the end a low loss does not strictly imply a good ADD-S score. Therefore you need to save a bunch of non optimal checkpoints and test them all. If you get lucky you can match the results...
Oh, the result is too strange, since most of the classes achieve comparable or even better results as in our paper but the two clamp classes are too low to be normal. There must be something wrong. The result in the paper is from the model trained by our internal architecture. But I've checked that code in this repo can achieve comparable results before I released it. Did you inference our released model and get normal results? I suspect that maybe there is something wrong with the data or some setting of these two classes, you can visualize data of these two classes by python3 -m datasets.ycb.ycb_dataset
what is the size of you pretrained model of the YCB dataset, thank you!
The pretrained model we released on one drive only saves the model_state
without the optim_state
, so it's much smaller. Related code to modify is at here
My best guess is that there is a fundamental problem with the loss function on symmetric objects. Or even with the keypoint approach since keypoints are not well defined on symmetric objects. In the end, a low loss does not strictly imply a good ADD-S score. Therefore you need to save a bunch of nonoptimal checkpoints and test them all. If you get lucky you can match the results```
Improving the design of loss function on symmetric objects can get even better results but our current setting should get good results as in our paper and is a lot better than the direct regression approaches, as is shown by our ablation study. For the checkpoint to test, the saved the best checkpoint selected from the validation usually gets good results.
Have you found the reason, my results show that it seems overfit.
For the overfitting concerned, We've also tested our pre-trained model on YCB-Video to a new camera without finetuning and compared DenseFusion in our robotic grasping demo, which still shows quite better results. One example is as follows:
@aiai84 I also found it overfit . Do you have experience with tuning the hyperparameters?
Hello,
I am currently trying to reproduce your results. But I can't achieve the results from your paper. Especially the results for the large and extra large clamp are not even close. I trained with your original settings and I also trained with many different settings. You can see the results of your original code below. Did you do something different than the code in the GitHub repository? How did you achieve these results? I would greatly appreciate any insights you can provide.
Cheers, Tobias