The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
I was able to finetune SAM on custom image segmentation dataset which trained the mask decoder and got very high IoU. Then I used the same model weights to test video predictor but didn't see improvement. I assume it has to do with memory encoder. How can I finetune the whole model? Didn't want to reinvent the wheel if someone has done it.
I was able to finetune SAM on custom image segmentation dataset which trained the mask decoder and got very high IoU. Then I used the same model weights to test video predictor but didn't see improvement. I assume it has to do with memory encoder. How can I finetune the whole model? Didn't want to reinvent the wheel if someone has done it.