Closed mattmdjaga closed 6 months ago
@mattmdjaga can you change the PR title to reflect we are putting CLIP and OWL here?
I added a link to my YOLO mention. However I linked https://github.com/ultralytics/ultralytics instead of the YOLO paper as I think it's more relevant to the topic. Not sure if everyone else agrees with that?
@mattmdjaga you can add YOLO paper and add ultralytics in references. If you can commit rest of my suggestions I'll give another review and we can merge 😊
I think we can get this merged @mattmdjaga and if you feel like it you can add a small PyTorch implementation of CLIP in another PR or this PR, what do you think?
Yeh I'm down to add a PyTroch CLIP implementation and it can be in another PR as I might first need to work on other parts of this chapter which haven't been completed yet.
@mattmdjaga sure, no worries! this one is good as is
This section is part of showcasing of CLIP relatives and it covers CLIP and OWL-ViT, an open vocabulary object detection model. The aim is to briefly introduce the model, show how to use it and provide resources to find more about the models.
The whole section structure: Models: