-
Hi. GroupViT is an excellent work.
I wonder if the groupViT has any Open-vocabulary characteristic, likes, if we want to segement a cat, we don't input the lable of "cat" , instead, we input some wo…
-
# Describe the feature
**Motivation**
Recently, a series of open-vocabulary semantic segmentation works have emerged. For example: Scaling Open-Vocabulary Image Segmentation with Image-Level Label…
-
### Model description
MaskCLIP represents a transformative step in the realm of open-vocabulary universal image segmentation. Built upon the robust foundation of pre-trained CLIP models, it negates…
-
Hi,
Thanks for your work, I found it very interesting.
I was wondering whether it is possible to get more per-pixel features using your pre-trained model. Currently, using the provided example scr…
-
Dear Author,
I hope this message finds you well.
I have some concerns regarding the experimental setup and results presented in your paper, which I hope you can clarify.
**1. Dataset Usage Is…
-
@dingjiansw101 Hi Jian, thanks for your great work! I am wondering did you happen to test your trained coco-stuff model directly on the ADE-20K dataset? Because in the concurrent works, like [1][2], t…
-
[Unsupervised Open-Vocabulary Object Localization in Videos](http://openaccess.thecvf.com/content/ICCV2023/html/Fan_Unsupervised_Open-Vocabulary_Object_Localization_in_Videos_ICCV_2023_paper.html) fro…
-
Hi,
Thank you for sharing your impressive work!
I got confused about Table 2: How are the open vocabulary segmentation metrics calculated?
Also, could you please explain how Osprey outputs the…
-
Hi,
In the demos it shows, that you can also extract the type of the segment (e.g a door, a human etc.) can it be extracted from the model or do we need an additional library for that?
Cheers
-
Hi,
I recently came across your article and found it very inspiring. However, I wanted to ask if you've noticed the growing trend in 3D open vocabulary semantic scene understanding. These works hav…