Closed yifliu3 closed 1 year ago
First, the cross entropy loss does not influence the mask generation of novel classes. Secondly, we want to predict high scores of novel classes and low scores of seen classes for novel instances. From my view, the "no-object" textual embeddings are learnable. The cross-entropy loss will make the "no-object" embeddings dissimilar to the seen class embeddings. If a novel instance is matched with "no-object", the features of a novel instance will also be dissimilar to the seen class embeddings, which may help to predict low scores of seen classes for the novel instances.
Thanks a lot for your clear explanation!
Hi, another two problems are:
I see, thanks a lot for your clear clarification!
Hello authors, thank you for this great work!
In my understanding, during the training period, instances with novel classes are also exposed to the network, so the queries of these novel instances will match the "no-object" textual embeddings by the cross entropy loss. In this case, the model may tend to ignore the novel instances during inference and cause severe performance drop. But it seems not a problem in the experiements, so I'm wondering if my understanding has something wrong.