-
Hi @davide-coccomini,
When I read your paper, I had a question about Cross ViT architecture. Can you help me to answer?
In your convolutional Cross ViT architecture, which used convolutional archi…
-
First of all Kudos on creating this amazing library ❤️.
I think all of us in self supervised learning community have heard of [DINO](https://github.com/facebookresearch/dino). For the past couple …
-
[The ConTeXt interface](http://mirrors.ctan.org/macros/generic/markdown/markdown.html#context) would benefit from a proper user interface such as [the one LaTeX has](http://mirrors.ctan.org/macros/gen…
-
How to get the pretrained weights in custom datasets? Not ACDC or Synapse. Thanks, Looking forward to your reply.
@Dootmaan
-
@jongwook Thanks for this great work!
I am trying to train CLIP VIT B/32 from scratch, but cannot get a higher score on imagenet versus CLIP resnet-50. May I ask what initialization you use in tra…
-
Hello Dear Authors,
I am trying to replicate your results for ViT benchmark model on COCO detection. I was able to successfully run the inference but I am getting a CUDA out of memory error during…
-
hi,大家好,非常高兴的告诉大家,百度飞桨论文复现赛第五期已经开始了,本次**论文复现赛**共将有100篇的经典&前沿论文供大家复现,以及新增了**工程落地赛**,详细信息可以参考[AI Studio](https://aistudio.baidu.com/aistudio/competition/detail/126/0/introduction),大家是否已经迫不及待了呢~
为了帮助…
-
Thanks for sharing your code. Can you provide the hyper-parameters (e.g. learning rate, weight decay, optimizer type, training epochs) of ViT-B/16 training from scratch on ImageNet dataset? Many thank…
-
My log says
```
404 Client Error: Not Found for url: https://huggingface.co/sentence-transformers/clip-ViT-B-32/resolve/main/config.json
----------------------------------------------------------…
-
In section 3 of paper 'Augmenting Convolutional networks with attention-based aggregation':
···
We can easily specialize the attention maps per class by replacing the CLS vector with a k × d matrix,…