-
-
I have been running the Swin_Transformer and VMamba models on the same A800 GPU, using same batch sizes and the COCO2017 detection dataset.
However, I've observed that VMamba performs at least 5 tim…
-
I had a doubt about layers in Swin Transformer. As it is mentioned in the architecture of Swin-T that there are 2, 2, 6, 2 layers at stage 1,2,3 and 4.
What does it mean by 2 layers at 1st stage an…
-
Hello, as we know, the basic ability of `swin-transformer` is better than `vit`. Can you provide `swin-transformer` in `oml/models` (including pre-train weight) ?
upd:
[Contribution instruction…
-
> Repeat the process with the Swin Transformer. Detect the right seed, but hallucinate a fourth one.
![Image](https://github.com/ai-cfia/nachet-model/assets/96267006/cdfcad57-36ae-4e23-b628-cd836a0f7…
-
Related to https://github.com/keras-team/keras-cv/issues/668, this is another transformer based model which we'd like to be supported. For simplicity, focus on only classification task, having in mind…
-
Hi,
I found model parameter `use_cuda` already removed in the commit: 00711a20b7a8dfb6d6d1d892216997daeeee6288, but in `usage_examples/swinT_example.py` still use `use_cuda`. So the user may got e…
-
Thanks for your great work! I try to leverage the STViT-R-Swin-S in the semantic segmentation task according to Sec 6.3 in the paper. I use the pretrained checkpoint of STViT-R-Swin-S in #5 with the T…
-
-
从提供的swin_transformer权重开始训练感觉和没用预训练权重一样,前几个epoch精度基本都是0,up也是这样的吗?