-
-
Hi! I wonder from which piece of code this result was derived.
| Top-1 | Top-5
-- | -- | --
DeiT-T | 72.2 | 91.1
+BatchFormerV2 | 72.7 | 91.5
DeiT-S | 79.8 | 95.0
+BatchFormerV2 | 80.4 | 95.2
…
-
Hi, thanks for your great work!
I'm confused about the setting of ablation experiment of DeiT below:
![image](https://user-images.githubusercontent.com/100334147/233563641-0a00103c-f21f-4db8-afb3-aa…
-
Nice work!
**bash scripts/pretrain/visda/run_visda.sh deit_base
bash scripts/uda/visda/run_visda.sh deit_base**
**What different between 'pretrain' and 'uda'?**
-
제 configuration을 script하기 전까지 issue에서 관리하도록 하겠습니다...ㅎㅎ
# 2nd trial 현재 리더보드 4위
-
**Describe the bug**
Model I am using (trocr):
The problem arises when using:
* [ ] the official example scripts: (give details below)
python pic_inference.py
A clear and concise descriptio…
-
- https://arxiv.org/abs/2012.12877
- 2020
近年、純粋に注意力に基づくニューラルネットワークが、画像分類などの画像理解タスクに対応することが示された。
しかし、これらの視覚変換器は、高価なインフラを使って何億枚もの画像を使って事前に学習されているため、その採用には限界があります。
本研究では、Imagenetのみで学習することにより、競争力のある…
e4exp updated
3 years ago
-
Hello, I have a question about the transformations in the MiniViT paper.
I could find the first transformation (implemented in the MiniAttention class) in the code:
https://github.com/microsoft/Cr…
gudrb updated
2 months ago
-
Hi. thanks for your great job.
I trained deit-base with your QAT method using quant_train.py.
while training, it showed good eval TOP1 accuracy over 82 in early epoches.
but when I load the sav…
-
Hi,
In section 4 of the Lumina-Next paper, you demonstrate that the model can be adapted for image classification. Could you provide guidance on how to reproduce the model architecture using the co…