-
Hi, I want to ask that whether the code of classification parts is completed yet. Seems like the code now is only for testing, but I want to implement full training (30 epochs in your paper) to reprod…
-
Hello guys.
Have you successfully reproduced the performance of LeViT?
I've attempted to train the LeViT, which seems to be among the top-performing ViT models. But I failed to reproduced the result…
-
Excuse me, do you have any plan to release codes or instructions for pretraining?
-
_torch.stack(template_imgs).float().div(255)_ on line 103 in /MixSort/yolox/mixsort_oc_tracker/mixformer.py
```
template_imgs = normalize(
torch.stack(template_imgs).float().div(255),
self.cfg.DAT…
-
Hi @imisra, @rohitgirdhar and rest of 3DETR team
3DETR is quite interesting and it wold be great to make it more visible to the rest of the Machine Learning Ecosystem!
Would you be interested i…
-
Hi,
I have following 3 questions and would be really grateful if anyone could provide some insights:
1. While pertaining the model on the text lines extracted from the PDFs and synthetic data, do …
-
Hi!
Let's bring the documentation to all the Japanese-speaking community 🌐
Who would want to translate? Please follow the 🤗 [TRANSLATING guide](https://github.com/huggingface/transformers/blob/…
-
Hello, I would like to ask if the warmup strategy is not used, but the keep rate is directly set to the target value, will the experimental results differ greatly?
-
- https://arxiv.org/abs/2105.02723
- https://github.com/lukemelas/do-you-even-need-attention
- 2021
視覚変換器の注目層をフィードフォワード層に置き換えてみたところ、ImageNetでもうまく機能することがわかりました。
画像分類などの視覚タスクにおいてビジョントランスフォーマーが優れ…
e4exp updated
3 years ago
-
Using `is_flash_attn_available` is deprecated and will be removed in v4.38. Please use `is_flash_attn_2_available` instead.
Traceback (most recent call last):
File "/opt/ai/test/inference_example_…