-
Hello @lucidrains ,
I use vit-transform for spesific data.Image size is 320x320 and number of classes equal to 2. I set parameters for my dataset and it reached %64.5 test accuracy.Have you any …
-
Hey, thanks for your great work I love it! :) A quick question - in your repo for the Vision Transformer [https://github.com/lucidrains/vit-pytorch] there is a module called `Attention`. Can I simply …
-
We are trying to apply this method on a **medical dataset**, and have about **70K images** (224 res) for **5 classes**. However, our training doesn't converge (we tried a range of learning rates e.g. …
-
Thanks for your great work!
These days I try to compress MobileVit, and reduce the FLOPs and parameters, but the inference time almost doesn't change.
Only when I remove some MV2 blocks, the inf…
-
Hello, author, your vit code is very concise and excellent. I found that the pre-training weight provided by "timm" is very inconsistent with your code. The key and value are not in line with your cod…
-
# 🌟 New model addition
My teammates and I (including @ice-americano) would like to use efficient self attention methods such as Linformer, Performer and Nystromformer
## Model description
The…
-
# 🐛 Bug
This is a minor bug (AFAICT) but the name of the registered nodes (attention, pe, ...) are listed as required in the hydra nodes when "retrieved" from the config store.
## Command
## …
-
Hi, thanks for sharing the implementation, could you pls share some reproduction results, possibly on some benchmarks?
-
**Describe the bug**
Audio-Webui does not install the requirements properly, precisely on audiolm, saying it failed to install.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'audio-…
-
Congrats on your great work!
I am verifying your method on vision tasks and have a small concern on the influence of the "conv_kernel_size" of the 2D group-convolution in your task and I find that …