-
### 1. Error Loading `state_dict` for ViT
While attempting to train a pretrained model using the `LaMed/script/pretrained.sh` script with a self-trained vision model specified as:
```bash
-…
-
have anyone who training from scratch, not use pre-trained weight of ViT-B?
Can author or anyone released the training log of train from scratch?
the author said it takes more epochs to training fro…
-
Hi @KaimingHe,
thank you for your wonderful work and initiative to open-sourcing it.
I have been banging my head against this for a month now and any help would be deeply appreciated!
I am tryi…
-
Thanks for your work and Detailed answer in issues.
I am reproducing the ViT B-16 in Tensorflow based on your Paper and answers. (In this issue, I only deal with original ViT paper)
But I just reach…
-
Can you share the img2text model.pth. I want to use this work for inference on other datasets.
-
Any idea how to solve this? I am clueless. Thanks!
(midas-py310) C:\Midas\MiDaS>python run.py --model_type dpt_next_vit_large_384 --input_path "C:\Midas\MiDaS\input" --output_path "C:\Midas\MiDaS\o…
-
Thank you for your GREAT work.
The default ViT size of paper and released weights is ViT-g, which is too large for researchers and users with limited resourses. Could you please release the weights…
-
Hi Author,
Thank you for sharing this project and for your kindness for answering my previous questions. I have some of questions want to ask about training:
1. What is your default learning rat…
-
Hi, I have read your tutorials about ViT and finetuning on custom dataset, and based on one of your links [https://blog.roboflow.com/how-to-train-vision-transformer/](https://blog.roboflow.com/how-to-…
-
Thank you for publishing this code !!
I want to run it and cite it in an academic paper.
Can you please add a license? Preferably MIT license if possible.
link to the procedure:
https://docs.git…