-
Issue converting the pytorch model to onnx with the attached script
![image](https://github.com/isl-org/MiDaS/assets/33397127/d50b72ac-54c5-47b0-9b8f-336fc04ec1bc)
[pt_to_onnx.py.gz](https://gith…
-
I am trying to fine-tuning the BEiT3ForVQAV2 model based on the pretrained parameters.
I downloaded the pretrained parameters from this link:
https://conversationhub.blob.core.windows.net/beit-sh…
-
Hi, this ran for me fine yesterday, but today gives the following error. Any ideas on how to solve it? I am using timm 0.6.13. Thank you for any help!
```
Initialize
Device: cuda
/usr/local/lib/…
-
**Describe the bug**
The encoder accuracy values are not stable throughout training as it skyrockets and dips.
**Reproduction**
1. What command or script did you run?
```CUDA_VISIBLE_DEVICES=…
-
Hi!
Thank you for this great repo.
In the Table 4 of your paper you show ablation studies with a ViT-S (Small & 1x384x6).
Is it possible to have access to those pretrained weights or to have the …
-
Error loading script: depthmap.py
Traceback (most recent call last):
File "D:\stable-diffusion-webui\modules\scripts.py", line 229, in load_scripts
script_module = script_loading.load_module(…
-
- https://arxiv.org/abs/2106.08254
- 2021
自己教師付き視覚表現モデルBEiT(Bidirectional Encoder representation from Image Transformersの略)を紹介する。
自然言語処理分野で開発されたBERTに倣い、ビジョントランスフォーマーを事前学習するためのマスク付き画像モデリングタスクを提案す…
e4exp updated
3 years ago
-
Hi
I tried to inference the segmentation model(mask2former_beit_adapter_large_896_80k_cityscapes_ss) and got some weird point.
(1)
Run image_demo.py as
CUDA_VISIBLE_DEVICES=1 python image_demo…
-
## タイトル: EDADepth:単眼深度推定のための強化データ拡張手法
## リンク: https://arxiv.org/abs/2409.06183
## 概要:
拡散モデルは、テキストから画像を生成する機能により、最近では深度推定などの視覚認識タスクで注目されています。しかし、高品質なデータセットの不足は、拡散モデルが詳細な意味的コンテキストを抽出することを困難にしています。詳細…
-
Hi, until recently I could run this via a Google Colab notebook (with timm 0.6.13 and torch 2.0.1 to avoid other errors), but today running the following cell:
```
%cd /content/MiDaS/
!python run…