-
# Vision Transformer Adapter for Dense Predictions
Info.
- ICLR 2023 spotlight
- https://github.com/czczup/ViT-Adapter
- https://arxiv.org/abs/2205.08534
### Summary
- plain ViT
- whi…
-
### Model description
I know the transformers library has not included object tracking models in the past, but this one can either plug into any object detection model or be an end-to-end open world …
-
# Supervised Transformer Network for Efficient Face Detection #
- Author: Dong Chen, Gang Hua, Fang Wen, Jian Sun
- Origin: https://arxiv.org/abs/1607.05477
- Related:
-
Argilla integration, dataset integration etc.
detail to follow.
-
### Describe the bug
I trained a lora with simpletuner using ai-toolkit preset (I used all+ffs and others and it doesnt train correctly on hard concepts).
And Now I have this issue when loading the …
-
Can you add support so that we can add language as a parameter ? I am trying to use the diarization code given at the readme file and receiving the following error.
Due to a bug fix in https://github…
-
Hello, as you mention you borrowed some parts from the [Swin-Transformer-Object-Detection](https://github.com/SwinTransformer/Swin-Transformer-Object-Detection) repository for the detector part. It se…
-
## 一言でいうと
物体検知/セグメンテーションにTransformer (#329 を適用した研究。ResNetで特徴抽出しTransformerに入れるシンプルな構成だが、End2Endで予測するため(=重複領域除去処理(NMS)フリーにするため)固定数の検出予測と実際をマッチングさせるlossを導入している。
![image](https://user-images.github…
-
## 一言でいうと
物体検出を予測領域と実領域(ラベル)との組み合わせ問題として解く研究。CNNで得た特徴と空間位置を表すpositional encoding(PE)をTransformer(Encoder)に入力し、出力をTransformer(Decoder)に渡し物体検出用PEでクエリする。Decoderはクラス確率と物体(中央)位置を出力。ラベルとの適合から学習する
![im…
-
Vision Transformers should be supported out-of-the-box by `quanto`.
The goal of this issue is to add some examples under `examples/vision`.
At the very minimum, there should be a classification …