-
If you could also add vision/video transformer models, please. Thanks in advance
-
Hi,
I'm trying to constrain the generation of my VLMs using this repo; however i can't figure out the way to personalize the pipeline for handling inputs (query+image). Whereas it is documented as …
-
I see that the multi-model models in the example all use TensorRT directly to deploy vision encoders, why not use TensorRT-LLM? Are there known issues or challenges associated with integrating Context…
-
Curious why you made that decision?
-
features = self.dino_block.forward_features(x.to("cuda"))['x_norm_patchtokens']
File "/root/.cache/torch/hub/facebookresearch_dinov2_main/dinov2/models/vision_transformer.py", line 258, in forward_…
-
### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
binary
### TensorFlow version
2.15
### Custom code
Yes
### OS platform and dist…
-
Hi,
Niels here from the open-source team at Hugging Face. It's great to see you're releasing models + data on HF, I discovered your work through the paper page: https://huggingface.co/papers/2407.1…
-
```
>python pdf_extract.py --pdf ./pdf/第一单元.pdf
Namespace(pdf='./pdf/第一单元.pdf', output='output', vis=False, render=False)
2024-07-15 16:43:01
Started!
Traceback (most recent call last):
Fi…
-
# Vision Transformer Adapter for Dense Predictions
Info.
- ICLR 2023 spotlight
- https://github.com/czczup/ViT-Adapter
- https://arxiv.org/abs/2205.08534
### Summary
- plain ViT
- whi…
-
## Expected Behavior
When using the given model initialization code:
```
from open_flamingo import create_model_and_transforms
model, image_processor, tokenizer = create_model_and_transforms…