-
-
### System Info
- `transformers` version: 4.44.2
- Platform: Windows-10-10.0.22631-SP0
- Python version: 3.9.13
- Huggingface_hub version: 0.24.7
- Safetensors version: 0.4.5
- Accelerate vers…
-
### Feature request
Currently, the [visual-question-answering pipeline/task](https://huggingface.co/tasks/visual-question-answering) in transformers is not supported for onnx export:
https://githu…
-
![image](https://github.com/user-attachments/assets/75ef8603-9cba-404c-9435-7a946e9bb2b0)
-
I always encounter this error in this path ("Retrieval-Augmented-Visual-Question-Answering/runway_for_ml/experiment.py") appears ([ERROR] - runway_for_ml.experiment : Uncaught exception: --> Trainer.…
-
- [VQAv2](https://arxiv.org/pdf/1612.00837v3)
- [TallyQA: Answering Complex Counting Questions](https://arxiv.org/pdf/1810.12440)
- [GQA: A New Dataset for Real-World Visual Reasoning and Compos…
-
Hello,
How can we have the confidence of the visual question answering prediction?
Thanks
-
Hi, It is great work and I'm following your work. I have some questions and hope you give me some solution. First, how to code "Entity Relevance" and "Relational Relevance" in VQA task, just use the "…
-
Dear Authors,
We'd like to add "GITA: Graph to Visual and Textual Integration for Vision-Language Graph Reasoning" to this repository, which has been accepted by NeurIPS 2024. [**Paper**](https:/…
-
|id|title|author|year|
|---|---|---|---|
|1|Graph-Structured Representations for Visual Question Answering|Teney, Damien and Liu, Lingqiao and van den Hengel, Anton|2017|