-
Hello, after reading your paper, I was really impressed by not only the innovative research but also the beautiful and intuitive way the graphs were presented. I wanted to ask you, how did you create …
-
How about add Visual Question Answering to hezar ?
I saw a few days ago that we have a visual question answering benchmark for persian and i thought its nice to have vqa in hezar
I would like also …
-
getting this error while running the default videos:
visual_question_answering(("how many boats are there in the video?", 0)) is not a valid tool, try one of [caption_retrieval, segment_localization,…
-
Hi, I'm confused I did some visual answering with the InternVL2-26B model and it performs very badly in that. The only model that passes that question are Gemini 1.5 pro/flash, gpt4-o, and Claude.
…
-
Hello,
How can we have the confidence of the visual question answering prediction?
Thanks
-
### Feature request
Currently, the [visual-question-answering pipeline/task](https://huggingface.co/tasks/visual-question-answering) in transformers is not supported for onnx export:
https://githu…
-
Hi, It is great work and I'm following your work. I have some questions and hope you give me some solution. First, how to code "Entity Relevance" and "Relational Relevance" in VQA task, just use the "…
-
The following tasks are available in the model hub, and seem to have inference support, but are not yet listed in the [Inference API docs](https://huggingface.co/docs/api-inference/detailed_parameters…
-
|id|title|author|year|
|---|---|---|---|
|1|Graph-Structured Representations for Visual Question Answering|Teney, Damien and Liu, Lingqiao and van den Hengel, Anton|2017|
-
Enjoying the recent gradio notebook stuff!
Was curious about when/if support for an additional hugging face task option of ["visual question answering“](https://huggingface.co/models?pipeline_tag=…