-
### User Story
As a PUL Librarian in Area Studies, I am in contact with documentary film makers and campus partners who are interested in the library hosting, preserving and making available documenta…
-
> Technically, we employ the lightweight image captioning model Tag2Text [59] for the finer scale, which describes videos at low fps in a frame-by-frame manner. These individual image captions are the…
-
I have dedicated several days, working over 12 hours each day, on SUPIR (Scaling-UP Image Restoration), a cutting-edge image enhancement and upscaling model introduced in the paper [Scaling Up to Exce…
-
**1. Requester Information:**
Name: Ibrahim Demir
Position: Associate Professor
Institution: University of Iowa
Contact: +1 (319) 335-5780, ibrahim-demir@uiowa.edu
**2. Project Information:**
…
-
Thx for awesome job!
I have tried to understand how it works judging from the code but it's hard for me:(
How does it describes what's happening on the images? What tech do you use? Thx!
-
The framework is very nice and I downloaded the motion feature extracted from resnext-101, but found the shape of feature are different for each video. So when I run the code, it goes wrong ,
![image…
-
I've run the requirements installation, but the model spits out random words when I run the video examples from gradio.
The image captioning works fine.
The hugging face demo for video captioning wo…
-
Hi,
I find your project intriguing and believe it could greatly assist in working with multiple data sources. However, I noticed that you haven't mentioned how the vector data generated by your proje…
-
The various challenges involved in making sense of an image found on social media is summarized by this image
![Screenshot 2023-12-04 at 15-13-05 Tech Interventions against Online Harms](https://gith…
-
Hi, thank you so much for the great works! I have questions about sampled frame number, in the paper mentioned
> During inference, we uniformly sample 6 frames with center crop.
I am keen to kn…