-
Hello,
I tried Kosmos-2 with some high-quality images: The code generated images and labels files in the folder.
but all the labels were empty. Here is the code that I was running on Apple M1. Wha…
-
Hi,
Thank you so much for developing this impactful and impressive work! This work really bridges the gap in multimodal grounding capability to the visual world.
I would like to kindly ask if y…
-
**Describe the bug**
Model I am using: Kosmos-2.5
The problem arises when using:
* [ x ] the official example scripts: Using the exact required custom-libraries and dependencies to run the suppli…
-
After working whole day I finally coded it and published it.
Also I found a better prompt that captions better.
**You can download auto installer at here : https://www.patreon.com/posts/90744385…
-
### Feature request
Some of our models interpolate its positional embeddings, enabling pretrained checkpoints to be used on different input resolutions. For example, [here in ViT](https://github.co…
-
Is there a way to train novel concepts into your blip model, like the way that textual inversions work for stable diffusion image generation? If so is there a training script provided or would one nee…
-
Hello, when I tried to run it, I found that it stopped working, what should I do?
-
Hi
I am still getting this error despite trying all the fixes noted for previous issues on the subject? Is it broken again?
```
Traceback (most recent call last):
File "C:\SDAI\ComfyUI\ComfyUI…
-
Hello,
Thank you for your time and for sharing your work!
I'm interested in reproducing the generated caption dataset provided in [here](https://connecthkuhk-my.sharepoint.com/personal/jhyang13_…
-
Hello everyone, thank you very much for your contribution. I appreciate the effort and consistency in uploading the code for such many models and maintaining this repository.
I saw Kosmos-2 and I q…