-
### Feature request
It is on the road map to have images embeddings models?
### Motivation
This is very useful since there are many VLLMs coming out.
### Your contribution
Any thing that is neede…
-
After installing CLIP, I get this error
File "C:\Users\sloom\AppData\Local\NVIDIA\ChatRTX\env_nvd_rag\lib\site-packages\transformers\image_transforms.py", line 386, in normalize
raise ValueE…
-
### Describe the bug
`transformers` added `sdpa` and FA2 for CLIP model in https://github.com/huggingface/transformers/pull/31940. It now initializes the vision model like https://github.com/huggingf…
-
Hello InternVideo team,
You guys have done a great job with this project!
In your paper, you use the Stage 2 model for the task of temporal grounding on QVHighlight [Lei et al., 2021] and Charad…
-
does anyone else get a similar issue during training?
```Traceback (most recent call last):
File "./train_long_clip.py", line 425, in
raise e
File "./train_long_clip.py", line 395, in
…
-
Loading a model requires is_timm, is_clip flags but these should be inferred from model name
-
### What happened?
作者您好,我基于分支minicpm-v2.5构建了应用程序,MiniCPM-V-2.5 模型运行良好。
从 Hugging Face 上下载了 MiniCPM-V-2-GGUF文件,运行失败,请问是不支持 MiniCPM-V2 吗?
### Name and Version
./bin/minicpmv-cli -m /home/wuzhi/share…
-
Common issue with FPS games
Possible workarounds:
-a shader???
-make the gun move out of the way when close to a surface; issues - a lot of work, possibly many edge cases
-make the guns very, ve…
-
### Feature request
I wonder if the task text-classification can to be supported in the ONNX export for clip? Ich want to use the openai/clip-vit-large-path14 model for zero-shot image classification…
-
1.我利用CN-CLIPViT-B/16模型向量化几张狗的图片,并将向量化的结果加到Groma向量数据库中;之后我在搜索匹配时输入特殊符号!@#¥时却会搜索到狗的图片?为什么与向量化图片无关的输入会搜到结果呢?
2. 并且我在搜索“狗”时出现的结果中,有几张图片的相似度要比搜索!@#¥时要低,为什么与向量化图片无关输入的相似度会比有关系的输入还要低呢?
3. 我的代码如下:
`import …