-
Thanks for your excellent work :)
We admire the potential of vision-language models with unified vocabularies, and wish if some templates of the training code can be released for academic experiments…
-
This is an issue to collect requests for model abliterations.
No one is required to abliterate your request, but it does make for a good place to check if someone else has used this process on the…
-
### Model description
A new large language and vision model (LLVM) that uses auxiliary visual information and natural language for prediction.
It uses 2 modules: 𝙈𝙤𝘼𝙄-𝘾𝙤𝙢𝙥𝙧𝙚𝙨𝙨𝙤𝙧 and 𝙈𝙤𝘼𝙄-𝙈𝙞𝙭𝙚𝙧. He…
-
Hi,
Congrats on the impressive work! Our paper FAITHSCORE: Evaluating Hallucinations in Large Vision-Language Models is very related to your topic.
I wonder if you would mind adding our work to your…
-
CUDA_VISIBLE_DEVICES=0 python3 inference.py --model-path ./PCIResearch/TransCore-M --vision-path ./openai/clip-vit-large-patch14-336
You are using a model of type transcorem to instantiate a model of…
-
## LINKs
[[paper](https://arxiv.org/abs/2405.02246)](https://arxiv.org/abs/2405.02246)
[[models](https://huggingface.co/HuggingFaceM4/idefics2-8b)](https://huggingface.co/HuggingFaceM4/idefics2-8b)…
-
in load_pretrained_model
model = CambrianLlamaForCausalLM.from_pretrained(
File "/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py", line 3531, in from_pretrained
) =…
-
Thanks for your excellent work. Could you provide the code for the robot in a real-world environment?
-
- [ ] [Title: "Yi Model Family: Powerful Multi-Dimensional Language and Multimodal Models"](https://arxiv.org/html/2403.04652v1)
# Title: "Yi Model Family: Powerful Multi-Dimensional Language and Mul…
-
Hi, @wondervictor, a huge shoutout for your remarkable contributions!
I've seamlessly integrated YOLO-World into [X-AnyLabeling](https://github.com/CVHub520/X-AnyLabeling), marking a significant ad…