-
Hello, is the sum of the Multimodal Interaction (MMI) Module.Module part add or concat?
-
# [24’ CVPR] AnyRef: Multi-modal Instruction Tuned LLMs with Fine-grained Visual Perception - Blog by rubatoyeong
Find Directions
[https://rubato-yeong.github.io/multimodal/anyref/](https://rubato-…
-
**Is your feature request related to a problem? Please describe.**
- obsidian in many usecases can contain a lot of non-text message, unstructured data such as
1. imgs
2. pdfs
3. pdfs with c…
-
### Feature request
Hi, I am wondering that can this repository supports the unified multimodal model like Show-o? [https://github.com/showlab/Show-o](https://github.com/showlab/Show-o)
### Motivati…
-
Whether GroundingDINO can support TensorRT-LLM multimodal ?
[TensorRT-LLM multimodal ](https://github.com/NVIDIA/TensorRT-LLM/blob/main/examples/multimodal/README.md)
-
Hi,
I hope this message finds you well. I am very interested in your project, Social Media Popularity Prediction using Multimodal Deep Learning, and the innovative approach you've taken with multim…
-
Add support to multimodal models as discussed with @haileyschoelkopf
- This PR #1832 would be a great starting point.
- List all tasks we want to support for the first iteration.
-
Is there a sample code or can you guide to pass additional context to llm, like in this pipeline agents example with new openai multimodal example?
https://github.com/livekit/agents/blob/main/examp…
-
Link : https://if.kakao.com/session?t.RVmWsFVgRd=0
---
AI mate와의 새로운 연결 Kanana - 이상호
Kakao AI Platform : AI 서비스의 품질은 높이고, 개발은 빠르게 - 홍석용
이미지까지 이해하는 Multimodal LLM의 학습 방법 밝혀내기 - 강우영
-
### Describe the issue linked to the documentation
Users need to understand the list of supported loss functions for their problem type in our doc, specifically at [here](https://auto.gluon.ai/stable…