-
# URL
- https://arxiv.org/abs/2411.02571
# Authors
- Sheng-Chieh Lin
- Chankyu Lee
- Mohammad Shoeybi
- Jimmy Lin
- Bryan Catanzaro
- Wei Ping
# Abstract
- State-of-the-art retrieval mod…
-
**New Feature**:
The final LLM answering to the query should take as input the images extracted from the files.
**Specification**:
- Multimodal LLMs supported
- Easy to extend for new architec…
-
Do you have any plans to support multimodal LLMs, such as MiniGPT-4/MiniGPT v2 (https://github.com/Vision-CAIR/MiniGPT-4/) and LLaVA (https://github.com/haotian-liu/LLaVA/)? That would be a significan…
-
### Feature request
Is it possible to rum multimodal LLMs like Qwen VL or LLaVa 1.5 using openllm?
### Motivation
_No response_
### Other
_No response_
-
### Title
Dreamcatcher: decoding dream-event in EEG data with multimodal language models and interpretability tools
### Leaders
Lorenzo Bertolini
### Collaborators
_No response_
###…
-
### Feature Description
AWS Bedrock has a few multimodal LLMs such as Claude Opus. It would be great if this can be added as a multi-modal-llm integration. There is already an anthropic multimodal …
-
### Description
When using memory=True for a crew that uses Azure Open AI, there is an error creating long term memory.
### Steps to Reproduce
```
import os
from chromadb.utils.embedding_…
-
Is it possible to merge multimodal LLMs?
For example, could Llava and CodeLlama be merged? It might be beneficial for some software engineering tasks
-
### 🚀 The feature, motivation and pitch
Let's build an example demo app, perhaps in pytorch-labs, which will become a forcing function to improve developer experience from a user perspective. A pos…
-
### 🚀 The feature, motivation and pitch
***Please note that since the actual implementation is going to be simple, and the design has already been reviewed, the purpose of this GitHub Issue is to l…