-
### System Info
- `n1-standard-16` from GCP with 4x NVIDIA T4s
- nvidia-smi: `NVIDIA-SMI 550.54.15`, `Driver Version: 550.54.15`, `CUDA Version: 12.4`
- Using a `NVIDIA GPU Optimized` base image
#…
-
### System Info
-GPU A100
NVIDIA-SMI 535.183.01 Driver Version: 535.183.01 CUDA Version: 12.2
NVIDIA A100-SXM4-80GB
### Who can help?
@byshiue @kaiyux
### Information
- [X] …
-
Is it possible to merge multimodal LLMs?
For example, could Llava and CodeLlama be merged? It might be beneficial for some software engineering tasks
-
### System Info
- 2x H100 80GB on docker container (nvidia/cuda:12.4.1-devel-ubuntu22.04)
- last version of the library
### Who can help?
_No response_
### Information
- [X] The official example…
-
**What would you like to be added/modified**:
A benchmark suite for multimodal large language models deployed at the edge using KubeEdge-Ianvs:
1. Modify and adapt the existing edge-cloud data c…
-
Request to Add Multimodal LLMs in unsloth
Rivising my Previous Issue: https://github.com/unslothai/unsloth/issues/376
-
### Describe your use-case.
There are multiple simple models used in this repository: Blip, Clip and WD-taggers. However, when it comes to detailed description, they are all dwarfed by modern multi…
-
Hi TensorRT-LLM team, Your work is incredible.
By following the READme file for [multi-modeling](https://github.com/NVIDIA/TensorRT-LLM/blob/main/examples/multimodal/README.md), we were sucess to run…
-
### System Info
...
### Who can help?
_No response_
### Information
- [ ] The official example scripts
- [ ] My own modified scripts
### Tasks
- [x] An officially supported task in the `example…
-
such as,i want to use a Visual Pretrained Language Models to take the image embedding,and add it to llm input to get the output