-
Hey @sjahangard
1) I assume this function is used to draw frames from the video and feed them to the image-based model. Is that correct?
https://github.com/JRDB-dataset/JRDB-Social/blob/b9b5ee…
-
Welcome to try the MiniGPT-4 Colab notebook: https://github.com/Czi24/Awesome-MLLM-LLM-Colab/tree/master/MLLM/MiniGPT-4-colab. It provides a convenient environment for testing and running the model.
Czi24 updated
1 month ago
-
Posting to gauge/express interested in MiniGPT-v2 support being added.
-
I can not understand why size=(batch_size, 577, 768), in vit , the size = (batch_size, 257, 1408), in vit and q-former , the size = (batch_size, 32, 768)
![image](https://github.com/user-attachments/…
-
I see in the code that in many places, self.llama_tokenizer.padding_side is set to "right,"(https://github.com/Vision-CAIR/MiniGPT-4/blob/23671f986135de7947117e0e1a514eeb6b67bea2/minigpt4/models/minig…
-
Hi, In one of the issues, you said that the test data does not need data processing. Is `MiniGPT-v2` or `TextBind` applied to this data? What about the results reported in the article? Because many of…
-
Do you have any plans to support multimodal LLMs, such as MiniGPT-4/MiniGPT v2 (https://github.com/Vision-CAIR/MiniGPT-4/) and LLaVA (https://github.com/haotian-liu/LLaVA/)? That would be a significan…
-
**Objective:** Improve clarity of miniGPT through comments or code modifications that do not alter functionality.
- We can check http://nlp.seas.harvard.edu/annotated-transformer/ for inspiration
…
-
Hi, I run the imgJP-based Jailbreak(Multiple Harmful Behaviors) method for MiniGPT-4(LLaMA2) attacks, Run the provided code python v1_mprompt.py --cfg-path eval_configs/minigpt4_llama2_eval.yaml --gpu…
-