-
Hi,
I am currently working with the NExT-QA dataset and I ran your code using the model meta-llama/Meta-Llama-3-8B, as GPT-3.5 and GPT-4 are not open source. Could you please provide details on the r…
-
hi, thanks for your great work!
i had some issues when launching the demo, as no image_processor was loaded by default (same bug as a comment mentioned in the youtube demo video iirc).
i found a…
-
server:
```
export CUDA_VISIBLE_DEVICES="3,4,5,6"
python -m sglang.launch_server --model-path lmms-lab/llava-next-72b --tokenizer-path lmms-lab/llavanext-qwen-tokenizer --port=30010 --host="0.0.0.0…
-
I have never been able to correctly quantify awq for llava-llama3 in the official format of llava。
Can anyone help me?
-
It would be nice to add LLava to recognize images in web pages so that it works together with RAG models.. LLava is however a textual AI model so it can be used as the main model to describe pages, t…
-
I notice that the image processor crops rectangular images into square images, which inevitably loses some information.
It seems that cropping is also used during training.
What if we want to ca…
-
### Describe the issue
Issue:
We are trying to perform inference on the LoRA weights provided for vicuna-13b-v1.3 [here](https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md#model-zoo)…
-
[paper](https://arxiv.org/pdf/2310.03744.pdf)
see llava https://github.com/long8v/PTIR/issues/128#issue-1749571159 here
## TL;DR
- **I read this because.. :** aka LLaVA1.5 / ShareGPT4V에서 LL…
-
I have a question regarding the AnyRes feature in LLaVA-NeXT-Video. The documentation mentions that AnyRes enables high-resolution image processing. However, when examining the demo code at https://gi…
-
안녕하세요, 우선 좋은 모델과 데이터셋 공유 감사드립니다.
다름이 아니라 여러 모델들이 있는데, 한국 멀티모달 벤치마크 성능 비교가 없는 것 같아 문의드립니다. 테스트 데이터셋 기준이라도 모델별 성능을 비교한 걸 공개해주실 수 있을까요? 감사합니다.