-
Hello,
I've encountered an issue while trying to run the `scripts/evaluate_pretrained_model.sh` script as instructed in the `TRAINING_README.md` file.
```shell
python -m training.offline.online…
-
Hello, I finetuned LLaVA onevision with Qwen2-7B. In the finetuning script, I set it to finetune just the adapter. When I am trying to inference my model, I am using the example code given in the rep…
-
I know that legacy uses openaiclip's github and main branch uses clip_trt's github. I used main branch to make coco's vector.bin and the search perfomance is very bad. I would like to ask where I didn…
-
Thank you for the incredible set of repositories (this one and prismatic-vlms), it has been a great joy using them. Very well-designed, configurable, and easy to use for researchers.
I'm running in…
-
-
Hi, I get this error when preprocessing text using the mSigLIP model. Any idea what may be wrong? I didn't change anything in the [demo colab ](https://colab.research.google.com/github/google-research…
-
Hi, I am using sglang to deploy llava-next-interleave-qwen-7b, but i found there is no preprocessor_config.json for llava-next-interleave-qwen-7b model, could we add this to huggingface? Or do we hav…
-
Thanks for your great work, but due to my poor knowledge, what's the 'hf_token' in your code of inference? How can I get it?
-
Hello,
I watched a video about Live ViLA and was impressed by the 3B model running on the edge. Regarding this, I'm curious about how frame sequences are processed during video understanding.
1. H…
-
Dear all,
I have fine-tuned TinyLLaVA-Phi-2-SigLIP-3.1B for 1 epoch and then continued the fine-tuning for another epoch starting from trained models saved after the first epoch. Both training runs w…