-
I followed the paper for re-implementation. I got good results on DF and FS, but pool results on F2F and NT. Did you encounter a similar situation?
-
### 0.5b response is norm but 7b wrong
the same image,where i chage the code is` pretrained = "/home/shihongyu/MMLM_models/lmms-lab/llava-onevision-qwen2-7b-ov"
model_name = "llava_qwen"
device = "…
-
Would you please tell the torch seed of 2 pretrained models.
I found the result varies with torch seed, and I ran many times but failed to realize 76.2 for caption retrieval. I only got 75.6 instead.
-
**Expected Outcomes**
- Prompt: Summarize the content from the url (do not emit the url back) https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/9/html/managing_file_systems/ind…
-
Is there a plan to release the evaluation code? How can I quantitatively evaluate the quality of video generation?
-
File "/root/.local/lib/python3.12/site-packages/vlmeval/dataset/image_mcq.py", line 181, in evaluate
answer_map = {i: c for i, c in zip(meta['index'], meta['answer'])}
…
-
You will see the problem in the text below, this is with using gpt-4o and version 0.5 of agent zero, but have similar issues with other models
User message ('e' to leave):
> Write a college level …
-
Did anyone try to run NVDS in Windows? I suspect it is not that easy because of GMFlow...
-
Hi,
thanks to the authors of this paper and this code for making the effort to share their work with the community.
I am trying to use Dino on a non-imagenet dataset and started to alter the cod…
-
Hey all!
The video models are all supported in Transformers now and will be part of the v4.42 release. Feel free to check out the model checkpoints [here](https://huggingface.co/collections/llava-h…