Open WesleyHsieh0806 opened 7 months ago
[Issue solved]
Looks like the output_ids
from LLaVA-v1.6 does not include the input prompt. We should thus change
outputs = self.tokenizer.decode(output_ids[0, input_ids.shape[1]:])
to
outputs = self.tokenizer.decode(output_ids[0])
to avoid truncation.
I'm not sure if we need to update cli.py for llava-v1.6? @haotian-liu
That is the problem I observed several days ago. but I didn't make a PR since it was not clear how should we distinguish different behaviors from model. I hope @haotian-liu can either create a unified behavior between models or determine which criterion to use to distinguish between two different kinds of behavior
Thanks for reporting, yes I missed that file, and we should make that change. Just pushed the change to main
That is the problem I observed several days ago. but I didn't make a PR since it was not clear how should we distinguish different behaviors from model. I hope @haotian-liu can either create a unified behavior between models or determine which criterion to use to distinguish between two different kinds of behavior
Definitely. We're working on a major refactor to make these behaviors more consistent. Thank you!
Describe the issue
Issue:
Hi, I tried to evaluate LLaVA-v1.6 on Science-QA, but the model keeps generating empty responses as shown in the log. Did I miss something?
Prompt and Response (Empty String)
Code