mbzuai-oryx / Video-ChatGPT

[ACL 2024 🔥] Video-ChatGPT is a video conversation model capable of generating meaningful conversation about videos. It combines the capabilities of LLMs with a pretrained visual encoder adapted for spatiotemporal video representation. We also introduce a rigorous 'Quantitative Evaluation Benchmarking' for video-based conversational models.
https://mbzuai-oryx.github.io/Video-ChatGPT
Creative Commons Attribution 4.0 International
1.23k stars 108 forks source link

The responses from the offline_demo are garbled #95

Closed hlh2023214 closed 8 months ago

hlh2023214 commented 8 months ago

Hi @mmaaz60 I have followed the steps in the offline_demo.md meticulously for local deployment, but I do not know why the output responses are always garbled. Could you please help me to solve the issue? 屏幕截图 2024-03-27 122601

mmaaz60 commented 8 months ago

Hi @hlh2023214,

Thank you for your interest in our work. Most probably, the issue could be related to the unsupported transformers version. Please follow the following steps and let me know if it resolves the issue.

# Stop the demo

pip install -r requirements.txt

# Run the demo again and verify if the issue is still there.

# If the issue persists, please try downloading the model checkpoints again from HuggingFace.

In case if it does not solve the issue, please share the output of conda list command to further investigate. Thank You.

hlh2023214 commented 8 months ago

Hi @hlh2023214,

Thank you for your interest in our work. Most probably, the issue could be related to the unsupported transformers version. Please follow the following steps and let me know if it resolves the issue.

# Stop the demo

pip install -r requirements.txt

# Run the demo again and verify if the issue is still there.

# If the issue persists, please try downloading the model checkpoints again from HuggingFace.

In case if it does not solve the issue, please share the output of conda list command to further investigate. Thank You.

My version of the transformer was consistent, but later I found out that the Llama model I used was incorrect. I was using the Llama-2-7b-hf, which is not the same as your version, hence the garbled phenomenon occurred. Thank you, my issue has been resolved. 屏幕截图 2024-03-28 213820