Open DwanZhang-AI opened 1 week ago
This is the output of the model.
python -W ignore llava/eval/run_vila.py \ --model-path Efficient-Large-Model/Llama-3-VILA1.5-8b \ --conv-mode llama_3 \ --query "\n Please describe the traffic condition." \ --image-file "demo_images/av.png"
This is the inference code. Why the output is wrong?
BTW, I have muted the flash attention module.
seems works properly on my side, should not be a issue for flash-attn. Could you make sure you have done a fresh install following readme?
This is the output of the model.
python -W ignore llava/eval/run_vila.py \ --model-path Efficient-Large-Model/Llama-3-VILA1.5-8b \ --conv-mode llama_3 \ --query "\n Please describe the traffic condition." \
--image-file "demo_images/av.png"
This is the inference code. Why the output is wrong?