Closed JarvisUSTC closed 1 month ago
Hi, thanks for the question, and sorry for the confusion. To clarify:
If you directly use the original implementation of llava, it's likely the safety of the model is very bad --- we mentioned it later in this paper https://arxiv.org/abs/2310.03693. Basically, llava fine-tunes the LLM backbone as well during modality alignment. This can lead to regression of the safety alignment in the LLM backbones. To maintain the safety performance, we fine-tune one version of our own llava models using llava's repository. The only change is we freeze the weights of the llama2 model backbone. Then, the safety would be better. We use that version of the model in the AAAI version of the paper.
Get it. Thanks for your response. Could you please provide the finetuned weight of your own llava model for reproduction?
BTW, is the InstructBLIP used in this paper also finetuned by yourself?
BTW, is the InstructBLIP used in this paper also finetuned by yourself?
Unfortunately, I may no longer keep that checkpoint. But it should be implementable by running the llava fine-tuning script by freezing the llama2 backbone.
BTW, is the InstructBLIP used in this paper also finetuned by yourself?
No, for InstructBLIP, we just used the official checkpoint.
BTW, is the InstructBLIP used in this paper also finetuned by yourself?
Unfortunately, I may no longer keep that checkpoint. But it should be implementable by running the llava fine-tuning script by freezing the llama2 backbone.
Get it. So, what kind of training data is used for fine-tuning? If possible, I still want to reproduce this experiment.
BTW, is the InstructBLIP used in this paper also finetuned by yourself?
Unfortunately, I may no longer keep that checkpoint. But it should be implementable by running the llava fine-tuning script by freezing the llama2 backbone.
Get it. So, what kind of training data is used for fine-tuning? If possible, I still want to reproduce this experiment.
Hey, you can just follow the instructions of llava --- it provides training codes. The only change is probably to just freeze weights for the LLM backbone. Sorry for the inconvenience.
Oh, I misunderstanded before. Thanks for your kind response.
Hello, I am currently working on reproducing the results of Table 3 in your paper but have encountered some discrepancies in the performance of LLaVA (LLaMA-2-Chat). Despite following the provided instructions and using the same dataset and configuration, I am unable to reproduce the reported results (has a large margin: 9.2 vs. 60.9).
Details:
Steps to reproduce:
python -u llava_llama_v2_inference.py --image_file adversarial_images/clean.jpeg --output_file outputs/baseline/clean_llava_llama_v2_13b/result.jsonl
python get_metric.py --input outputs/baseline/clean_llava_llama_v2_13b/result.jsonl --output outputs/baseline/clean_llava_llama_v2_13b/rtp_prompts_challenges_metrics.jsonl
Questions:
Best regards