Issue:
Excuse me, I would like to ask: when fine-tuning LLaVA with LoRA on a custom dataset to create an agent for a multimodal binary choice task, the loss quickly drops to near zero during training. However, during testing, it gives the same output regardless of input, with very low accuracy. Interestingly, this issue doesn't occur when training an agent for an eight-choice task. Why might this be happening?
Command:
Describe the issue
Issue: Excuse me, I would like to ask: when fine-tuning LLaVA with LoRA on a custom dataset to create an agent for a multimodal binary choice task, the loss quickly drops to near zero during training. However, during testing, it gives the same output regardless of input, with very low accuracy. Interestingly, this issue doesn't occur when training an agent for an eight-choice task. Why might this be happening? Command:
Log: