Open Yang-bug-star opened 8 months ago
@zhuyiche @JLM-Z @Yang-bug-star Were you able to conduct training on your GPUs? I'm also interested in know about it.
Sorry, I haven't tried it.
I found the training time is very close to original LLaVA: llama with clip 336 batch 14 except the pretrain is 1-hour shorter, is that normal?
@feiyu12138 how many gpu's / which gpu did you use?
Is it possible to train on 2 v100 gpus due to the small language model used ?