Open lucasjinreal opened 4 months ago
Hi Could you please elaborate on it? Try using Cambrian7M with system prompts: https://huggingface.co/datasets/nyu-visionx/Cambrian-10M/resolve/main/jsons/Cambrian7M_withsystemprompt.jsonl
I tried using Cambiran 7M with systme prompts.
I don't know why, maybe the more longer training process, shouldn't use same wramup ratio?
I tried the intermediate model, the inference is not as good as expected, almost can be treat as over-fitted.
I am not sure how to dealing with so larger instruction data, any suggestions?
Haven't u guys met similar issue?
Hi,
maybe try following our training hyperparameters? Also, we encountered similar issues but we later realized its bug with transformer version in the inference time. Maybe this will help a bit
@tsb0601 Hi! Could you please provide the detailed training hyperparameters or which file should I refer to?
Hello, I meet the same problem. Using this dataset results in inferior results than LLaVA. @lucasjinreal What are your setting?
same problem
Normaly, I using my data got a reasonable result, but same training process, just replace data with Cambrian 7M, got some extremly bad result, why?