Open Mikael17125 opened 1 month ago
model = dict(
freeze_llm=True,
freeze_visual_encoder=True,
llm=dict(
attn_implementation='eager',
pretrained_model_name_or_path='microsoft/Phi-3-mini-4k-instruct', # ---------------
type='transformers.AutoModelForCausalLM.from_pretrained',
trust_remote_code=True),
pretrained_pth='/home/oem/xtuner/pretrained/phi-3/model.pth',
type='xtuner.model.LLaVAModel',
visual_encoder=dict(
type=CLIPVisionModel.from_pretrained,
pretrained_model_name_or_path=visual_encoder_name_or_path))
it gives me an error like this:
The attention mask is not set and cannot be inferred from input because pad token is same as eos token.As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
The `seen_tokens` attribute is deprecated and will be removed in v4.41. Use the `cache_position` model input instead.
hi, can I knew the exact syntax, mine is still error:
it said I have no config.json
even when I specified the .pth it still give me error: