Open rasbt opened 2 months ago
I recommend investigating it.
Just by quickly checking I saw that neither HF modeling file, nor the weights were updated.
The error message says that it didn't get norm_2.weight
but got post_attention_norm.weight
, although it has to be mapped perfectly:
https://github.com/Lightning-AI/litgpt/blob/b0ea1772f7498e2e8cc58d2ac1640b0255ced757/litgpt/scripts/convert_hf_checkpoint.py#L150
Well, maybe not perfect, because I see that unexpected weights start with transformer.h
, while it has to be model.layers
.
Something is off.
Bug description
It seems that they updated the Gemma v1 2B weights. Something to look into:
We can either fix or remove these. Because there's Gemma 2, not sure why someone would care about Gemma 1. What do you think @Andrei-Aksionov ?
What operating system are you using?
Unknown
LitGPT Version