Open fireicewolf opened 2 months ago
We will support the T5XXL to run on fp8 in the future, so please wait a little longer.
I tried in Linux and Windows and I cannot train DB in Kohya as I get this error. No, I did not try and train Clip-L, nor T5, YET this error persists.
In issue #1453 I made a PR #1452 to fix AttributeError: 'T5EncoderModel' object has no attribute 'text_model', while loading T5 into GPU without 'cache_text_encoder_outputs'.
But I did't check
T5EncoderModel
dict.😅 So it has bugs while load T5 with different dtype(ex. Using FP8 without 'cache_text_encoder_outputs'.)T5EncoderModel Dict:
T5EncoderModel
have 3 itemsshared
,encoder.embed_tokens
,encoder.block[0].layer[0].SelfAttention.relative_attention_bias
have attributeEmbedding
So I fixed it in PR #1508 Second commit.With
fp8_base
if t5xxl load in GPU usefloat8_e4m3fn
, will cause loss nan, So I change it tofloat8_e5m2
.@kohya-ss Please check PR #1508, its first commit also fix an issue metioned in #1509. Thanks.