a-r-r-o-w / cogvideox-factory

Memory optimized finetuning scripts for CogVideoX using TorchAO and DeepSpeed
Apache License 2.0
177 stars 16 forks source link

add "max_sequence_length": model_config.max_text_seq_length, #15

Closed glide-the closed 2 days ago

glide-the commented 2 days ago

prepare_dataset.sh Include max_sequence_length, bug train verify log not it

sayakpaul commented 2 days ago

Works for me.

If we are serializing the text embeddings beforehand, I think this argument shouldn't affect the training script, no?

a-r-r-o-w commented 2 days ago

This is only used during validation, so prepared embeddings should already have the same max sequence length. Basically, when someone is preparing the precomputed embeddings, they'd have to make sure what the model is training with, and what was specified as an argument to the prepare_dataset.py script are the same (and even if not, training will remaining unaffected).