jzhang38 / EasyContext

Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Apache License 2.0
529 stars 33 forks source link

integrate it into the Transformers Trainer? #20

Open jkl375 opened 2 months ago

jkl375 commented 2 months ago

Great work! Is there a plan to integrate it into the Transformers Trainer?

jzhang38 commented 2 months ago

Currently there is no such plan.