jzhang38 / EasyContext

Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Apache License 2.0
653 stars 47 forks source link

how to infer the model? #27

Open laoda513 opened 6 months ago

laoda513 commented 6 months ago

Shall I also integrate with the same libs with training? Or just load the model as usual