jzhang38 / EasyContext

Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Apache License 2.0
653 stars 47 forks source link

dependency confilct #53

Open SihengLi99 opened 2 months ago

SihengLi99 commented 2 months ago

Hi,

Now, ring-flash-attention is conflicted with transformers==4.39.1