dvmazur / mixtral-offloading

Run Mixtral-8x7B models in Colab or consumer desktops
MIT License
2.29k stars 227 forks source link

Update typo in README.md #20

Open kaushalpowar opened 10 months ago

kaushalpowar commented 10 months ago

There was a typo (pack -> back).

Old: Each expert per layer is offloaded separately and only brought pack to GPU when needed.

Changed: Each expert per layer is offloaded separately and only brought back to GPU when needed.