EleutherAI / gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
https://www.eleuther.ai/
Apache License 2.0
6.95k stars 1.02k forks source link

Post training examples #1277

Closed dmahan93 closed 2 months ago

dmahan93 commented 2 months ago
dmahan93 commented 2 months ago

will need to edit the configs once whatever changes in https://github.com/EleutherAI/gpt-neox/pull/1276 get finalized

Quentin-Anthony commented 2 months ago

will need to edit the configs once whatever changes in #1276 get finalized

It's been merged. You're clear to do a final pass + test run.

Quentin-Anthony commented 2 months ago

Tested and working for me