Closed yananchen1989 closed 1 month ago
hello. is it ok to ppo train 7b level models on single gpu where each sample has around 6k length in prompt and round 2k length in completion ?
any advice for this scenario ? thanks
A single GPU is not enough to run 7B + long text
hello. is it ok to ppo train 7b level models on single gpu where each sample has around 6k length in prompt and round 2k length in completion ?
any advice for this scenario ? thanks