Open lyzKF opened 1 year ago
To train a model(7B) with megatron-deepspeed, tensor_parallelism=2 pipeline_parallelism=8
how many GPUS do i need?
Roughly calculated, about 280G of memory is required, which is divided into 16 cards, and each card occupies about 17G.
To train a model(7B) with megatron-deepspeed, tensor_parallelism=2 pipeline_parallelism=8
how many GPUS do i need?