Closed JingxinLee closed 1 year ago
Unfortunately we don't have the compute resources to do that.
Unfortunately we don't have the compute resources to do that.
So how many H100/H800 GPUs do you need to train a OpenLLaMA-65B model? May I ask the minimum number?
I believe you will need at least 1000 of them to finish the training within a month.
If I don't mind the amount of month to train, for example, I have 5 months to train a 65B model, can I say that I need at least 1000/5=200 GPUs? If I have 10 months to train, can I say I need at least 1000/10=100 GPUs? Thanks
That sounds about right.
Do you have a plan for releasing OpenLLaMA-65B? Maybe met the GPUs amount bottleneck? Thanks