Closed cnut1648 closed 1 year ago
Yeah that sounds about right. You have a huge list of tasks, after all.
Are there other contexts where you get substantially more than 8.7 forward passes through a 6B model per second?
Hi @StellaAthena thanks I am still new to the field so you probably is right that this is an expected speed. Just want to check if I enable all accelerating configurations. Thanks. I am closing this issue.
Hello, I want to evaluate some 7B models using multi-gpu on a cluster of tasks. Right now I use the master branch's latest commit and run the following commands:
Basically it runs every model in
MODELS
with 0,1,5 shots on several tasks. Running gpt-j-6b 1 shot gives me 231k total examples to evaluate and takes about 8 hours to finish 0 shot on 8 A100-40G GPUs (each takes about 34G memory usage, around 8.7 instances per second). I wonder if this is an expected speed since it seems to be slow (e.g. running all might takes 8x3x6 = 144 hours). I tried bf16 (bydtype="bfloat16"
) but I am not yet sure if bf16 is used. I also madeuse_accelerate=True
. Is there any way to speed up this process? Thank you!