Closed mathemakitten closed 2 years ago
I'm in favor of just making the disk space large enough for all zero shot submissions, so we don't need the large model dict. Disk space shouldn't cost that much anyway unless I'm missing something.
OK! In that case I've just bumped up the default to 200 GB, which accounts for 145GB for 66B and then some for the system, saving preds, etc. It also assumes that we'll route requests for models > 66B elsewhere (for now; we can change it when we start supporting 175B+ inference)—let me know if that works.
Minimal example test: