Closed philipkiely-baseten closed 2 months ago
Llama 3 70B only needs 2, not 4, 80 GB GPUs to run inference. I've changed the config and tested the updated config in production.
Llama 3 70B only needs 2, not 4, 80 GB GPUs to run inference. I've changed the config and tested the updated config in production.