Closed xinsu626 closed 2 weeks ago
Hi @xinsu626, please set this variable: PT_HPU_ENABLE_LAZY_COLLECTIVES=true
. It is required to make HPU graphs working with tensor parallelism.
Please check: Environment variables
Hi @xinsu626, please set this variable:
PT_HPU_ENABLE_LAZY_COLLECTIVES=true
. It is required to make HPU graphs working with tensor parallelism. Please check: Environment variables
@kdamaszk Got it. Thank you for your help!
Is this functionally the same as PT_HPU_LAZY_MODE being set? (eg, per the readme worning, should only be set with eager mode?)
Your current environment
I am using the following Docker image: vault.habana.ai/gaudi-docker/1.17.0/ubuntu22.04/habanalabs/pytorch-installer-2.3.1:latest.
🐛 Describe the bug
On the main branch of the vllm-fork repository, I attempted to run the "meta-llama/Meta-Llama-3-70B" model using the following code:
However, I encountered the following error: