vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
26.8k stars 3.93k forks source link

[Usage]: Does VLLM support starting multiple cards using mpirun? Want to bind different CPUs to each card. #8069

Open xiabo123 opened 2 weeks ago

xiabo123 commented 2 weeks ago

Your current environment

Does VLLM support starting multiple cards using mpirun? Want to bind different CPUs to each card. When testing performance, use mpirun to bind different CPUs to different cards.

How would you like to use vllm

Does VLLM support starting multiple cards using mpirun? Want to bind different CPUs to each card. When testing performance, use mpirun to bind different CPUs to different cards.

Before submitting a new issue...

youkaichao commented 1 week ago

vllm does not support mpirun. please check out https://docs.vllm.ai/en/latest/serving/distributed_serving.html for how to use multiple gpus.