PKU-Alignment / align-anything

Align Anything: Training All-modality Model with Feedback
https://align-anything.readthedocs.io
Apache License 2.0
260 stars 47 forks source link

[Question] Slow Speed of vLLM when evaluating MMLU #35

Open cby-pku opened 4 months ago

cby-pku commented 4 months ago

Required prerequisites

Questions

When evaluating MMLU, the codebase supports vLLM inference, but the speed is slow (20 minutes for a single task). According to my experience, the normal speed is 20 minutes for all tasks.

Kass123777 commented 4 months ago

Thank you for your question! This is a known issue. Since the current architecture implements the BaseInference class based on deepspeed and vllm in the same Python file, importing deepspeed-related dependencies causes vllm to fail to start properly. Therefore, I set distributed_executor_backend="ray" when starting vllm. This does significantly affect efficiency. We will further modify the framework in the next version to completely decouple the two backends and fully unleash the inference speed of vllm.