vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
30.27k stars 4.59k forks source link

Fatal Python error: Segmentation fault #3053

Open lmx760581375 opened 8 months ago

lmx760581375 commented 8 months ago

SIGSEGV received at time=1709021359 on cpu 44 PC: @ 0x7f3c5f628350 (unknown) (unknown) @ 0x7f3c945c8630 (unknown) (unknown) @ 0x55b11321eaf0 1247139872 (unknown) @ ... and at least 2 more frames [2024-02-27 16:09:19,106 E 19394 19394] logging.cc:361: SIGSEGV received at time=1709021359 on cpu 44 [2024-02-27 16:09:19,108 E 19394 19394] logging.cc:361: PC: @ 0x7f3c5f628350 (unknown) (unknown) [2024-02-27 16:09:19,108 E 19394 19394] logging.cc:361: @ 0x7f3c945c8630 (unknown) (unknown) [2024-02-27 16:09:19,110 E 19394 19394] logging.cc:361: @ 0x55b11321eaf0 1247139872 (unknown) [2024-02-27 16:09:19,110 E 19394 19394] logging.cc:361: @ ... and at least 2 more frames Fatal Python error: Segmentation fault

Stack (most recent call first): File "/root/.conda/envs/py39/lib/python3.9/site-packages/torch/cuda/graphs.py", line 77 in capture_begin File "/root/.conda/envs/py39/lib/python3.9/site-packages/torch/cuda/graphs.py", line 192 in enter File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/worker/model_runner.py", line 782 in capture File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/worker/model_runner.py", line 725 in capture_model File "/root/.conda/envs/py39/lib/python3.9/site-packages/torch/utils/_contextlib.py", line 115 in decorate_context File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/worker/worker.py", line 160 in warm_up_model File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/engine/llm_engine.py", line 1006 in _run_workers File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/engine/llm_engine.py", line 360 in _init_cache File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/engine/llm_engine.py", line 123 in init File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/engine/async_llm_engine.py", line 366 in _init_engine File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/engine/async_llm_engine.py", line 321 in init File "/root/.conda/envs/py39/lib/python3.9/site-packages/vllm/engine/async_llm_engine.py", line 625 in from_engine_args File "/apdcephfs_cq10/share_1567347/share_info/xli/Coder/service/dev/service/model_infer.py", line 268 in

chuanldong commented 7 months ago

Was it finally resolved?

github-actions[bot] commented 2 weeks ago

This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!