Closed burling closed 4 weeks ago
I am getting a very similar problem, however I am not sure if it is the exact same error. I also have a python decupled backend. After starting tritonserver I run stress testing which sends a lot of requests to the tritonserver. Within the first 10 minutes of testing I quite consistently get this error which completely crushes my tritonserver. Unfortunately I have a custom build of tritonserver based on 24.05, so I don't know how relevant the information is.
E0702 19:02:48.658289 148148 infer_handler.h:187] "[INTERNAL] Attempting to access current response when it is not ready"
Signal (11) received.
0.773678183555603
0# 0x0000561EA6BD83ED in tritonserver
1# 0x00007F6E5E5D3090 in /usr/lib/x86_64-linux-gnu/libc.so.6
2# 0x0000561EA6C4DBE4 in tritonserver
3# 0x0000561EA6C4E740 in tritonserver
4# 0x0000561EA6C46DFA in tritonserver
5# 0x0000561EA6C31AB5 in tritonserver
6# 0x00007F6E5E9D4793 in /usr/lib/x86_64-linux-gnu/libstdc++.so.6
7# 0x00007F6E5EB64609 in /usr/lib/x86_64-linux-gnu/libpthread.so.0
8# clone in /usr/lib/x86_64-linux-gnu/libc.so.6
Segmentation fault (core dumped)
I assume the error occurs because of this check, however I have no clue why this is the case: https://github.com/triton-inference-server/server/blob/c61d993ff3e3608d9b76e9c89a65e5ce41497d49/src/grpc/infer_handler.h#L183-L192
Do you have any logs about Stub being unhealthy and being restarted prior to this Signal 11 crash ?
Closing due to in-activity. Please provide a full reproducer in case you're still running into this issue with the latest version.
Description After using the Python vllm backend, Triton crashed with signal 11. The model had been loaded and preheated for some time before the crash occurred.
Triton Information What version of Triton are you using?
Are you using the Triton container or did you build it yourself? Yes
trace info: