Open wilsoniya opened 9 months ago
Hrmm, so in a well-running multiprocess setup, a heartbeat should be immediate and barely queued. Because the primary process should be processing the cross-process heartbeat almost immediately.
So the TimeoutError arises not from queue polling, but from executing a coroutine on an event loop.
Well, it's more like from the lack of queue polling. I believe if the heartbeat queue is not being polled/processed fast enough, this is where this call gets hung.
As to why the Future.result(10) call times out, I can't understand why.
The only reason why I can think this would fail is if your primary process is not processing asyncio quickly enough. This often happens if you're doing a blocking call in some async def
somewhere but I suppose could be caused by sending too many heartbeats too fast (or a combination of both things).
I am not sure there is much that can be done here. I believe the worker is not processing heartbeats faster than you are sending them and at some point it's a timeout. We can't queue heartbeats indefinitely. Open to suggestions though. For whatever suggestion, we likely need a replication to even confirm it solves your issue.
Alternatively, you may consider standard threaded (or async) activities that you, inside the activity, invoke multiprocessing.
Thanks for replying and generally being so responsive!
The only reason why I can think this would fail is if your primary process is not processing asyncio quickly enough. This often happens if you're doing a blocking call in some async def somewhere but I suppose could be caused by sending too many heartbeats too fast (or a combination of both things).
I suspect you're right. I'm using Pydantic 2.x, and using the custom data converter recipe you provide. The activity with frequent heartbeats is sending the details parameter, thus invoking the data converter (if I understand correctly). Pydantic 2.x appears to have some serious unresolved perf issues, so I'm wondering if the primary process is bogged down doing Pydantic serde of heartbeat details. I'm going to experiment with omitting heartbeat details and see if that resolves the issue.
I am not sure there is much that can be done here. I believe the worker is not processing heartbeats faster than you are sending them and at some point it's a timeout. We can't queue heartbeats indefinitely. Open to suggestions though. For whatever suggestion, we likely need a replication to even confirm it solves your issue.
Is this return
call potentially problematic?
It seems like returning from _MultiprocessingSharedStateManager._heartbeat_processor()
when we catch a TimeoutError
when evaluating the heartbeat future here
will result in processing no further heartbeats, and thus the heartbeat queue filling up and blocking further activity liveness. I have an incomplete understanding of this code, but I suspect that the error I'm describing is unrecoverable.
Is this return call potentially problematic?
I fear it will encourage lossy heartbeats as opposed to just stopping the processor. If we had an easy way to surface the exception we would, but in the meantime I am worried about just accepting some heartbeats and not others
will result in processing no further heartbeats, and thus the heartbeat queue filling up and blocking further activity liveness
Yes, this is a dangerous state to get in. I am not exactly sure what we should do when the shared heartbeat processor fails. It should never fail. I do want to avoid adding some kind of advanced code (e.g. draining just this specific activity's heartbeat queue somehow) to this never-should-happen failure state. Really it shouldn't even be a queue, it should be almost immediate where the queue never has a size more than one.
Open to suggestions about how to better surface this. Should I try to find a way to throw an exception and shut the whole worker down?
Sorry for the delayed response.
Should I try to find a way to throw an exception and shut the whole worker down?
That would be ideal for my use case. But more generally, I think encountering situations one has consciously conceived of as "impossible" merits a hard crash. And given a choice between a process crashing or technically persisting in a dysfunctional state, I think the former is more conducive to detecting and recovering from a problem.
Thanks!
With Temporal, we often favor partially disfunctional workers that still work for other things (e.g. things that aren't heartbeating and workflows) over fail-fast. But maybe we can make it an option.
But I am now wondering if we should deprecate multiprocess activities and instead tell users they can use normal Python subprocessing in their activities if they must have separate processes. It doesn't have much value in the SDK anymore than any other form of remote activity would. The details of dealing with out-of-process activities is probably best left to the (advanced) user.
What are you really trying to do?
Operate a heavily-loaded multiprocess worker.
Describe the bug
In a heavily-loaded multiprocess worker with activities that may possibly be issuing heartbeats too frequently, I observe logging indicating the parent process times out reading from the
_MultiprocessingSharedStateManager._heartbeat_queue
thus the worker tree effectively hangs, i.e., exhibits no further liveness in its ability to process tasks. If this situation is unrecoverable, I would expect the worker process tree to terminate, thus allowing a restart (i.e., via systemd, k8s policy, etc.).It's not clear to me why this timeout is happening, but it also results in the heartbeat queue filling up and blocking further progress on activities which heartbeat.
I see the following logging. First, presumably logs emitted by the parent of the process tree (i.e., the reader of the heartbeat queue):
And second, from the a child in the process tree, as it attempts to push heartbeats to the queue:
Minimal Reproduction
Unfortunately, I have no reproduction, but I do have some analysis to share:
I think the log message
Failed during multiprocess queue poll for heartbeat
is potentially misleading, as raised here on the topException
type. Digging into the first stack trace above, the exception is aTimeoutError
and the point of origin appears to be here: https://github.com/temporalio/sdk-python/blob/1.5.0/temporalio/worker/_activity.py#L644 - the body of that expression ends withFuture.result(10)
, which can raiseTimeoutError
. So theTimeoutError
arises not from queue polling, but from executing a coroutine on an event loop. As to why theFuture.result(10)
call times out, I can't understand why.Environment/Versions
5.4.0-42-generic #46-Ubuntu SMP Fri Jul 10 00:24:02 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux