Open alterge1st opened 1 week ago
When the size of channel ctrlStream(ctrlStreamBufLen) is increased, we need to create more watch connections and disconnect them to reproduce the blocking problem.
We recently fixed a watch related goroutine leak issue, https://github.com/etcd-io/etcd/pull/18784
The fix will be included in 3.5.17, which is supposed to be released this week. Please try again with the new version once available.
We recently fixed a watch related goroutine leak issue, #18784
The fix will be included in 3.5.17, which is supposed to be released this week. Please try again with the new version once available.
Okay, thanks, we'll try the new version
We recently fixed a watch related goroutine leak issue, #18784
The fix will be included in 3.5.17, which is supposed to be released this week. Please try again with the new version once available.
Unfortunately, I modified my local code to follow the latest changes, but the problem persists.
Not following whether the issue is etcd or K8s related. In the repro you provide and discuss code for kubernetes API while the folowing debugging is about etcd. I would like to clarify this, because K8s apiserver demultiplexes watch connections to etcd. So the issue with client cancelation should not happen for K8s, as 100 watches opened to apiserver still opens only 1 watch to etcd.
Not following whether the issue is etcd or K8s related. In the repro you provide and discuss code for kubernetes API while the folowing debugging is about etcd. I would like to clarify this, because K8s apiserver demultiplexes watch connections to etcd. So the issue with client cancelation should not happen for K8s, as 100 watches opened to apiserver still opens only 1 watch to etcd.
Kube-apiserver is integrated with etcd. The in-process server of etcd is used to directly invoke APIs instead of using etcd service ports. This problem occurs when the watch command of K8s is used. When another process is started to create a large number of watch requests for the configmap cyclically, killing the process will cause the watch command of the configmap by the Kubernetes to become invalid.
The in-process server of etcd is used to directly invoke APIs instead of using etcd service ports.
This should not change the fact that apiserver will demultiplexes watch, or have you disabled watch cache?
The in-process server of etcd is used to directly invoke APIs instead of using etcd service ports.
This should not change the fact that apiserver will demultiplexes watch, or have you disabled watch cache?
Yes, we did disable the watch cache.
If ctrlStreamBufLen is set to hundreds, will etcd run properly?
Could anyone create a test using etcd client SDK instead of k8s client-go to reproduce this?
Bug report criteria
What happened?
We used the in-process etcdserver of v3client. Then we created a client, created a watch connection to the same resource every second, without freeing them, and ran it for more than 1 minute. When the client maintains a large number of watch connections, we kill the client process. After the client process is killed, when other clients attempt to establish watch connections for the same resource, the new watch connections cannot obtain new event changes.
What did you expect to happen?
After the client is killed, the new watch connection for the same resource can properly listen to event changes. And after analysis, the blocking problem exists. Although it is unreasonable for the client to establish a large number of watch connections with the same resource at the same time, can the etcd server do something to avoid the blocking?
How can we reproduce it (as minimally and precisely as possible)?
We created a large number of Watch connections to the same configmap resource in a loop through a process using code similar to the following: main.txt After running this program for 1 minute, kill the program. When you continue to run the kubectl get configmap -A -w command, after the configmap is modified, the configmap change cannot be watched.
Anything else we need to know?
After the client is killed, a large number of watch connections are disconnected. The code analysis shows that the
Send()
function ofWatchCancelRequest
incase ws := <-w.closingc
of the(w *watchGrpcStream) run()
method inetcd/client/v3/watch.go
is blocked and unable to continue processing. It is suspected that a large number ofWatchCancelRequests
cause the channel inwatchGrpcStream
to be fully occupied. As a result, newWatchResponse
cannot be pushed intosws.ctrlStream
. TheWatchResponse
obtained fromctrlStream
and newWatchResponse
are blocked incase pbresp := <-w.respc
andcase ws := <-w.closingc
in(w *watchGrpcStream) run()
.Etcd version (please run commands below)
Etcd configuration (command line flags or environment variables)
Etcd debug information (please run commands below, feel free to obfuscate the IP address or FQDN in the output)
Relevant log output
No response