Closed cjbottaro closed 1 year ago
Probably you are running into 13205. Could you try to reproduce this issue using the latest code in release-3.5 or main branches.
Please also provide the following info so that others can take a closer look.
@cjbottaro any update on this? Have you enabled auth in the cluster?
Ahh, I stopped using Etcd for locks and went back to using single node Redis for now. Will come back to Etcd if the need for HA ever arises though.
This issue has been automatically marked as stale because it has not had recent activity. It will be closed after 21 days if no further activity occurs. Thank you for your contributions.
What happened?
Etcd thinks these leases still exist...
Each lease has a ttl of 5s. We shutdown our system (no new leases created) and these leases still exist over 24h later.
Trying to revoke any of these leases results in an error saying they don't exist.
We had to shutdown our system because asking Etcd for a lease would result in a timeout. After a few hours of our system being shutdown, Etcd stopped timing out when asking for a lease.
What did you expect to happen?
Etcd to continue to grant leases under heavy load and not lose track of leases.
How can we reproduce it (as minimally and precisely as possible)?
Locking/unlocking unique locks at a rate of 80 per second, each using a lease with ttl=5. Letting that run for a day or so.
Anything else we need to know?
We use Etcd for distributed locking and nothing else. Each lease is 5s. We have a 3 node cluster running on a single Core i3-8100 machine. We are only requesting about 80 locks per second.
It seems this setting is reasonable given that our locks have a ttl of 5s.
Etcd version (please run commands below)
quay.io/coreos/etcd:v3.5.3
Etcd configuration (command line flags or environment variables)
Etcd debug information (please run commands blow, feel free to obfuscate the IP address or FQDN in the output)
Relevant log output
No response