Over time we are seeing a distinct increase in memory usage for the lvmd process (it doubles in about a week). It can be seen in the screenshots of memory usage over time. I have included information for 2x pods in the same kubernetes cluster, one which has been running for 8 days, and one that has been running for about 30 mins. It seems to be the "working set/wss" where we see the increase over time. I have also included pprof heap dumps for both pods. I have looked through the heap dumps and nothing stands out for me.
The NVMe disks on m6gd are being converted to a Logical volume, which is what's being provisioned by Topolvm, if it is relevant.
Describe the bug
Over time we are seeing a distinct increase in memory usage for the lvmd process (it doubles in about a week). It can be seen in the screenshots of memory usage over time. I have included information for 2x pods in the same kubernetes cluster, one which has been running for 8 days, and one that has been running for about 30 mins. It seems to be the "working set/wss" where we see the increase over time. I have also included pprof heap dumps for both pods. I have looked through the heap dumps and nothing stands out for me.
The NVMe disks on
m6gd
are being converted to a Logical volume, which is what's being provisioned by Topolvm, if it is relevant.Environments
v1.28.12-eks-a18cd3a
To Reproduce Steps to reproduce the behavior:
Expected behavior We would expect more of a flat line for memory usage.
Additional context
Pod running for 8 days:
topolvm-provisioner-lvmd-0-qx57c.heap.gz
Pod running for 30m:
topolvm-provisioner-lvmd-0-v4j4l.heap.gz
This was originally referenced in https://github.com/topolvm/topolvm/pull/931, for context. Happy to provide as much information as you may need, let me know.