Open ShaunMaxwell opened 1 year ago
@justinsb I remember you had looked into similar issues before and attempted a refactoring to make the cnrm-resource-stats-recorder
controller more performant. This looks to be related?
I have encountered the same issue with version 1.110. It appears that cnrm-resource-stats-recorder
is attempting to access v1alpha1
CRDs that need to be manually installed(https://cloud.google.com/config-connector/docs/how-to/install-alpha-crds).
I have encountered the same issue with version 1.110. It appears that
cnrm-resource-stats-recorder
is attempting to accessv1alpha1
CRDs that need to be manually installed(https://cloud.google.com/config-connector/docs/how-to/install-alpha-crds).
Is there any way to prevent the cnrm-resource-stats-recorder
from looking for the v1alpha1
CRDs? We do not want to maintain them manually. Right now our cnrm-resource-stats-recorder
is flooding our log system with thousands of error logs when looking for these CRDs every minute.
Checklist
Bug Description
The
configconnector_applied_resources_total
metric is not always available. It appears for ~8 minutes and then disappears for ~10 minutes.I have tried using
kubectl port-forward
to expose the pod onlocalhost
and usedcurl
to retrieve the metrics myself and confirmed that the whole series ofconfigconnector_applied_resources_total
metrics disappears for a few minutes.I suspect the disappearance is due to the resetting of the metric and the length of time it is unavailable for is due to client-side throttling of the Kubernetes API requests causing a complete run of the metrics endpoint to take a long time.
In the logs of the
cnrm-resource-stats-recorder
pod, there are a lot of throttling messages and errors related to missing resources for some CRDs.Additional Diagnostic Information
None
Kubernetes Cluster Version
v1.25.10-gke.2700
Config Connector Version
1.108.0
Config Connector Mode
cluster mode
Log Output
Steps to reproduce the issue
configconnector_applied_resources_total
metricYAML snippets
No response