Open hybby opened 8 years ago
Hi @hybby !
This RDS Aurora instance is probably missing a metric the check expects to be there, and it fails because of this. Could you enable debug logs and send a flare to support at datadoghq.com so that we can troubleshoot this further ?
I'm experiencing this issue as well, any resolution?
@degemer To add some more context to this, running SHOW /*!50000 ENGINE*/ INNODB STATUS \G
on an Aurora writer gives the usual innodb status output:
mysql> SHOW /*!50000 ENGINE*/ INNODB STATUS \G
*************************** 1. row ***************************
Type: InnoDB
Name:
Status:
=====================================
2016-09-23 04:59:38 2b10b7ec5700 INNODB MONITOR OUTPUT
=====================================
Per second averages calculated from the last 2 seconds
-----------------
BACKGROUND THREAD
-----------------
srv_master_thread loops: 2746859 srv_active, 0 srv_shutdown, 3179 srv_idle
srv_master_thread log flush and writes: 0
----------
SEMAPHORES
----------
.
.
.
While running the same command on an Aurora replica yeilds:
mysql> SHOW /*!50000 ENGINE*/ INNODB STATUS \G
Empty set (0.00 sec)
I know that we can disable innodb stats on replicas with version 5.7 or later but it seems like there should be a more graceful way to handle this problem.
I've configured the MySQL Integration against two instances of an AWS RDS Aurora database, which is MySQL compatible. The first is a writer (master) and is reporting fine. The second is a reader (read-replica) is returning a Python stack trace when invoking
dd-agent info
:In this case,
instance #11
is the writer (master) andinstance #12
is the reader (read-replica). I have other RDS instances configured up with this integration on this host, which is why the numbers are larger than you'd expect.Configuration file
/etc/dd-agent/conf.d/mysql.yaml
looks like so, with site-specific items removed:How can I troubleshoot this issue further?
As noted, the writer (master) instance seems to report as OK and send metrics just fine. It's the reader (read-replica) that's throwing this traceback.
Curious as to whether I should just configure this up pointing at the cluster endpoint rather than a configuration stanza for each node of the cluster.