Open electroma opened 11 months ago
Yes. We are aware of this gap. We expect having these metrics start to flow by Jan 2024. We will update this issue when we have a better ETA.
Hey @vinisoto. Is there any ETA on this ?
@vinisoto , Any update on when we can expect this feature to be in GA ?
@vinisoto Any update? :D
@vinisoto We are also looking forward to this feature very much since we would like to gain more insight into our ACA jobs.
Hey @vinisoto. We are also looking forward to this feature very much since we would like to gain more insight into our ACA jobs.
@vinisoto is it possible to give us an ETA?
hi sorry for the lack of updates. We hit some delays on our release. We are targetting for these metrics to be available by the week of 4/22 or sooner.
Waiting for this feature please
Is there an update on progress with this?
Any update on this?
ETA?
also waiting
Any news on this?
This is a vital enabler for the automation of failure monitoring. The offering seems incomplete without metrics.
I'm waiting for updates on this feature.
Insights to metrics are vital for us, in order to do right sizing and consolidations of Dedicated Plans.
I see the Otel integration is in preview, will Otel integration get the System metrics for ACA Jobs?
hi - the metrics UX has been enabled for Jobs:
we currently support:
cc: @anthonychu
Thank you for the update @vinisoto. In the original request I have requested one more important metric - Error Ratio. Would it be possible to compute this metric by splitting "Job Executions" on State?
@vinisoto when can we except this to be rolled out for all jobs? At least for my subscription its not present yet.
@rbange The metrics blade should be available everywhere now and should appear as in the screenshot that @vinisoto shared. Could you please check again?
@anthonychu when I check our Job metrics, all 86 jobs report a stable 0.05nc CPU usage and 536.9 MB on average, min and max aggregation.
Doesn't seem to be right, I wouldn't expect a flatline and the same values for all jobs.
@anthonychu Yes the metrics appeared approx 3 days later. However interestingly they are not selectable when trying to access them via scope in the metrics section. There only the regular resources appear, so you basically have to manually go to each job to check them...
Also I can report the same issue as @dinoo. Ram is always maxed out even though they require far less locally and CPU Usage is stuck at around 0.0x nc in min/max and average aggregation. I have a job which runs approx 20 minutes at each full hour and a flat line is absolutely unrealistic.
@anthonychu when I check our Job metrics, all 86 jobs report a stable 0.05nc CPU usage and 536.9 MB on average, min and max aggregation.
Doesn't seem to be right, I wouldn't expect a flatline and the same values for all jobs.
I am seeing the exact same numbers as @dinoo - 0.05nc and 536.9MB average. Something seems wrong...
Thanks all for reporting. We'll investigate.
Can you clarify what does the metric - Number of job executions mean? I thought it just counted every time the job is started, but as I can see I have more than 7 million executions in the past 30 days on a event driven job that approx. runs 30-50 times a month.
Hi, thanks for reporting. Some container apps jobs are not displaying the correct runtime value of the CPU and Memory metrics. We are preparing a configuration change and will be rolling it out in the next few days.
Regarding Job Executions: This metric displays the current number of job executions in an environment (Consistent with the execution list displayed in the Execution History blade). When a job execution ends, it remains in the environment (currently the last 100 successful and 100 failed executions are kept around). One way to visualize executions in time is to apply a split by Execution Name:
We are working on two new metrics: Executions Started and Executions Ended to display how many jobs executions start/end at a point in time. Will communicate here once we have a timeline to share.
Regarding Job Executions: This metric displays the current number of job executions in an environment (Consistent with the execution list displayed in the Execution History blade). When a job execution ends, it remains in the environment (currently the last 100 successful and 100 failed executions are kept around). One way to visualize executions in time is to apply a split by Execution Name:
This is not working on my Container App Jobs. It runs every hour and I definitely see the runs in the Execution History blade, but nothing in the Metrics section:
Are there any plans to make metrics accessible outside the "Container App Job | Metrics" tab?
I can access them through the ACA Job, but not through Metrics or Dashboard Hub.
While selecting the scope I can't see any Container App Jobs even though they are in that Resource Group.
[edit]
Container App Jobs Metrics also can't be selected within "Monitor | Alerts" which would be a great thing to have.
@rodyvansambeek - there was a regression that caused Jobs metrics to stop showing for some customers. We are rolling out a fix which will be fully deployed to all regions by end of this week.
@xdawxd - we are in the process of fixing both issues: Jobs being available outside of the Job Metrics blade (for example: making Jobs available as a Metrics scope) and being able to create alerts based on Jobs metrics.
We will update here when we have ETAs for both fixes.
Is your feature request related to a problem? Please describe.
Regular ACAs have a good set of built-in metrics: replica count, CPU and memory utilization. I can't find anything like that for ACA Jobs.
Describe the solution you'd like.
Make the following ACA Job metrics available:
Describe alternatives you've considered.
I do not see any alternative, and I believe metrics should be published, so the teams can plugin proper monitoring.