Closed wandgitlabbot closed 3 years ago
In GitLab, by Daniel Oosterwijk on 2020-08-23
I'm considering using some form of automated parameter tuning to see if I can reduce the workload of doing this manually. This would involve a fair bit of work before it functions, but is likely to be useful on other workloads once complete.
Currently, we have an entrypoint that runs all the detectors against the NAB dataset. We then do some post-processing of the results in Python before running it against the NAB scorer, also in Python (although we run the scorer via a Bash script). Automating this would involve the following steps:
I'm going to spin these steps off into several issues, and maybe even try out Gitlab's Milestones feature.
In GitLab, by Daniel Oosterwijk on 2021-01-21
I've finished parsing the logs for the final optimisation run, and talked about it in the wiki at Parameter Tuning Results. I'll include the script I used to run the tests in the repo, and close the issue.
In GitLab, by Daniel Oosterwijk on 2020-08-21
The existing detectors (baseline, changepoint, distdiff, mode, spike) have been tested against the NAB dataset, but the results were not very impressive. Three of the five detectors produced no events, and the two that did had a low accuracy. I believe that these scores could be improved by tuning the detectors' configurations to the dataset. NAB testing mandates that each detector must have a single config for the entire dataset, and that detectors may not look ahead at what's to come to pre-tune themselves. Luckily, that's how we're already set up :)
The results are as follows:
The official NAB scoreboard as of the time of writing is replicated below: