-
Notifications
You must be signed in to change notification settings - Fork 75
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AD model performance benchmark #724
Conversation
This PR adds an AD model performance benchmark so that we can compare model performance across versions. Regarding benchmark data, we randomly generated synthetic data with known anomalies inserted throughout the signal. In particular, these are one/two/four dimensional data where each dimension is a noisy cosine wave. Anomalies are inserted into one dimension with 0.003 probability. Anomalies across each dimension can be independent or dependent. We have approximately 5000 observations per data set. The data set is generated using the same random seed so the result is comparable across versions. We also backported opensearch-project#600 so that we can capture the performance data in CI output. Testing done: * added unit tests to run the benchmark. Signed-off-by: Kaituo Li <kaituo@amazon.com>
Codecov Report
@@ Coverage Diff @@
## 2.1 #724 +/- ##
============================================
- Coverage 79.23% 79.05% -0.18%
+ Complexity 4222 4212 -10
============================================
Files 296 296
Lines 17686 17686
Branches 1880 1880
============================================
- Hits 14013 13982 -31
- Misses 2779 2812 +33
+ Partials 894 892 -2
Flags with carried forward coverage won't be shown. Click here to find out more.
|
143442f
to
2869a74
Compare
Signed-off-by: Tyler Ohlsen <ohltyler@amazon.com>
2869a74
to
c95bd15
Compare
@@ -366,7 +379,7 @@ task integTestRemote(type: RestIntegTestTask) { | |||
} | |||
} | |||
|
|||
String bwcVersion = "1.13.0.0" | |||
String bwcVersion = "1.1.0.0" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for updating these too :)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
np :)
Description
This PR adds an AD model performance benchmark so that we can compare model performance across versions.
For the single stream detector, we refactored tests in DetectionResultEvalutationIT and moved it to SingleStreamModelPerfIT.
For the HCAD detector, we randomly generated synthetic data with known anomalies inserted throughout the signal. In particular, these are one/two/four dimensional data where each dimension is a noisy cosine wave. Anomalies are inserted into one dimension with 0.003 probability. Anomalies across each dimension can be independent or dependent. We have approximately 5000 observations per data set. The data set is generated using the same random seed so the result is comparable across versions.
We also backported #600 so that we can capture the performance data in CI output. We backported #625 to update bwc zip since odfe zip is unavaiable.
Testing done:
Signed-off-by: Kaituo Li kaituo@amazon.com
By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.
For more information on following Developer Certificate of Origin and signing off your commits, please check here.