You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The benchmark-cli's performance detection workflow fails if the performance regression exceeds the threshold value. Currently, the threshold is set to 1, meaning that the workflow only fails when the new commit is twice as slow as the original one.
Need a better strategy to calculate the performance gap between the original commit and the new commit to generate a desired and accurate threshold value.
In the short term, we should increase the warm-up and iteration numbers to make sure the results become stable.
in the long term,
We should compare the new commit not only to the most recent one but also to the average value from the past months to prevent unnoticed false negatives from occurring.
We also need to reduce the noise caused by the machine. Currently, we only run performance regression detection on Ubuntu but we plan to add other OSs in the future. To ensure more consistent benchmark results on Linux, we can control factors such as CPU affinity - https://easyperf.net/blog/2019/08/02/Perf-measurement-environment-on-Linux
We should create more specific files that target specific data models. For example, if a new commit introduces changes that impact the speed of string serialization, it can be hard to find. By having files that include numerous test cases for Ion string values, we can magnify this regression and easily detect it.
We can also mimic JMH (Java Microbenchmark Harness) and implement another warm-up mode based on"clock time" instead of "execution numbers" which ensures that the benchmark-cli warm up for a specific duration. For example, instead of executing the specific benchmark code 20 times, we warm up the benchmark-cli for 20 minutes.
We can try running each command for both previous and current commits together and see if it has less variance (currently we are running all commands for each commit each time).
The benchmark-cli's performance detection workflow fails if the performance regression exceeds the threshold value. Currently, the threshold is set to 1, meaning that the workflow only fails when the new commit is twice as slow as the original one.
Need a better strategy to calculate the performance gap between the original commit and the new commit to generate a desired and accurate threshold value.
Related issue - amazon-ion/ion-java-benchmark-cli#53
The text was updated successfully, but these errors were encountered: