Hello, Performance evaluation of Linux Kernel v6.10-rc3 is finished. Comparison is done to baseline Linux Kernel version: v6.9. The following configuration was used for all targets: - Instance CPU: graviton3 - Number of cores: 2 - RAM size, GB: 7.6 - OS: Debian Linux 12 (bookworm)
Regressions by test
Benchmark/Test/Metric, Units | Baseline Median | Current Median | Diff, % | SD | Samples | Confidence interval ------------------------------------------------------------------------------------------------------------------------------------------------------ aim9/brk_test/System Memory Allocations, OPS | 2024133.335 | 2013262.470 | 0.537 | 16053.459 | 60 | 2024467.853 / 2032591.875 aim9/exec_test/Program Loads, OPS | 2019.665 | 2005.835 | 0.685 | 17.517 | 60 | 2015.828 / 2024.693 aim9/page_test/System Allocations & Pages, OPS | 570520.000 | 550091.667 | 3.581 | 8356.271 | 60 | 568085.990 / 572314.769 stressng/context/Context Switches, switches/s - context-1 | 2413.745 | 2376.445 | 1.545 | 5.545 | 21 | 2411.760 / 2416.503 - context-2 | 4841.600 | 4769.580 | 1.488 | 8.553 | 21 | 4836.387 / 4843.704 - context-4 | 4752.505 | 4712.600 | 0.840 | 56.213 | 21 | 4716.312 / 4764.397 - context-8 | 4753.285 | 4724.605 | 0.603 | 42.188 | 21 | 4728.812 / 4764.899 - context-16 | 4753.605 | 4726.140 | 0.578 | 26.756 | 21 | 4740.173 / 4763.060 - context-128 | 4814.425 | 4774.555 | 0.828 | 43.430 | 21 | 4786.041 / 4823.191 stressng/fork/Forks, OPS - fork-32 | 6659.030 | 6412.310 | 3.705 | 174.515 | 21 | 6588.293 / 6737.573 - fork-64 | 6516.635 | 6200.470 | 4.852 | 266.061 | 21 | 6394.023 / 6621.612 - fork-128 | 6128.850 | 5843.435 | 4.657 | 353.797 | 21 | 6005.178 / 6307.815 stressng/get/Read Throughput, MB/s - get-1 | 2835.125 | 2785.405 | 1.754 | 25.632 | 21 | 2823.417 / 2845.342 - get-2 | 3295.495 | 3250.850 | 1.355 | 50.308 | 21 | 3271.035 / 3314.069 - get-16 | 3097.905 | 3062.850 | 1.132 | 36.459 | 21 | 3081.725 / 3112.912 stressng/madvise/Access calls, OPS - madvise-1 | 39.845 | 37.240 | 6.538 | 0.864 | 21 | 39.425 / 40.164 - madvise-2 | 76.950 | 72.480 | 5.809 | 2.384 | 21 | 76.016 / 78.055 - madvise-4 | 86.900 | 83.080 | 4.396 | 1.283 | 21 | 86.317 / 87.414 - madvise-8 | 87.615 | 82.920 | 5.359 | 1.634 | 21 | 86.943 / 88.341 - madvise-16 | 87.230 | 82.345 | 5.600 | 1.879 | 21 | 86.472 / 88.079 - madvise-32 | 85.895 | 80.780 | 5.955 | 1.543 | 21 | 85.294 / 86.614 - madvise-64 | 85.955 | 80.625 | 6.201 | 2.257 | 21 | 84.615 / 86.546 - madvise-128 | 85.005 | 79.375 | 6.623 | 1.706 | 21 | 84.143 / 85.603 stressng/vm-splice/Transfer Rate, MB/s - vm-splice-1 | 393825.950 | 384976.075 | 2.247 | 1326.151 | 21 | 393146.946 / 394281.333 - vm-splice-2 | 985022.850 | 970009.675 | 1.524 | 3305.843 | 21 | 983522.945 / 986350.758 - vm-splice-4 | 1111278.105 | 1098156.625 | 1.181 | 5058.131 | 21 | 1108947.236 / 1113273.953 - vm-splice-8 | 1195749.630 | 1183949.325 | 0.987 | 5994.316 | 21 | 1192658.405 / 1197785.934 - vm-splice-16 | 1238939.755 | 1226909.580 | 0.971 | 7273.147 | 21 | 1235968.212 / 1242189.651 - vm-splice-32 | 1241984.390 | 1227662.560 | 1.153 | 5226.860 | 21 | 1239895.031 / 1244366.079 - vm-splice-64 | 1241623.855 | 1227060.990 | 1.173 | 5099.827 | 21 | 1239863.809 / 1244226.193 - vm-splice-128 | 1248609.720 | 1233938.390 | 1.175 | 7083.372 | 21 | 1245315.449 / 1251374.554
* index after test name indicate number of threads or processes used for the test.
Detailed test results and raw data can be found here: https://artifacts.codelinaro.org/artifactory/linaro-373-reports/aws-graviton...
Best regards.