[mm/filemap.c] 06c0444290: stress-ng.sendfile.ops_per_sec 26.7% improvement
by kernel test robot
Greeting,
FYI, we noticed a 26.7% improvement of stress-ng.sendfile.ops_per_sec due to commit:
commit: 06c0444290cecf04c89c62e6d448b8461507d247 ("mm/filemap.c: generic_file_buffered_read() now uses find_get_pages_contig")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
in testcase: stress-ng
on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
with following parameters:
nr_threads: 100%
disk: 1HDD
testtime: 30s
class: pipe
cpufreq_governor: performance
ucode: 0x5003003
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
pipe/gcc-9/performance/1HDD/x86_64-rhel-8.3/100%/debian-10.4-x86_64-20200603.cgz/lkp-csl-2sp5/stress-ng/30s/0x5003003
commit:
723ef24b9b ("mm/filemap/c: break generic_file_buffered_read up into multiple functions")
06c0444290 ("mm/filemap.c: generic_file_buffered_read() now uses find_get_pages_contig")
723ef24b9b379e59 06c0444290cecf04c89c62e6d44
---------------- ---------------------------
%stddev %change %stddev
\ | \
14865658 +26.7% 18839172 ± 2% stress-ng.sendfile.ops
495501 +26.7% 627957 ± 2% stress-ng.sendfile.ops_per_sec
17943 ± 12% +36.5% 24500 ± 5% proc-vmstat.numa_hint_faults
1585 ±170% +3019.7% 49447 ± 63% numa-numastat.node0.other_node
85104 ± 3% -56.2% 37244 ± 83% numa-numastat.node1.other_node
169349 ± 19% -21.8% 132475 meminfo.AnonHugePages
301479 ± 10% -11.8% 265842 meminfo.AnonPages
333754 ± 9% -10.6% 298511 meminfo.Inactive
333754 ± 9% -10.6% 298511 meminfo.Inactive(anon)
11540 ± 12% -15.2% 9785 ± 3% sched_debug.cfs_rq:/.load.avg
17531 ± 77% -77.9% 3878 ± 9% sched_debug.cfs_rq:/.load.stddev
28103 ± 22% +50.3% 42227 ± 30% sched_debug.cpu.avg_idle.min
6188 ± 20% +38.9% 8595 ± 4% sched_debug.cpu.curr->pid.min
495.48 ± 24% -46.2% 266.34 ± 18% sched_debug.cpu.curr->pid.stddev
3.336e+10 -5.6% 3.148e+10 ± 6% perf-stat.i.branch-instructions
0.03 ± 7% +0.0 0.04 ± 43% perf-stat.i.dTLB-load-miss-rate%
0.01 ± 18% +0.0 0.01 ± 10% perf-stat.i.dTLB-store-miss-rate%
6253 ± 3% -18.2% 5117 ± 9% perf-stat.i.instructions-per-iTLB-miss
0.64 +0.0 0.67 perf-stat.overall.branch-miss-rate%
3.264e+10 -5.5% 3.084e+10 ± 5% perf-stat.ps.branch-instructions
0.01 ± 5% +13.5% 0.01 ± 2% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64
0.05 ± 83% +1464.2% 0.85 ±157% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
494.00 ± 4% +734.5% 4122 ± 3% perf-sched.wait_and_delay.count.preempt_schedule_common._cond_resched.__splice_from_pipe.splice_from_pipe.direct_splice_actor
5392 ± 3% -98.4% 87.00 ± 10% perf-sched.wait_and_delay.count.preempt_schedule_common._cond_resched.generic_file_buffered_read.generic_file_splice_read.splice_direct_to_actor
18.80 ± 20% -68.8% 5.86 ± 73% perf-sched.wait_and_delay.max.ms.preempt_schedule_common._cond_resched.generic_file_buffered_read.generic_file_splice_read.splice_direct_to_actor
0.59 ± 24% -45.0% 0.32 ± 22% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.path_openat
18.80 ± 20% -68.8% 5.86 ± 73% perf-sched.wait_time.max.ms.preempt_schedule_common._cond_resched.generic_file_buffered_read.generic_file_splice_read.splice_direct_to_actor
7626 -38.2% 4716 ± 33% interrupts.CPU30.NMI:Non-maskable_interrupts
7626 -38.2% 4716 ± 33% interrupts.CPU30.PMI:Performance_monitoring_interrupts
18521 ± 44% +241.7% 63289 ± 80% interrupts.CPU42.RES:Rescheduling_interrupts
30879 ± 63% +91.0% 58971 ± 31% interrupts.CPU49.RES:Rescheduling_interrupts
37970 ± 19% +115.4% 81806 ± 33% interrupts.CPU5.CAL:Function_call_interrupts
48131 ± 35% -55.7% 21307 ± 23% interrupts.CPU65.RES:Rescheduling_interrupts
33689 ± 39% +186.7% 96598 ± 88% interrupts.CPU7.CAL:Function_call_interrupts
37234 ± 52% +76.5% 65709 ± 45% interrupts.CPU71.CAL:Function_call_interrupts
22154 ± 18% +126.8% 50249 ± 70% interrupts.CPU82.RES:Rescheduling_interrupts
16632 ± 60% +310.7% 68311 ± 51% interrupts.CPU9.CAL:Function_call_interrupts
17920 ± 45% +264.1% 65238 ± 48% interrupts.CPU95.CAL:Function_call_interrupts
83.19 -3.7 79.44 ± 5% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
83.28 -3.8 79.53 ± 5% perf-profile.children.cycles-pp.do_syscall_64
0.68 ± 3% +0.0 0.71 ± 2% perf-profile.children.cycles-pp.sched_clock
0.32 ± 12% +0.1 0.39 ± 6% perf-profile.children.cycles-pp.set_next_buddy
0.00 +0.1 0.07 ± 12% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
1.31 +0.1 1.39 ± 3% perf-profile.children.cycles-pp.__update_load_avg_se
3.41 +0.2 3.56 ± 2% perf-profile.children.cycles-pp.switch_mm_irqs_off
0.08 +0.0 0.10 ± 5% perf-profile.self.cycles-pp.perf_trace_run_bpf_submit
0.08 ± 10% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.__bitmap_and
0.28 ± 9% +0.1 0.33 ± 9% perf-profile.self.cycles-pp.pipe_poll
0.93 +0.1 0.99 perf-profile.self.cycles-pp.switch_mm_irqs_off
0.29 ± 13% +0.1 0.36 ± 7% perf-profile.self.cycles-pp.set_next_buddy
1.28 +0.1 1.36 ± 3% perf-profile.self.cycles-pp.__update_load_avg_se
9207 ± 21% -40.3% 5500 ± 26% softirqs.CPU10.SCHED
8966 ± 22% -36.9% 5653 ± 18% softirqs.CPU11.SCHED
9107 ± 21% -41.4% 5341 ± 19% softirqs.CPU12.SCHED
8925 ± 23% -41.6% 5215 ± 14% softirqs.CPU13.SCHED
9040 ± 21% -41.5% 5285 ± 16% softirqs.CPU14.SCHED
8972 ± 23% -40.6% 5327 ± 16% softirqs.CPU15.SCHED
9021 ± 22% -41.5% 5279 ± 13% softirqs.CPU16.SCHED
8932 ± 20% -40.0% 5357 ± 15% softirqs.CPU17.SCHED
8870 ± 21% -39.0% 5409 ± 20% softirqs.CPU18.SCHED
8866 ± 23% -39.9% 5330 ± 19% softirqs.CPU19.SCHED
9042 ± 21% -37.1% 5683 ± 16% softirqs.CPU2.SCHED
8898 ± 22% -40.7% 5274 ± 20% softirqs.CPU20.SCHED
8989 ± 22% -39.8% 5412 ± 18% softirqs.CPU21.SCHED
8876 ± 22% -41.2% 5223 ± 19% softirqs.CPU22.SCHED
8892 ± 19% -38.7% 5455 ± 16% softirqs.CPU23.SCHED
6924 ± 31% +58.8% 10997 ± 7% softirqs.CPU24.SCHED
7098 ± 30% +64.3% 11663 ± 6% softirqs.CPU25.SCHED
7040 ± 32% +63.8% 11534 ± 7% softirqs.CPU26.SCHED
6977 ± 32% +63.6% 11416 ± 7% softirqs.CPU27.SCHED
7088 ± 30% +58.8% 11255 ± 8% softirqs.CPU28.SCHED
6857 ± 33% +65.5% 11352 ± 6% softirqs.CPU29.SCHED
9142 ± 25% -41.4% 5358 ± 19% softirqs.CPU3.SCHED
7061 ± 29% +62.5% 11472 ± 7% softirqs.CPU30.SCHED
6878 ± 30% +65.2% 11362 ± 8% softirqs.CPU31.SCHED
7173 ± 30% +64.9% 11828 ± 5% softirqs.CPU32.SCHED
7013 ± 31% +62.6% 11405 ± 8% softirqs.CPU33.SCHED
14139 ± 27% +30.8% 18492 ± 24% softirqs.CPU34.RCU
7033 ± 32% +58.8% 11166 ± 7% softirqs.CPU34.SCHED
6963 ± 29% +61.5% 11248 ± 7% softirqs.CPU35.SCHED
7012 ± 30% +61.6% 11332 ± 9% softirqs.CPU36.SCHED
6923 ± 29% +63.4% 11310 ± 8% softirqs.CPU37.SCHED
7070 ± 32% +59.8% 11298 ± 8% softirqs.CPU38.SCHED
6818 ± 31% +67.0% 11389 ± 7% softirqs.CPU39.SCHED
9088 ± 20% -42.2% 5250 ± 15% softirqs.CPU4.SCHED
7040 ± 29% +61.5% 11368 ± 7% softirqs.CPU40.SCHED
6980 ± 29% +62.2% 11321 ± 7% softirqs.CPU41.SCHED
6926 ± 29% +64.3% 11379 ± 8% softirqs.CPU42.SCHED
7062 ± 30% +57.4% 11114 ± 7% softirqs.CPU43.SCHED
6960 ± 31% +62.1% 11279 ± 9% softirqs.CPU44.SCHED
6854 ± 31% +65.0% 11310 ± 7% softirqs.CPU45.SCHED
7152 ± 29% +58.9% 11362 ± 9% softirqs.CPU46.SCHED
6828 ± 31% +64.2% 11210 ± 8% softirqs.CPU47.SCHED
8903 ± 21% -41.9% 5170 ± 17% softirqs.CPU48.SCHED
9048 ± 21% -41.0% 5336 ± 19% softirqs.CPU49.SCHED
8974 ± 23% -42.9% 5124 ± 18% softirqs.CPU5.SCHED
8742 ± 23% -40.8% 5177 ± 17% softirqs.CPU50.SCHED
8647 ± 23% -38.3% 5335 ± 19% softirqs.CPU51.SCHED
8783 ± 22% -41.7% 5118 ± 10% softirqs.CPU52.SCHED
8659 ± 26% -40.2% 5175 ± 18% softirqs.CPU53.SCHED
8852 ± 23% -40.2% 5292 ± 15% softirqs.CPU54.SCHED
9070 ± 20% -43.2% 5153 ± 16% softirqs.CPU55.SCHED
9191 ± 17% -42.7% 5266 ± 15% softirqs.CPU56.SCHED
8884 ± 24% -41.8% 5171 ± 16% softirqs.CPU57.SCHED
8986 ± 23% -40.5% 5344 ± 18% softirqs.CPU58.SCHED
9501 ± 24% -44.9% 5233 ± 20% softirqs.CPU59.SCHED
8897 ± 21% -38.5% 5467 ± 18% softirqs.CPU6.SCHED
9260 ± 18% -42.4% 5335 ± 22% softirqs.CPU60.SCHED
8966 ± 21% -42.3% 5170 ± 15% softirqs.CPU61.SCHED
8963 ± 22% -39.2% 5454 ± 16% softirqs.CPU62.SCHED
8948 ± 22% -39.0% 5462 ± 17% softirqs.CPU63.SCHED
8980 ± 22% -41.1% 5289 ± 16% softirqs.CPU64.SCHED
8969 ± 20% -41.4% 5260 ± 15% softirqs.CPU65.SCHED
8891 ± 23% -41.4% 5211 ± 20% softirqs.CPU66.SCHED
9193 ± 23% -40.0% 5520 ± 14% softirqs.CPU67.SCHED
8936 ± 24% -40.7% 5296 ± 16% softirqs.CPU68.SCHED
8871 ± 22% -40.0% 5320 ± 16% softirqs.CPU69.SCHED
8962 ± 19% -41.8% 5216 ± 17% softirqs.CPU7.SCHED
8671 ± 24% -38.7% 5315 ± 19% softirqs.CPU70.SCHED
7198 ± 29% +53.9% 11076 ± 4% softirqs.CPU72.SCHED
7133 ± 29% +61.0% 11488 ± 9% softirqs.CPU73.SCHED
6952 ± 30% +66.9% 11602 ± 8% softirqs.CPU74.SCHED
6975 ± 31% +60.8% 11214 ± 7% softirqs.CPU75.SCHED
6985 ± 31% +58.4% 11065 ± 10% softirqs.CPU76.SCHED
6811 ± 31% +63.6% 11146 ± 6% softirqs.CPU77.SCHED
7006 ± 29% +62.0% 11347 ± 7% softirqs.CPU78.SCHED
6827 ± 32% +65.8% 11316 ± 9% softirqs.CPU79.SCHED
8957 ± 19% -40.8% 5304 ± 18% softirqs.CPU8.SCHED
7102 ± 32% +59.7% 11345 ± 8% softirqs.CPU80.SCHED
7023 ± 30% +60.3% 11258 ± 8% softirqs.CPU81.SCHED
7046 ± 31% +57.8% 11121 ± 6% softirqs.CPU82.SCHED
6966 ± 30% +57.1% 10941 ± 8% softirqs.CPU83.SCHED
6953 ± 30% +62.0% 11261 ± 10% softirqs.CPU85.SCHED
6884 ± 31% +63.0% 11220 ± 9% softirqs.CPU86.SCHED
6765 ± 32% +66.3% 11249 ± 8% softirqs.CPU87.SCHED
6963 ± 29% +63.8% 11403 ± 7% softirqs.CPU88.SCHED
6869 ± 31% +63.6% 11241 ± 8% softirqs.CPU89.SCHED
9002 ± 21% -43.2% 5115 ± 18% softirqs.CPU9.SCHED
6759 ± 33% +69.4% 11450 ± 9% softirqs.CPU90.SCHED
7003 ± 30% +58.9% 11130 ± 9% softirqs.CPU91.SCHED
6994 ± 28% +64.9% 11533 ± 9% softirqs.CPU92.SCHED
6839 ± 32% +66.1% 11358 ± 6% softirqs.CPU93.SCHED
7121 ± 29% +59.2% 11339 ± 7% softirqs.CPU94.SCHED
6798 ± 32% +67.9% 11412 ± 7% softirqs.CPU95.SCHED
stress-ng.sendfile.ops
2e+07 +-----------------------------------------------------------------+
| O OO OO |
1.9e+07 |-+O OOOOOOO OO OOOO |
|OOOO O O OOO O |
| O O O O |
1.8e+07 |-+ |
| |
1.7e+07 |-+ |
| |
1.6e+07 |-+ |
| |
| |
1.5e+07 |+++++ + ++++++++++ +++++++++++ ++++++ +++++++ ++++++++++ + +++++|
|+ ++++ + +++ +++++ + ++ +++++ ++++++++ +++++ + + ++++ + |
1.4e+07 +-----------------------------------------------------------------+
stress-ng.sendfile.ops_per_sec
660000 +------------------------------------------------------------------+
640000 |-+ OO OOOOO O |
|OOOOOOOO O OOOOOOO |
620000 |-+OO O |
600000 |-+ O O O |
| |
580000 |-+ |
560000 |-+ |
540000 |-+ |
| |
520000 |-+ |
500000 |-+ + + + + + |
|++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++|
480000 |-+ ++ + + ++ + + + + + +++++ ++ + + ++ + + |
460000 +------------------------------------------------------------------+
7000 +--------------------------------------------------------------------+
| |
6000 |-+ + + + + + + |
|++++++:++++++++++++:+++++++++++::++++:+ + ++++++:+++++++++++::++++++|
5000 |-+ ++++ + ++ + ++ + + + ++ + ++++ + + ++ + + + ++ + ++ |
| |
4000 |-+ |
| |
3000 |-+ |
| |
2000 |-+ |
| |
1000 |-+ |
| |
0 +--------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Oliver Sang
1 year, 4 months
[nohz] 7dbadb2bd2: WARNING:at_kernel/time/tick-sched.c:#tick_nohz_idle_exit
by kernel test robot
Greeting,
FYI, we noticed the following commit (built with gcc-9):
commit: 7dbadb2bd2ebca693ab8fdcf80b032bb3e06ad2b ("[PATCH V1] nohz: exit nohz idle before invoking softirq")
url: https://github.com/0day-ci/linux/commits/chenshiyan/nohz-exit-nohz-idle-b...
base: https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git 5c8fe583cce542aa0b84adc939ce85293de36e5e
in testcase: rcuscale
version:
with following parameters:
runtime: 300s
scale_type: srcu
on test machine: qemu-system-x86_64 -enable-kvm -cpu SandyBridge -smp 2 -m 8G
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
+------------------------------------------------------------------+-----------+------------+
| | v5.11-rc1 | 7dbadb2bd2 |
+------------------------------------------------------------------+-----------+------------+
| boot_successes | 62 | 0 |
| WARNING:at_kernel/time/tick-sched.c:#tick_nohz_idle_exit | 0 | 4 |
| RIP:tick_nohz_idle_exit | 0 | 4 |
| RIP:_raw_spin_unlock_irq | 0 | 3 |
| WARNING:at_kernel/locking/lockdep.c:#lockdep_hardirqs_on_prepare | 0 | 4 |
| RIP:lockdep_hardirqs_on_prepare | 0 | 4 |
| kernel_BUG_at_kernel/sched/core.c | 0 | 4 |
| invalid_opcode:#[##] | 0 | 4 |
| RIP:preempt_schedule_irq | 0 | 4 |
| RIP:vprintk_emit | 0 | 2 |
| RIP:d_revalidate | 0 | 1 |
| Kernel_panic-not_syncing:Fatal_exception | 0 | 4 |
| RIP:free_unref_page_list | 0 | 1 |
| RIP:__orc_find | 0 | 1 |
| RIP:__init_waitqueue_head | 0 | 1 |
| RIP:__unwind_start | 0 | 1 |
| RIP:unwind_next_frame | 0 | 1 |
| RIP:__call_rcu | 0 | 1 |
+------------------------------------------------------------------+-----------+------------+
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang(a)intel.com>
[ 2.985918] WARNING: CPU: 0 PID: 1 at kernel/time/tick-sched.c:1237 tick_nohz_idle_exit+0x6c/0x17a
[ 2.985918] Modules linked in:
[ 2.985918] CPU: 0 PID: 1 Comm: swapper/0 Not tainted 5.11.0-rc1-g7dbadb2bd2eb #1
[ 2.985918] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.12.0-1 04/01/2014
[ 2.985918] RIP: 0010:tick_nohz_idle_exit+0x6c/0x17a
[ 2.985918] Code: e0 2a 48 c1 ea 03 8a 04 02 4c 89 e2 83 e2 07 38 d0 7f 0c 84 c0 74 08 4c 89 e7 e8 5a d5 19 00 f6 45 4c 01 75 07 e8 b0 30 04 00 <0f> 0b e8 a9 30 04 00 48 8d bd b0 00 00 00 b8 ff ff 37 00 48 89 fa
[ 2.985918] RSP: 0000:ffffc9000001fb70 EFLAGS: 00010093
[ 2.985918] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000000
[ 2.985918] RDX: ffff8881001b0000 RSI: ffffffff984354a9 RDI: ffff8881f1ff3420
[ 2.985918] RBP: ffff8881f1fe6ee0 R08: 0000000000000000 R09: 0000000000000001
[ 2.985918] R10: fffffbfff367b7c7 R11: ffffffff9b3dbe37 R12: ffff8881f1fe6f2c
[ 2.985918] R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000000
[ 2.985918] FS: 0000000000000000(0000) GS:ffff8881f1e00000(0000) knlGS:0000000000000000
[ 2.985918] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 2.985918] CR2: ffff88823ffff000 CR3: 0000000204216000 CR4: 00000000000406b0
[ 2.985918] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[ 2.985918] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
[ 2.985918] Call Trace:
[ 2.985918] __irq_exit_rcu+0x72/0xde
[ 2.985918] irq_exit_rcu+0x5/0xd
[ 2.985918] sysvec_apic_timer_interrupt+0xa0/0xae
[ 2.985918] asm_sysvec_apic_timer_interrupt+0x12/0x20
[ 2.985918] RIP: 0010:_raw_spin_unlock_irq+0x27/0x30
[ 2.985918] Code: b0 8b fe 55 48 8b 74 24 08 48 89 fd 48 83 c7 18 e8 e8 b3 96 fe 48 89 ef e8 2c 48 97 fe e8 b1 8b a7 fe fb 65 ff 0d 31 ac 5e 66 <75> 05 e8 e1 eb 7c fe 5d c3 55 48 89 fd 48 83 c7 18 53 48 89 f3 48
[ 2.985918] RSP: 0000:ffffc9000001fc70 EFLAGS: 00000286
[ 2.985918] RAX: 000000000000041d RBX: ffff8881001b0000 RCX: ffffffff98395155
[ 2.985918] RDX: ffffffff98395155 RSI: 0000000000000008 RDI: ffffffff99a34307
[ 2.985918] RBP: ffff8881002a7228 R08: fffffbfff386608f R09: 0000000000000001
[ 2.985918] R10: fffffbfff386608f R11: ffffffff9c330477 R12: 7fffffffffffffff
[ 2.985918] R13: ffff8881002a7228 R14: ffffc9000001fcd8 R15: ffff8881001b2540
[ 2.985918] ? test_bit+0x23/0x2e
[ 2.985918] ? _raw_spin_unlock_irq+0x1f/0x30
[ 2.985918] __wait_for_common+0x2d0/0x2ff
[ 2.985918] ? usleep_range+0x11b/0x11b
[ 2.985918] ? try_to_wake_up+0x476/0x807
[ 2.985918] ? bit_wait_io_timeout+0x8a/0x8a
[ 2.985918] ? __kthread_bind_mask+0x8e/0x99
[ 2.985918] kthread_park+0xb8/0xe1
[ 2.985918] __smpboot_create_thread+0x197/0x2ef
[ 2.985918] smpboot_create_threads+0x54/0xa0
[ 2.985918] ? idle_thread_get+0x8c/0x8c
[ 2.985918] cpuhp_invoke_callback+0x247/0x763
[ 2.985918] _cpu_up+0x22c/0x360
[ 2.985918] cpu_up+0x8d/0xa8
[ 2.985918] bringup_nonboot_cpus+0xa3/0xa5
[ 2.985918] smp_init+0x29/0xef
[ 2.985918] kernel_init_freeable+0x1b3/0x3f0
[ 2.985918] ? rest_init+0x185/0x185
[ 2.985918] kernel_init+0xd/0x160
[ 2.985918] ret_from_fork+0x22/0x30
[ 2.985918] irq event stamp: 1054
[ 2.985918] hardirqs last enabled at (1053): [<ffffffff99a34307>] _raw_spin_unlock_irq+0x1f/0x30
[ 2.985918] hardirqs last disabled at (1054): [<ffffffff99a1d020>] sysvec_apic_timer_interrupt+0xb/0xae
[ 2.985918] softirqs last enabled at (878): [<ffffffff9824a24b>] fpu__copy+0x12b/0x386
[ 2.985918] softirqs last disabled at (876): [<ffffffff9824a199>] fpu__copy+0x79/0x386
[ 2.985918] random: get_random_bytes called from init_oops_id+0x28/0x42 with crng_init=0
[ 2.985918] ---[ end trace 162d069718170ff7 ]---
To reproduce:
# build kernel
cd linux
cp config-5.11.0-rc1-g7dbadb2bd2eb .config
make HOSTCC=gcc-9 CC=gcc-9 ARCH=x86_64 olddefconfig prepare modules_prepare bzImage
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> job-script # job-script is attached in this email
Thanks,
Oliver Sang
1 year, 4 months
[blokc/blk] d828aefa7a: xfstests.xfs.372.fail
by kernel test robot
Greeting,
FYI, we noticed the following commit (built with gcc-9):
commit: d828aefa7a1bdf7fbc5073fdb66da9bfb1264c21 ("[PATCH v2] blokc/blk-merge: remove the next_bvec label in __blk_bios_map_sg()")
url: https://github.com/0day-ci/linux/commits/Hui-Su/blokc-blk-merge-remove-th...
base: https://git.kernel.org/cgit/linux/kernel/git/axboe/linux-block.git for-next
in testcase: xfstests
version: xfstests-x86_64-d41dcbd-1_20201218
with following parameters:
disk: 4HDD
fs: xfs
test: xfs-reflink-18
ucode: 0x21
test-description: xfstests is a regression test suite for xfs and other files ystems.
test-url: git://git.kernel.org/pub/scm/fs/xfs/xfstests-dev.git
on test machine: 4 threads Intel(R) Core(TM) i3-3220 CPU @ 3.30GHz with 8G memory
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang(a)intel.com>
2019-02-14 10:12:19 export TEST_DIR=/fs/sdb2
2019-02-14 10:12:19 export TEST_DEV=/dev/sdb2
2019-02-14 10:12:19 export FSTYP=xfs
2019-02-14 10:12:19 export SCRATCH_MNT=/fs/scratch
2019-02-14 10:12:19 mkdir /fs/scratch -p
2019-02-14 10:12:19 export SCRATCH_DEV=/dev/sdb5
2019-02-14 10:12:19 export SCRATCH_LOGDEV=/dev/sdb3
2019-02-14 10:12:19 export SCRATCH_XFS_LIST_METADATA_FIELDS=u3.sfdir3.hdr.parent.i4
2019-02-14 10:12:19 export SCRATCH_XFS_LIST_FUZZ_VERBS=random
2019-02-14 10:12:19 export MKFS_OPTIONS=-mreflink=1
2019-02-14 10:12:19 sed "s:^:xfs/:" //lkp/benchmarks/xfstests/tests/xfs-reflink-18
2019-02-14 10:12:19 ./check xfs/372 xfs/373
FSTYP -- xfs (debug)
PLATFORM -- Linux/x86_64 lkp-ivb-d04 5.10.0-10938-gd828aefa7a1b #6 SMP Wed Dec 23 22:57:37 CST 2020
MKFS_OPTIONS -- -f -mreflink=1 /dev/sdb5
MOUNT_OPTIONS -- /dev/sdb5 /fs/scratch
xfs/372 - output mismatch (see /lkp/benchmarks/xfstests/results//xfs/372.out.bad)
--- tests/xfs/372.out 2020-12-18 04:10:02.000000000 +0000
+++ /lkp/benchmarks/xfstests/results//xfs/372.out.bad 2019-02-14 10:14:04.137007985 +0000
@@ -1,4 +1,5 @@
QA output created by 372
Format and populate
Fuzz refcountbt
+xfs_mdrestore: error reading from metadump file
Done fuzzing refcountbt
...
(Run 'diff -u /lkp/benchmarks/xfstests/tests/xfs/372.out /lkp/benchmarks/xfstests/results//xfs/372.out.bad' to see the entire diff)
xfs/373 - output mismatch (see /lkp/benchmarks/xfstests/results//xfs/373.out.bad)
--- tests/xfs/373.out 2020-12-18 04:10:02.000000000 +0000
+++ /lkp/benchmarks/xfstests/results//xfs/373.out.bad 2019-02-14 10:15:47.481014528 +0000
@@ -1,4 +1,5 @@
QA output created by 373
Format and populate
Fuzz refcountbt
+xfs_mdrestore: error reading from metadump file
Done fuzzing refcountbt
...
(Run 'diff -u /lkp/benchmarks/xfstests/tests/xfs/373.out /lkp/benchmarks/xfstests/results//xfs/373.out.bad' to see the entire diff)
Ran: xfs/372 xfs/373
Failures: xfs/372 xfs/373
Failed 2 of 2 tests
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Thanks,
Oliver Sang
1 year, 4 months
[drm/i915/gt] eedd8093ed: assertion_failure
by kernel test robot
Greeting,
FYI, we noticed the following commit (built with gcc-9):
commit: eedd8093ed11ccdf4a4465f1c499c945e093aeaa ("[Intel-gfx] [CI 1/9] drm/i915/gt: Replace direct submit with direct call to tasklet")
url: https://github.com/0day-ci/linux/commits/Chris-Wilson/drm-i915-gt-Replace...
base: git://anongit.freedesktop.org/drm/drm-tip drm-tip
in testcase: igt
version: igt-x86_64-c88c4cc7-1_20201215
with following parameters:
group: group-01
ucode: 0xe2
on test machine: 8 threads Intel(R) Core(TM) i7-6700 CPU @ 3.40GHz with 28G memory
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang(a)intel.com>
[ 513.814951] With one display connected: 2.96us
[ 513.814952]
[ 513.816036] [IGT] gem_exec_nop: starting dynamic subtest vecs0
[ 513.820274] Without a display connected (headless): 3.87us
[ 513.826714]
[ 513.835841] (gem_exec_nop:2988) CRITICAL: Test assertion failure function headless, file ../tests/i915/gem_exec_nop.c:431:
[ 513.835842]
[ 513.850958] (gem_exec_nop:2988) CRITICAL: Failed assertion: (n_headless) <= (1.0 + 0.1f) * n_display && (n_headless) >= (1.0 - 0.1f) * n_display
[ 513.850959]
[ 513.867477] (gem_exec_nop:2988) CRITICAL: 'n_headless' != 'n_display' (0.000004 not within 10.000000% tolerance of 0.000003)
[ 513.867478]
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Thanks,
Oliver Sang
1 year, 4 months
[mm] 4df910620b: will-it-scale.per_process_ops 37.7% improvement
by kernel test robot
Greeting,
FYI, we noticed a 37.7% improvement of will-it-scale.per_process_ops due to commit:
commit: 4df910620bebb5cfe234af16ac8f6474b60215fd ("mm: memcg: relayout structure mem_cgroup to avoid cache interference")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
in testcase: will-it-scale
on test machine: 144 threads Intel(R) Xeon(R) CPU E7-8890 v3 @ 2.50GHz with 512G memory
with following parameters:
nr_task: 50%
mode: process
test: page_fault2
cpufreq_governor: performance
ucode: 0x16
test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
test-url: https://github.com/antonblanchard/will-it-scale
In addition to that, the commit also has significant impact on the following tests:
+------------------+------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_process_ops 2.0% improvement |
| test machine | 48 threads Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz with 112G memory |
| test parameters | cpufreq_governor=performance |
| | mode=process |
| | nr_task=50% |
| | test=page_fault2 |
| | ucode=0x42e |
+------------------+------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_process_ops 12.1% improvement |
| test machine | 144 threads Intel(R) Xeon(R) CPU E7-8890 v3 @ 2.50GHz with 512G memory |
| test parameters | cpufreq_governor=performance |
| | mode=process |
| | nr_task=100% |
| | test=page_fault2 |
| | ucode=0x16 |
+------------------+------------------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase/ucode:
gcc-9/performance/x86_64-rhel-8.3/process/50%/debian-10.4-x86_64-20200603.cgz/lkp-hsw-4ex1/page_fault2/will-it-scale/0x16
commit:
fa02fcd94b ("Merge tag 'media/v5.10-2' of git://git.kernel.org/pub/scm/linux/kernel/git/mchehab/linux-media")
4df910620b ("mm: memcg: relayout structure mem_cgroup to avoid cache interference")
fa02fcd94b0c8dff 4df910620bebb5cfe234af16ac8
---------------- ---------------------------
fail:runs %reproduction fail:runs
| | |
3:4 34% 4:4 perf-profile.calltrace.cycles-pp.error_entry.testcase
3:4 36% 5:4 perf-profile.children.cycles-pp.error_entry
3:4 30% 4:4 perf-profile.self.cycles-pp.error_entry
%stddev %change %stddev
\ | \
10516193 +37.7% 14476896 will-it-scale.72.processes
146057 +37.7% 201067 will-it-scale.per_process_ops
10516193 +37.7% 14476896 will-it-scale.workload
59980 +10.8% 66467 meminfo.max_used_kB
0.08 ± 2% +0.0 0.09 ± 4% mpstat.cpu.all.soft%
29092 ± 3% +23.6% 35961 ± 8% cpuidle.POLL.time
8949 ± 6% +25.1% 11194 ± 8% cpuidle.POLL.usage
5203 ± 10% -17.3% 4300 ± 8% numa-meminfo.node0.PageTables
68326 ± 12% -14.9% 58153 ± 9% numa-meminfo.node0.SUnreclaim
1675674 ± 3% +27.7% 2140554 ± 10% numa-meminfo.node1.AnonPages.max
36696 ± 11% -28.2% 26346 ± 12% numa-meminfo.node1.KReclaimable
36696 ± 11% -28.2% 26346 ± 12% numa-meminfo.node1.SReclaimable
4212 ± 5% +14.0% 4802 ± 9% numa-meminfo.node2.PageTables
4683 +1.9% 4772 proc-vmstat.nr_page_table_pages
3.168e+09 +37.6% 4.359e+09 proc-vmstat.numa_hit
3.168e+09 +37.6% 4.359e+09 proc-vmstat.numa_local
3.171e+09 +37.5% 4.362e+09 proc-vmstat.pgalloc_normal
3.161e+09 +37.6% 4.349e+09 proc-vmstat.pgfault
3.168e+09 +37.6% 4.358e+09 proc-vmstat.pgfree
8.176e+08 +33.8% 1.094e+09 numa-numastat.node0.local_node
8.177e+08 +33.8% 1.094e+09 numa-numastat.node0.numa_hit
7.905e+08 +37.8% 1.089e+09 numa-numastat.node1.local_node
7.906e+08 +37.8% 1.09e+09 numa-numastat.node1.numa_hit
15636 ± 82% +111.0% 32997 ± 16% numa-numastat.node1.other_node
7.865e+08 +38.5% 1.09e+09 numa-numastat.node2.local_node
7.866e+08 +38.5% 1.09e+09 numa-numastat.node2.numa_hit
7.764e+08 +40.4% 1.09e+09 numa-numastat.node3.local_node
7.764e+08 +40.4% 1.09e+09 numa-numastat.node3.numa_hit
11.53 ± 6% +66.7% 19.22 sched_debug.cfs_rq:/.nr_spread_over.avg
30.62 ± 9% +89.9% 58.17 ± 39% sched_debug.cfs_rq:/.nr_spread_over.max
1.17 ± 41% +128.6% 2.67 ± 25% sched_debug.cfs_rq:/.nr_spread_over.min
6.13 ± 15% +54.3% 9.46 ± 15% sched_debug.cfs_rq:/.nr_spread_over.stddev
-4142704 -17.9% -3401780 sched_debug.cfs_rq:/.spread0.min
0.00 ± 4% +115.1% 0.00 ± 25% sched_debug.cpu.next_balance.stddev
2904 ± 5% -12.2% 2550 ± 3% sched_debug.cpu.nr_switches.stddev
2487 ± 8% -11.9% 2191 ± 6% sched_debug.cpu.sched_count.stddev
1261 ± 8% -11.9% 1111 ± 6% sched_debug.cpu.sched_goidle.stddev
317588 ± 4% -6.8% 296039 ± 5% numa-vmstat.node0.nr_anon_pages
1300 ± 10% -17.4% 1074 ± 8% numa-vmstat.node0.nr_page_table_pages
17081 ± 12% -14.9% 14537 ± 9% numa-vmstat.node0.nr_slab_unreclaimable
4.069e+08 +33.8% 5.443e+08 numa-vmstat.node0.numa_hit
4.069e+08 +33.8% 5.443e+08 numa-vmstat.node0.numa_local
9173 ± 11% -28.2% 6586 ± 12% numa-vmstat.node1.nr_slab_reclaimable
3.941e+08 +37.4% 5.417e+08 numa-vmstat.node1.numa_hit
3.94e+08 +37.5% 5.415e+08 numa-vmstat.node1.numa_local
104864 ± 11% +17.4% 123112 ± 4% numa-vmstat.node1.numa_other
1054 ± 5% +13.7% 1199 ± 9% numa-vmstat.node2.nr_page_table_pages
3.92e+08 +38.3% 5.42e+08 numa-vmstat.node2.numa_hit
3.919e+08 +38.3% 5.419e+08 numa-vmstat.node2.numa_local
1222 ± 12% +19.1% 1456 ± 12% numa-vmstat.node3.nr_page_table_pages
3.868e+08 +40.4% 5.433e+08 numa-vmstat.node3.numa_hit
3.867e+08 +40.5% 5.432e+08 numa-vmstat.node3.numa_local
22655 ± 4% +11.5% 25263 ± 6% softirqs.CPU0.SCHED
10045 ± 2% -9.0% 9139 ± 6% softirqs.CPU1.RCU
9348 ± 6% -10.4% 8377 softirqs.CPU12.RCU
9601 ± 6% -12.6% 8396 ± 2% softirqs.CPU18.RCU
22521 ± 4% -22.2% 17522 ± 12% softirqs.CPU2.SCHED
9465 ± 5% -9.1% 8601 ± 5% softirqs.CPU25.RCU
9393 ± 5% -10.6% 8398 ± 3% softirqs.CPU26.RCU
9708 ± 4% -13.6% 8388 ± 2% softirqs.CPU27.RCU
9338 ± 7% -10.6% 8351 ± 2% softirqs.CPU28.RCU
9386 ± 5% -14.2% 8055 ± 8% softirqs.CPU31.RCU
9057 ± 7% -12.3% 7948 ± 6% softirqs.CPU36.RCU
18746 ± 20% -24.2% 14204 ± 22% softirqs.CPU4.SCHED
13616 ± 8% +22.3% 16654 ± 6% softirqs.CPU44.SCHED
13697 ± 16% +23.2% 16879 ± 12% softirqs.CPU48.SCHED
10234 ± 6% -14.8% 8719 softirqs.CPU5.RCU
12718 ± 21% +48.8% 18923 ± 23% softirqs.CPU58.SCHED
14544 ± 11% +11.4% 16196 ± 11% softirqs.CPU61.SCHED
9145 ± 7% -12.6% 7988 ± 3% softirqs.CPU70.RCU
21831 ± 6% +21.9% 26619 ± 9% softirqs.CPU74.SCHED
11254 ± 26% -28.0% 8101 ± 6% softirqs.CPU89.RCU
47398543 ± 12% -6.4e+06 41043074 ± 6% syscalls.sys_close.noise.100%
60480226 ± 8% -6e+06 54474054 ± 4% syscalls.sys_close.noise.2%
58842094 ± 9% -6.2e+06 52672651 ± 4% syscalls.sys_close.noise.25%
60405154 ± 8% -6e+06 54392442 ± 4% syscalls.sys_close.noise.5%
55525400 ± 10% -6.4e+06 49123129 ± 4% syscalls.sys_close.noise.50%
51399889 ± 10% -6.4e+06 45005108 ± 5% syscalls.sys_close.noise.75%
18095 ± 2% +11.9% 20244 syscalls.sys_mmap.med
4739 ± 6% -12.2% 4160 ± 3% syscalls.sys_mmap.min
1.337e+09 ± 2% +1.6e+08 1.501e+09 syscalls.sys_mmap.noise.100%
1.437e+09 +1.6e+08 1.592e+09 syscalls.sys_mmap.noise.2%
1.421e+09 +1.6e+08 1.58e+09 syscalls.sys_mmap.noise.25%
1.436e+09 +1.6e+08 1.592e+09 syscalls.sys_mmap.noise.5%
1.395e+09 ± 2% +1.7e+08 1.56e+09 syscalls.sys_mmap.noise.50%
1.368e+09 ± 2% +1.7e+08 1.538e+09 syscalls.sys_mmap.noise.75%
2934931 ± 3% -19.9% 2350376 ± 4% syscalls.sys_write.max
2.862e+09 ± 6% -6.2e+08 2.246e+09 ± 10% syscalls.sys_write.noise.100%
2.88e+09 ± 6% -6.1e+08 2.266e+09 ± 10% syscalls.sys_write.noise.2%
2.878e+09 ± 6% -6.1e+08 2.263e+09 ± 10% syscalls.sys_write.noise.25%
2.88e+09 ± 6% -6.1e+08 2.266e+09 ± 10% syscalls.sys_write.noise.5%
2.873e+09 ± 6% -6.1e+08 2.259e+09 ± 10% syscalls.sys_write.noise.50%
2.867e+09 ± 6% -6.1e+08 2.253e+09 ± 10% syscalls.sys_write.noise.75%
3191 ± 15% +68.3% 5370 ± 36% interrupts.CPU102.NMI:Non-maskable_interrupts
3191 ± 15% +68.3% 5370 ± 36% interrupts.CPU102.PMI:Performance_monitoring_interrupts
3249 ± 18% +97.1% 6404 ± 23% interrupts.CPU11.NMI:Non-maskable_interrupts
3249 ± 18% +97.1% 6404 ± 23% interrupts.CPU11.PMI:Performance_monitoring_interrupts
6027 ± 15% -46.9% 3197 ± 56% interrupts.CPU115.NMI:Non-maskable_interrupts
6027 ± 15% -46.9% 3197 ± 56% interrupts.CPU115.PMI:Performance_monitoring_interrupts
32.50 ± 9% +141.5% 78.50 ± 79% interrupts.CPU119.RES:Rescheduling_interrupts
5795 ± 31% -36.3% 3689 ± 7% interrupts.CPU12.NMI:Non-maskable_interrupts
5795 ± 31% -36.3% 3689 ± 7% interrupts.CPU12.PMI:Performance_monitoring_interrupts
31.25 ± 13% +244.0% 107.50 ± 79% interrupts.CPU127.RES:Rescheduling_interrupts
6072 ± 16% -32.6% 4094 ± 37% interrupts.CPU133.NMI:Non-maskable_interrupts
6072 ± 16% -32.6% 4094 ± 37% interrupts.CPU133.PMI:Performance_monitoring_interrupts
29.75 ± 13% +97.5% 58.75 ± 43% interrupts.CPU141.RES:Rescheduling_interrupts
3232 ± 22% +69.4% 5475 ± 28% interrupts.CPU25.NMI:Non-maskable_interrupts
3232 ± 22% +69.4% 5475 ± 28% interrupts.CPU25.PMI:Performance_monitoring_interrupts
5761 ± 20% +32.2% 7615 ± 4% interrupts.CPU29.NMI:Non-maskable_interrupts
5761 ± 20% +32.2% 7615 ± 4% interrupts.CPU29.PMI:Performance_monitoring_interrupts
5831 ± 24% -38.8% 3566 ± 32% interrupts.CPU3.NMI:Non-maskable_interrupts
5831 ± 24% -38.8% 3566 ± 32% interrupts.CPU3.PMI:Performance_monitoring_interrupts
4257 ± 14% +48.0% 6301 ± 10% interrupts.CPU36.NMI:Non-maskable_interrupts
4257 ± 14% +48.0% 6301 ± 10% interrupts.CPU36.PMI:Performance_monitoring_interrupts
6022 ± 26% -41.0% 3553 ± 12% interrupts.CPU48.NMI:Non-maskable_interrupts
6022 ± 26% -41.0% 3553 ± 12% interrupts.CPU48.PMI:Performance_monitoring_interrupts
4806 ± 16% -35.7% 3091 ± 12% interrupts.CPU99.NMI:Non-maskable_interrupts
4806 ± 16% -35.7% 3091 ± 12% interrupts.CPU99.PMI:Performance_monitoring_interrupts
17.42 -11.8% 15.36 perf-stat.i.MPKI
1.005e+10 +44.7% 1.455e+10 perf-stat.i.branch-instructions
0.52 -0.0 0.48 perf-stat.i.branch-miss-rate%
50729291 +33.0% 67453908 perf-stat.i.branch-misses
35.51 +2.5 37.97 perf-stat.i.cache-miss-rate%
3.216e+08 +34.8% 4.335e+08 perf-stat.i.cache-misses
9.03e+08 +26.0% 1.138e+09 perf-stat.i.cache-references
4.05 -30.3% 2.82 perf-stat.i.cpi
675.62 -25.2% 505.55 perf-stat.i.cycles-between-cache-misses
14707774 ± 9% +34.4% 19774179 ± 2% perf-stat.i.dTLB-load-misses
1.565e+10 +42.2% 2.225e+10 perf-stat.i.dTLB-loads
73922631 +37.6% 1.017e+08 perf-stat.i.dTLB-store-misses
9.737e+09 +36.5% 1.329e+10 perf-stat.i.dTLB-stores
93.39 +1.8 95.21 perf-stat.i.iTLB-load-miss-rate%
31412146 +37.9% 43313640 perf-stat.i.iTLB-load-misses
2165412 -3.2% 2096990 perf-stat.i.iTLB-loads
5.18e+10 +43.0% 7.408e+10 perf-stat.i.instructions
1657 +4.0% 1724 perf-stat.i.instructions-per-iTLB-miss
0.25 +43.4% 0.35 perf-stat.i.ipc
255.15 +40.9% 359.51 perf-stat.i.metric.M/sec
10465904 +37.7% 14408416 perf-stat.i.minor-faults
6.04 -2.6 3.48 ± 2% perf-stat.i.node-load-miss-rate%
15615698 -25.0% 11707799 perf-stat.i.node-load-misses
2.551e+08 +37.9% 3.516e+08 perf-stat.i.node-loads
8900850 +39.9% 12451889 perf-stat.i.node-store-misses
39327623 +40.3% 55186081 perf-stat.i.node-stores
10465907 +37.7% 14408418 perf-stat.i.page-faults
17.43 -11.9% 15.36 perf-stat.overall.MPKI
0.50 -0.0 0.46 perf-stat.overall.branch-miss-rate%
35.62 +2.5 38.09 perf-stat.overall.cache-miss-rate%
4.05 -30.3% 2.82 perf-stat.overall.cpi
652.44 -26.0% 482.69 perf-stat.overall.cycles-between-cache-misses
93.55 +1.8 95.38 perf-stat.overall.iTLB-load-miss-rate%
1649 +3.7% 1710 perf-stat.overall.instructions-per-iTLB-miss
0.25 +43.4% 0.35 perf-stat.overall.ipc
5.77 -2.5 3.22 perf-stat.overall.node-load-miss-rate%
1484194 +3.8% 1540211 perf-stat.overall.path-length
1.002e+10 +44.7% 1.449e+10 perf-stat.ps.branch-instructions
50561382 +32.9% 67185149 perf-stat.ps.branch-misses
3.205e+08 +34.8% 4.319e+08 perf-stat.ps.cache-misses
8.999e+08 +26.0% 1.134e+09 perf-stat.ps.cache-references
14656335 ± 9% +34.4% 19698887 ± 2% perf-stat.ps.dTLB-load-misses
1.559e+10 +42.2% 2.217e+10 perf-stat.ps.dTLB-loads
73657440 +37.6% 1.013e+08 perf-stat.ps.dTLB-store-misses
9.703e+09 +36.5% 1.324e+10 perf-stat.ps.dTLB-stores
31302020 +37.9% 43150913 perf-stat.ps.iTLB-load-misses
2157706 -3.2% 2088881 perf-stat.ps.iTLB-loads
5.163e+10 +43.0% 7.38e+10 perf-stat.ps.instructions
10429014 +37.6% 14354103 perf-stat.ps.minor-faults
15559965 -25.0% 11664101 perf-stat.ps.node-load-misses
2.542e+08 +37.8% 3.503e+08 perf-stat.ps.node-loads
8869865 +39.9% 12405459 perf-stat.ps.node-store-misses
39189951 +40.3% 54978655 perf-stat.ps.node-stores
10429016 +37.6% 14354105 perf-stat.ps.page-faults
1.561e+13 +42.9% 2.23e+13 perf-stat.total.instructions
16.08 ± 14% -11.3 4.75 ± 11% perf-profile.calltrace.cycles-pp.mem_cgroup_charge.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
6.45 ± 15% -5.0 1.43 ± 13% perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.mem_cgroup_charge.do_fault.__handle_mm_fault.handle_mm_fault
1.22 ± 10% -0.6 0.64 ± 9% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.page_add_new_anon_rmap.alloc_set_pte.finish_fault.do_fault
1.38 ± 11% -0.5 0.83 ± 9% perf-profile.calltrace.cycles-pp.page_add_new_anon_rmap.alloc_set_pte.finish_fault.do_fault.__handle_mm_fault
0.68 ± 14% +0.2 0.92 ± 9% perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
0.84 ± 13% +0.3 1.13 ± 9% perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.do_fault
0.77 ± 13% +0.3 1.07 ± 11% perf-profile.calltrace.cycles-pp.__irqentry_text_end.testcase
0.77 ± 13% +0.3 1.07 ± 13% perf-profile.calltrace.cycles-pp.__list_del_entry_valid.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages_nodemask
0.61 ± 13% +0.3 0.92 ± 12% perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap
0.63 ± 13% +0.3 0.95 ± 12% perf-profile.calltrace.cycles-pp.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap
0.63 ± 13% +0.3 0.96 ± 12% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
0.97 ± 13% +0.3 1.30 ± 9% perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.do_fault.__handle_mm_fault
1.07 ± 13% +0.4 1.52 ± 10% perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.do_fault.__handle_mm_fault.handle_mm_fault
1.17 ± 13% +0.5 1.64 ± 10% perf-profile.calltrace.cycles-pp.__do_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
1.10 ± 12% +0.6 1.73 ± 12% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_list.release_pages.tlb_flush_mmu.zap_pte_range
0.00 +0.6 0.65 ± 10% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.page_remove_rmap.zap_pte_range.unmap_page_range.unmap_vmas
0.81 ± 10% +0.7 1.48 ± 9% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge.mem_cgroup_charge.do_fault.__handle_mm_fault
1.29 ± 12% +0.7 1.98 ± 12% perf-profile.calltrace.cycles-pp.free_unref_page_list.release_pages.tlb_flush_mmu.zap_pte_range.unmap_page_range
1.27 ± 13% +0.8 2.08 ± 12% perf-profile.calltrace.cycles-pp.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma
0.00 +0.8 0.84 ± 12% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist
0.00 +0.8 0.85 ± 12% perf-profile.calltrace.cycles-pp._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages_nodemask
1.66 ± 13% +0.9 2.58 ± 12% perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.do_fault
1.17 ± 10% +0.9 2.10 ± 9% perf-profile.calltrace.cycles-pp.try_charge.mem_cgroup_charge.do_fault.__handle_mm_fault.handle_mm_fault
1.86 ± 13% +1.0 2.85 ± 12% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.do_fault.__handle_mm_fault
2.06 ± 13% +1.1 3.12 ± 12% perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.do_fault.__handle_mm_fault.handle_mm_fault
2.37 ± 13% +1.2 3.53 ± 12% perf-profile.calltrace.cycles-pp.alloc_pages_vma.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
2.03 ± 14% +1.7 3.73 ± 14% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.release_pages.tlb_flush_mmu.zap_pte_range
2.04 ± 14% +1.7 3.75 ± 14% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.release_pages.tlb_flush_mmu.zap_pte_range.unmap_page_range
4.69 ± 13% +2.5 7.17 ± 13% perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.zap_pte_range.unmap_page_range.unmap_vmas
5.01 ± 13% +2.6 7.63 ± 12% perf-profile.calltrace.cycles-pp.tlb_flush_mmu.zap_pte_range.unmap_page_range.unmap_vmas.unmap_region
3.28 ± 11% +3.1 6.34 ± 9% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.pagevec_lru_move_fn.lru_cache_add.alloc_set_pte
3.32 ± 10% +3.1 6.40 ± 9% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.pagevec_lru_move_fn.lru_cache_add.alloc_set_pte.finish_fault
7.62 ± 12% +3.3 10.96 ± 12% perf-profile.calltrace.cycles-pp.zap_pte_range.unmap_page_range.unmap_vmas.unmap_region.__do_munmap
7.62 ± 12% +3.4 10.97 ± 12% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.__do_munmap.__vm_munmap
7.62 ± 12% +3.4 10.98 ± 12% perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
4.91 ± 11% +3.4 8.30 ± 9% perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.lru_cache_add.alloc_set_pte.finish_fault.do_fault
8.80 ± 11% +3.4 12.22 ± 10% perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.do_fault.__handle_mm_fault.handle_mm_fault
8.83 ± 11% +3.4 12.26 ± 10% perf-profile.calltrace.cycles-pp.finish_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
5.01 ± 11% +3.4 8.45 ± 9% perf-profile.calltrace.cycles-pp.lru_cache_add.alloc_set_pte.finish_fault.do_fault.__handle_mm_fault
8.27 ± 12% +3.7 11.94 ± 12% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
8.27 ± 12% +3.7 11.94 ± 12% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
8.27 ± 12% +3.7 11.94 ± 12% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
8.27 ± 12% +3.7 11.94 ± 12% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
8.27 ± 12% +3.7 11.95 ± 12% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
8.27 ± 12% +3.7 11.95 ± 12% perf-profile.calltrace.cycles-pp.__munmap
8.27 ± 12% +3.7 11.95 ± 12% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
16.11 ± 14% -11.3 4.76 ± 11% perf-profile.children.cycles-pp.mem_cgroup_charge
6.48 ± 15% -5.0 1.45 ± 13% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
2.04 ± 12% -1.3 0.74 ± 10% perf-profile.children.cycles-pp.__count_memcg_events
2.30 ± 11% -0.9 1.43 ± 10% perf-profile.children.cycles-pp.__mod_memcg_state
2.30 ± 16% -0.8 1.51 ± 10% perf-profile.children.cycles-pp.native_irq_return_iret
1.38 ± 11% -0.5 0.83 ± 10% perf-profile.children.cycles-pp.page_add_new_anon_rmap
0.90 ± 12% -0.5 0.37 ± 10% perf-profile.children.cycles-pp.mem_cgroup_charge_statistics
0.28 ± 16% -0.2 0.09 ± 13% perf-profile.children.cycles-pp.uncharge_page
0.40 ± 14% -0.2 0.23 ± 10% perf-profile.children.cycles-pp.mem_cgroup_uncharge_list
0.06 ± 14% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.pte_alloc_one
0.07 ± 7% +0.0 0.09 ± 9% perf-profile.children.cycles-pp.find_vma
0.06 ± 14% +0.0 0.08 ± 12% perf-profile.children.cycles-pp.get_task_policy
0.07 ± 7% +0.0 0.09 ± 7% perf-profile.children.cycles-pp.__might_sleep
0.06 ± 13% +0.0 0.09 ± 9% perf-profile.children.cycles-pp._cond_resched
0.08 ± 6% +0.0 0.10 ± 12% perf-profile.children.cycles-pp.up_read
0.07 ± 11% +0.0 0.10 ± 10% perf-profile.children.cycles-pp.unlock_page
0.07 ± 14% +0.0 0.10 ± 10% perf-profile.children.cycles-pp.page_mapping
0.04 ± 58% +0.0 0.07 ± 5% perf-profile.children.cycles-pp.vmacache_find
0.07 ± 7% +0.0 0.10 ± 13% perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
0.06 ± 9% +0.0 0.09 ± 15% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.09 ± 14% +0.0 0.13 ± 10% perf-profile.children.cycles-pp.free_unref_page_commit
0.09 ± 14% +0.0 0.13 ± 14% perf-profile.children.cycles-pp.perf_swevent_get_recursion_context
0.01 ±173% +0.0 0.06 ± 9% perf-profile.children.cycles-pp.prep_new_page
0.09 ± 8% +0.0 0.14 ± 11% perf-profile.children.cycles-pp.__mod_zone_page_state
0.12 ± 16% +0.0 0.16 ± 10% perf-profile.children.cycles-pp.__list_add_valid
0.12 ± 14% +0.0 0.17 ± 9% perf-profile.children.cycles-pp.___might_sleep
0.08 ± 14% +0.0 0.12 ± 16% perf-profile.children.cycles-pp.cgroup_throttle_swaprate
0.03 ±100% +0.0 0.07 ± 11% perf-profile.children.cycles-pp.page_counter_uncharge
0.03 ±100% +0.0 0.07 ± 11% perf-profile.children.cycles-pp.page_counter_cancel
0.09 ± 8% +0.1 0.15 ± 14% perf-profile.children.cycles-pp.propagate_protected_usage
0.17 ± 8% +0.1 0.23 ± 11% perf-profile.children.cycles-pp.__mod_node_page_state
0.16 ± 13% +0.1 0.22 ± 12% perf-profile.children.cycles-pp.sync_regs
0.20 ± 14% +0.1 0.27 ± 11% perf-profile.children.cycles-pp.___perf_sw_event
0.24 ± 10% +0.1 0.32 ± 12% perf-profile.children.cycles-pp.__mod_lruvec_state
0.30 ± 14% +0.1 0.41 ± 8% perf-profile.children.cycles-pp.xas_load
0.34 ± 12% +0.1 0.47 ± 12% perf-profile.children.cycles-pp.__perf_sw_event
0.33 ± 13% +0.2 0.48 ± 10% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.68 ± 13% +0.2 0.92 ± 9% perf-profile.children.cycles-pp.find_get_entry
0.85 ± 13% +0.3 1.15 ± 9% perf-profile.children.cycles-pp.find_lock_entry
0.77 ± 13% +0.3 1.07 ± 11% perf-profile.children.cycles-pp.__irqentry_text_end
0.64 ± 13% +0.3 0.96 ± 12% perf-profile.children.cycles-pp.tlb_finish_mmu
0.98 ± 13% +0.3 1.31 ± 9% perf-profile.children.cycles-pp.shmem_getpage_gfp
1.08 ± 13% +0.4 1.52 ± 10% perf-profile.children.cycles-pp.shmem_fault
1.17 ± 12% +0.4 1.61 ± 12% perf-profile.children.cycles-pp.__list_del_entry_valid
1.17 ± 13% +0.5 1.64 ± 10% perf-profile.children.cycles-pp.__do_fault
0.82 ± 10% +0.7 1.50 ± 9% perf-profile.children.cycles-pp.page_counter_try_charge
1.25 ± 12% +0.7 1.96 ± 12% perf-profile.children.cycles-pp.free_pcppages_bulk
1.48 ± 12% +0.8 2.26 ± 12% perf-profile.children.cycles-pp.free_unref_page_list
1.29 ± 13% +0.8 2.10 ± 12% perf-profile.children.cycles-pp.rmqueue_bulk
1.68 ± 13% +0.9 2.61 ± 12% perf-profile.children.cycles-pp.rmqueue
1.17 ± 10% +0.9 2.12 ± 9% perf-profile.children.cycles-pp.try_charge
1.90 ± 13% +1.0 2.90 ± 12% perf-profile.children.cycles-pp.get_page_from_freelist
2.14 ± 13% +1.1 3.22 ± 12% perf-profile.children.cycles-pp.__alloc_pages_nodemask
2.38 ± 13% +1.2 3.54 ± 12% perf-profile.children.cycles-pp.alloc_pages_vma
2.97 ± 11% +1.3 4.23 ± 11% perf-profile.children.cycles-pp._raw_spin_lock
5.43 ± 13% +2.8 8.27 ± 12% perf-profile.children.cycles-pp.release_pages
5.64 ± 13% +2.9 8.59 ± 12% perf-profile.children.cycles-pp.tlb_flush_mmu
7.62 ± 12% +3.3 10.97 ± 12% perf-profile.children.cycles-pp.zap_pte_range
7.63 ± 12% +3.4 10.98 ± 12% perf-profile.children.cycles-pp.unmap_vmas
7.63 ± 12% +3.4 10.98 ± 12% perf-profile.children.cycles-pp.unmap_page_range
4.92 ± 11% +3.4 8.32 ± 9% perf-profile.children.cycles-pp.pagevec_lru_move_fn
8.81 ± 11% +3.4 12.22 ± 10% perf-profile.children.cycles-pp.alloc_set_pte
8.84 ± 11% +3.4 12.26 ± 10% perf-profile.children.cycles-pp.finish_fault
5.01 ± 11% +3.4 8.45 ± 9% perf-profile.children.cycles-pp.lru_cache_add
8.27 ± 12% +3.7 11.94 ± 12% perf-profile.children.cycles-pp.__vm_munmap
8.27 ± 12% +3.7 11.94 ± 12% perf-profile.children.cycles-pp.__x64_sys_munmap
8.27 ± 12% +3.7 11.94 ± 12% perf-profile.children.cycles-pp.unmap_region
8.27 ± 12% +3.7 11.95 ± 12% perf-profile.children.cycles-pp.__do_munmap
8.37 ± 12% +3.7 12.04 ± 12% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
8.36 ± 12% +3.7 12.04 ± 12% perf-profile.children.cycles-pp.do_syscall_64
8.27 ± 12% +3.7 11.95 ± 12% perf-profile.children.cycles-pp.__munmap
5.65 ± 12% +5.0 10.64 ± 11% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
6.27 ± 12% +5.7 12.01 ± 11% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
7.44 ± 15% -6.7 0.74 ± 13% perf-profile.self.cycles-pp.mem_cgroup_charge
6.43 ± 15% -5.0 1.43 ± 13% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
2.04 ± 12% -1.3 0.73 ± 10% perf-profile.self.cycles-pp.__count_memcg_events
2.29 ± 11% -0.9 1.42 ± 10% perf-profile.self.cycles-pp.__mod_memcg_state
2.30 ± 16% -0.8 1.51 ± 10% perf-profile.self.cycles-pp.native_irq_return_iret
0.28 ± 16% -0.2 0.08 ± 15% perf-profile.self.cycles-pp.uncharge_page
0.06 ± 9% +0.0 0.08 ± 10% perf-profile.self.cycles-pp.__might_sleep
0.07 ± 11% +0.0 0.10 ± 8% perf-profile.self.cycles-pp.up_read
0.06 ± 6% +0.0 0.09 ± 14% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.07 ± 7% +0.0 0.09 ± 11% perf-profile.self.cycles-pp.__perf_sw_event
0.07 ± 7% +0.0 0.09 ± 11% perf-profile.self.cycles-pp.__mod_lruvec_state
0.06 ± 9% +0.0 0.08 ± 12% perf-profile.self.cycles-pp.mem_cgroup_page_lruvec
0.04 ± 58% +0.0 0.07 ± 7% perf-profile.self.cycles-pp.vmacache_find
0.08 ± 14% +0.0 0.10 ± 12% perf-profile.self.cycles-pp.alloc_pages_vma
0.07 ± 14% +0.0 0.10 ± 10% perf-profile.self.cycles-pp.unlock_page
0.07 ± 14% +0.0 0.10 ± 10% perf-profile.self.cycles-pp.page_mapping
0.06 ± 11% +0.0 0.09 ± 11% perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
0.10 ± 12% +0.0 0.14 ± 11% perf-profile.self.cycles-pp.exc_page_fault
0.04 ± 59% +0.0 0.08 ± 10% perf-profile.self.cycles-pp.get_task_policy
0.10 ± 11% +0.0 0.13 ± 12% perf-profile.self.cycles-pp.find_lock_entry
0.09 ± 8% +0.0 0.13 ± 14% perf-profile.self.cycles-pp.__mod_zone_page_state
0.08 ± 13% +0.0 0.12 ± 13% perf-profile.self.cycles-pp.perf_swevent_get_recursion_context
0.13 ± 9% +0.0 0.17 ± 13% perf-profile.self.cycles-pp.do_user_addr_fault
0.10 ± 15% +0.0 0.14 ± 11% perf-profile.self.cycles-pp.lru_cache_add
0.12 ± 10% +0.0 0.16 ± 9% perf-profile.self.cycles-pp.alloc_set_pte
0.06 ± 13% +0.0 0.10 ± 14% perf-profile.self.cycles-pp.cgroup_throttle_swaprate
0.01 ±173% +0.0 0.06 ± 9% perf-profile.self.cycles-pp.free_unref_page_prepare
0.12 ± 13% +0.0 0.16 ± 9% perf-profile.self.cycles-pp.___might_sleep
0.14 ± 12% +0.0 0.18 ± 12% perf-profile.self.cycles-pp.___perf_sw_event
0.10 ± 15% +0.0 0.15 ± 10% perf-profile.self.cycles-pp.__list_add_valid
0.14 ± 11% +0.0 0.18 ± 9% perf-profile.self.cycles-pp.__alloc_pages_nodemask
0.01 ±173% +0.0 0.06 ± 11% perf-profile.self.cycles-pp.page_counter_cancel
0.09 ± 7% +0.1 0.14 ± 12% perf-profile.self.cycles-pp.propagate_protected_usage
0.17 ± 9% +0.1 0.22 ± 11% perf-profile.self.cycles-pp.__mod_node_page_state
0.03 ±100% +0.1 0.08 ± 10% perf-profile.self.cycles-pp.memcg_check_events
0.14 ± 11% +0.1 0.19 ± 11% perf-profile.self.cycles-pp.sync_regs
0.18 ± 11% +0.1 0.25 ± 11% perf-profile.self.cycles-pp.do_fault
0.23 ± 15% +0.1 0.31 ± 12% perf-profile.self.cycles-pp.rmqueue
0.22 ± 11% +0.1 0.31 ± 12% perf-profile.self.cycles-pp.handle_mm_fault
0.18 ± 11% +0.1 0.27 ± 10% perf-profile.self.cycles-pp.page_remove_rmap
0.11 ± 12% +0.1 0.20 ± 12% perf-profile.self.cycles-pp.shmem_fault
0.26 ± 12% +0.1 0.36 ± 8% perf-profile.self.cycles-pp.xas_load
0.38 ± 14% +0.1 0.50 ± 10% perf-profile.self.cycles-pp.find_get_entry
0.35 ± 12% +0.1 0.48 ± 9% perf-profile.self.cycles-pp.__handle_mm_fault
0.35 ± 11% +0.1 0.49 ± 11% perf-profile.self.cycles-pp.release_pages
0.43 ± 13% +0.1 0.58 ± 7% perf-profile.self.cycles-pp.__pagevec_lru_add_fn
0.32 ± 12% +0.2 0.48 ± 11% perf-profile.self.cycles-pp.free_pages_and_swap_cache
0.35 ± 10% +0.3 0.62 ± 9% perf-profile.self.cycles-pp.try_charge
0.77 ± 13% +0.3 1.07 ± 11% perf-profile.self.cycles-pp.__irqentry_text_end
0.81 ± 12% +0.3 1.14 ± 12% perf-profile.self.cycles-pp.free_pcppages_bulk
1.16 ± 12% +0.4 1.60 ± 12% perf-profile.self.cycles-pp.__list_del_entry_valid
0.62 ± 8% +0.5 1.10 ± 9% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
0.74 ± 10% +0.6 1.36 ± 9% perf-profile.self.cycles-pp.page_counter_try_charge
1.47 ± 10% +0.7 2.22 ± 10% perf-profile.self.cycles-pp.zap_pte_range
2.21 ± 12% +0.8 3.06 ± 10% perf-profile.self.cycles-pp.testcase
6.27 ± 12% +5.7 12.01 ± 11% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
will-it-scale.72.processes
1.5e+07 +----------------------------------------------------------------+
1.45e+07 |-+ O O O O O O |
| O O O O O O |
1.4e+07 |-+ O |
1.35e+07 |-+ |
| |
1.3e+07 |-+ |
1.25e+07 |-+ |
1.2e+07 |-+ |
| |
1.15e+07 |-+ |
1.1e+07 |-+ |
| ..+... ..+.... |
1.05e+07 |....+...+....+....+...+....+....+...+.. +.. + |
1e+07 +----------------------------------------------------------------+
will-it-scale.per_process_ops
210000 +------------------------------------------------------------------+
| O |
200000 |-+ O O O O O O O O O O O |
| O |
190000 |-+ |
| |
180000 |-+ |
| |
170000 |-+ |
| |
160000 |-+ |
| |
150000 |-+ |
|....+....+...+....+....+....+....+...+....+....+....+...+ |
140000 +------------------------------------------------------------------+
will-it-scale.workload
1.5e+07 +----------------------------------------------------------------+
1.45e+07 |-+ O O O O O O |
| O O O O O O |
1.4e+07 |-+ O |
1.35e+07 |-+ |
| |
1.3e+07 |-+ |
1.25e+07 |-+ |
1.2e+07 |-+ |
| |
1.15e+07 |-+ |
1.1e+07 |-+ |
| ..+... ..+.... |
1.05e+07 |....+...+....+....+...+....+....+...+.. +.. + |
1e+07 +----------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
***************************************************************************************************
lkp-ivb-2ep1: 48 threads Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz with 112G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase/ucode:
gcc-9/performance/x86_64-rhel-8.3/process/50%/debian-10.4-x86_64-20200603.cgz/lkp-ivb-2ep1/page_fault2/will-it-scale/0x42e
commit:
fa02fcd94b ("Merge tag 'media/v5.10-2' of git://git.kernel.org/pub/scm/linux/kernel/git/mchehab/linux-media")
4df910620b ("mm: memcg: relayout structure mem_cgroup to avoid cache interference")
fa02fcd94b0c8dff 4df910620bebb5cfe234af16ac8
---------------- ---------------------------
fail:runs %reproduction fail:runs
| | |
5:4 18% 6:4 perf-profile.calltrace.cycles-pp.error_entry.testcase
5:4 18% 6:4 perf-profile.children.cycles-pp.error_entry
4:4 15% 5:4 perf-profile.self.cycles-pp.error_entry
%stddev %change %stddev
\ | \
0.96 ± 16% -0.3 0.71 ± 14% perf-profile.calltrace.cycles-pp.page_add_new_anon_rmap.alloc_set_pte.finish_fault.do_fault.__handle_mm_fault
0.96 ± 15% -0.5 0.47 ± 6% perf-profile.children.cycles-pp.__count_memcg_events
1.42 ± 14% -0.4 1.05 ± 10% perf-profile.children.cycles-pp.__mod_memcg_state
0.97 ± 16% -0.3 0.72 ± 14% perf-profile.children.cycles-pp.page_add_new_anon_rmap
0.43 ± 10% -0.2 0.26 ± 5% perf-profile.children.cycles-pp.mem_cgroup_charge_statistics
0.61 ± 12% -0.2 0.44 ± 10% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
0.06 ± 11% +0.0 0.08 ± 10% perf-profile.children.cycles-pp.worker_thread
0.07 ± 7% +0.1 0.12 ± 32% perf-profile.children.cycles-pp.ret_from_fork
0.07 ± 7% +0.1 0.12 ± 32% perf-profile.children.cycles-pp.kthread
0.52 ± 12% +0.1 0.60 ± 10% perf-profile.children.cycles-pp.xas_load
0.95 ± 14% -0.5 0.47 ± 7% perf-profile.self.cycles-pp.__count_memcg_events
1.40 ± 15% -0.4 1.04 ± 10% perf-profile.self.cycles-pp.__mod_memcg_state
0.60 ± 12% -0.2 0.42 ± 10% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
0.41 ± 9% -0.1 0.34 ± 12% perf-profile.self.cycles-pp.mem_cgroup_charge
0.29 ± 13% +0.1 0.35 ± 9% perf-profile.self.cycles-pp.handle_mm_fault
0.43 ± 11% +0.1 0.51 ± 9% perf-profile.self.cycles-pp.xas_load
1.06 ± 11% +0.2 1.29 ± 10% perf-profile.self.cycles-pp.free_pages_and_swap_cache
1.67 ± 11% +0.3 1.94 ± 10% perf-profile.self.cycles-pp.free_pcppages_bulk
0.32 ± 25% +0.7 1.02 ± 71% perf-profile.self.cycles-pp.alloc_set_pte
5158061 +2.0% 5263503 will-it-scale.24.processes
214918 +2.0% 219312 will-it-scale.per_process_ops
5158061 +2.0% 5263503 will-it-scale.workload
31247 ± 57% -64.6% 11065 ±169% numa-numastat.node1.other_node
14458 ± 5% -10.2% 12989 ± 7% numa-vmstat.node0.nr_slab_unreclaimable
11137 ± 4% +8.3% 12057 ± 3% numa-vmstat.node1.nr_slab_reclaimable
12240 ± 8% +18.3% 14486 ± 5% numa-vmstat.node1.nr_slab_unreclaimable
2484583 ± 4% +20.9% 3003853 ± 10% sched_debug.cfs_rq:/.min_vruntime.max
1911501 ± 14% +33.8% 2557320 ± 13% sched_debug.cfs_rq:/.spread0.max
3425 ± 11% -11.9% 3018 ± 2% sched_debug.cpu.nr_switches.stddev
1.557e+09 +2.0% 1.589e+09 proc-vmstat.numa_hit
1.557e+09 +2.0% 1.589e+09 proc-vmstat.numa_local
1.557e+09 +2.0% 1.589e+09 proc-vmstat.pgalloc_normal
1.553e+09 +2.0% 1.584e+09 proc-vmstat.pgfault
1.557e+09 +2.0% 1.589e+09 proc-vmstat.pgfree
57832 ± 5% -10.2% 51959 ± 7% numa-meminfo.node0.SUnreclaim
101264 ± 4% -8.8% 92360 ± 5% numa-meminfo.node0.Slab
44549 ± 4% +8.3% 48229 ± 3% numa-meminfo.node1.KReclaimable
44549 ± 4% +8.3% 48229 ± 3% numa-meminfo.node1.SReclaimable
48963 ± 8% +18.3% 57947 ± 5% numa-meminfo.node1.SUnreclaim
93513 ± 6% +13.5% 106178 ± 4% numa-meminfo.node1.Slab
798.75 ± 6% +12.5% 898.75 ± 7% slabinfo.file_lock_cache.active_objs
798.75 ± 6% +12.5% 898.75 ± 7% slabinfo.file_lock_cache.num_objs
2208 ± 2% +14.7% 2533 ± 8% slabinfo.fsnotify_mark_connector.active_objs
2208 ± 2% +14.7% 2533 ± 8% slabinfo.fsnotify_mark_connector.num_objs
1848 ± 4% +9.0% 2014 ± 5% slabinfo.kmalloc-rcl-96.active_objs
1848 ± 4% +9.0% 2014 ± 5% slabinfo.kmalloc-rcl-96.num_objs
0.03 ± 5% +38.6% 0.04 ± 7% perf-sched.sch_delay.avg.ms.futex_wait_queue_me.futex_wait.do_futex.__x64_sys_futex
0.04 ± 7% +29.9% 0.06 ± 12% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.04 ± 7% +46.9% 0.05 ± 17% perf-sched.sch_delay.max.ms.futex_wait_queue_me.futex_wait.do_futex.__x64_sys_futex
0.80 ±158% -94.0% 0.05 ± 10% perf-sched.sch_delay.max.ms.preempt_schedule_common._cond_resched.stop_one_cpu.__set_cpus_allowed_ptr.sched_setaffinity
0.04 ± 15% +34.5% 0.05 ± 5% perf-sched.sch_delay.max.ms.rcu_gp_kthread.kthread.ret_from_fork
3290 ± 50% -60.9% 1286 ± 33% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork
0.37 ± 48% -44.6% 0.21 ± 5% perf-sched.wait_time.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
3290 ± 50% -60.9% 1286 ± 33% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork
15251 ± 17% -20.0% 12195 ± 19% softirqs.CPU0.RCU
25042 ± 17% -38.4% 15437 ± 28% softirqs.CPU2.SCHED
23971 ± 11% -19.1% 19385 ± 4% softirqs.CPU21.SCHED
25139 ± 7% -12.0% 22110 ± 10% softirqs.CPU22.SCHED
10852 ± 13% -24.4% 8203 ± 11% softirqs.CPU24.SCHED
20926 ± 20% +44.8% 30292 ± 12% softirqs.CPU26.SCHED
10419 ± 6% -13.7% 8991 softirqs.CPU34.RCU
21252 ± 12% +23.3% 26201 ± 4% softirqs.CPU45.SCHED
4558 ± 35% -39.5% 2760 ± 60% interrupts.CPU0.NMI:Non-maskable_interrupts
4558 ± 35% -39.5% 2760 ± 60% interrupts.CPU0.PMI:Performance_monitoring_interrupts
171.50 ± 17% +28.1% 219.75 ± 12% interrupts.CPU2.RES:Rescheduling_interrupts
150.50 ± 4% +33.9% 201.50 ± 7% interrupts.CPU21.RES:Rescheduling_interrupts
944.00 ± 8% +51.4% 1429 ± 31% interrupts.CPU22.CAL:Function_call_interrupts
143.50 ± 14% +31.9% 189.25 ± 7% interrupts.CPU22.RES:Rescheduling_interrupts
187.75 ± 13% -30.2% 131.00 ± 18% interrupts.CPU26.RES:Rescheduling_interrupts
6516 ± 24% -52.8% 3074 ± 7% interrupts.CPU31.NMI:Non-maskable_interrupts
6516 ± 24% -52.8% 3074 ± 7% interrupts.CPU31.PMI:Performance_monitoring_interrupts
6557 ± 23% -37.7% 4084 ± 26% interrupts.CPU36.NMI:Non-maskable_interrupts
6557 ± 23% -37.7% 4084 ± 26% interrupts.CPU36.PMI:Performance_monitoring_interrupts
819.25 ± 18% +42.9% 1170 ± 15% interrupts.CPU40.CAL:Function_call_interrupts
935.25 ± 5% +12.2% 1049 ± 7% interrupts.CPU42.CAL:Function_call_interrupts
5249 ± 28% -46.8% 2793 ± 41% interrupts.CPU5.NMI:Non-maskable_interrupts
5249 ± 28% -46.8% 2793 ± 41% interrupts.CPU5.PMI:Performance_monitoring_interrupts
1099 ± 4% -15.9% 924.50 ± 6% interrupts.CPU8.CAL:Function_call_interrupts
27.32 -1.2% 27.00 perf-stat.i.MPKI
4.61e+09 +2.1% 4.705e+09 perf-stat.i.branch-instructions
66.43 +0.7 67.16 perf-stat.i.cache-miss-rate%
4.384e+08 +2.0% 4.471e+08 perf-stat.i.cache-misses
6.595e+08 +0.9% 6.655e+08 perf-stat.i.cache-references
2.99 -1.9% 2.93 perf-stat.i.cpi
165.29 -2.0% 162.05 perf-stat.i.cycles-between-cache-misses
8.043e+09 +1.8% 8.187e+09 perf-stat.i.dTLB-loads
5.408e+09 +2.1% 5.524e+09 perf-stat.i.dTLB-stores
5985395 +1.7% 6086521 perf-stat.i.iTLB-load-misses
2.413e+10 +2.1% 2.464e+10 perf-stat.i.instructions
0.34 +1.9% 0.34 perf-stat.i.ipc
403.98 +1.9% 411.70 perf-stat.i.metric.M/sec
5141091 +2.1% 5246543 perf-stat.i.minor-faults
4.65 -0.6 4.07 perf-stat.i.node-load-miss-rate%
13596718 -11.5% 12039508 perf-stat.i.node-load-misses
2.839e+08 +1.7% 2.887e+08 perf-stat.i.node-loads
3.06 -0.5 2.57 perf-stat.i.node-store-miss-rate%
9836492 -14.4% 8417873 perf-stat.i.node-store-misses
3.192e+08 +2.3% 3.266e+08 perf-stat.i.node-stores
5141100 +2.1% 5246556 perf-stat.i.page-faults
27.32 -1.2% 27.00 perf-stat.overall.MPKI
66.48 +0.7 67.19 perf-stat.overall.cache-miss-rate%
2.99 -1.9% 2.93 perf-stat.overall.cpi
164.46 -1.8% 161.48 perf-stat.overall.cycles-between-cache-misses
0.33 +2.0% 0.34 perf-stat.overall.ipc
4.57 -0.6 4.00 perf-stat.overall.node-load-miss-rate%
2.99 -0.5 2.51 perf-stat.overall.node-store-miss-rate%
4.594e+09 +2.1% 4.69e+09 perf-stat.ps.branch-instructions
4.369e+08 +2.0% 4.456e+08 perf-stat.ps.cache-misses
6.572e+08 +0.9% 6.632e+08 perf-stat.ps.cache-references
8.016e+09 +1.8% 8.159e+09 perf-stat.ps.dTLB-loads
5.39e+09 +2.1% 5.505e+09 perf-stat.ps.dTLB-stores
5965014 +1.7% 6065670 perf-stat.ps.iTLB-load-misses
2.405e+10 +2.1% 2.456e+10 perf-stat.ps.instructions
5123634 +2.0% 5228497 perf-stat.ps.minor-faults
13550037 -11.5% 11998434 perf-stat.ps.node-load-misses
2.829e+08 +1.7% 2.877e+08 perf-stat.ps.node-loads
9802170 -14.4% 8389142 perf-stat.ps.node-store-misses
3.182e+08 +2.3% 3.255e+08 perf-stat.ps.node-stores
5123644 +2.0% 5228510 perf-stat.ps.page-faults
7.271e+12 +2.1% 7.426e+12 perf-stat.total.instructions
***************************************************************************************************
lkp-hsw-4ex1: 144 threads Intel(R) Xeon(R) CPU E7-8890 v3 @ 2.50GHz with 512G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase/ucode:
gcc-9/performance/x86_64-rhel-8.3/process/100%/debian-10.4-x86_64-20200603.cgz/lkp-hsw-4ex1/page_fault2/will-it-scale/0x16
commit:
fa02fcd94b ("Merge tag 'media/v5.10-2' of git://git.kernel.org/pub/scm/linux/kernel/git/mchehab/linux-media")
4df910620b ("mm: memcg: relayout structure mem_cgroup to avoid cache interference")
fa02fcd94b0c8dff 4df910620bebb5cfe234af16ac8
---------------- ---------------------------
fail:runs %reproduction fail:runs
| | |
3:4 5% 3:4 perf-profile.calltrace.cycles-pp.error_entry.testcase
3:4 6% 3:4 perf-profile.children.cycles-pp.error_entry
2:4 5% 3:4 perf-profile.self.cycles-pp.error_entry
%stddev %change %stddev
\ | \
10826219 +12.1% 12136182 will-it-scale.144.processes
75181 +12.1% 84278 will-it-scale.per_process_ops
6427 -0.8% 6374 will-it-scale.time.minor_page_faults
10826219 +12.1% 12136182 will-it-scale.workload
10304 ± 3% +12.7% 11616 ± 3% slabinfo.skbuff_head_cache.active_objs
10320 ± 3% +12.6% 11616 ± 3% slabinfo.skbuff_head_cache.num_objs
3952279 ± 6% -10.7% 3528860 ± 5% numa-meminfo.node0.AnonPages.max
3835803 ± 4% -10.6% 3428691 ± 3% numa-meminfo.node1.AnonPages.max
40533 ± 17% -25.1% 30339 ± 10% numa-meminfo.node3.KReclaimable
40533 ± 17% -25.1% 30339 ± 10% numa-meminfo.node3.SReclaimable
52626 ± 15% -21.2% 41452 ± 2% numa-meminfo.node3.SUnreclaim
93160 ± 15% -22.9% 71791 ± 4% numa-meminfo.node3.Slab
22237 ± 2% +3.7% 23055 proc-vmstat.nr_active_anon
22237 ± 2% +3.7% 23055 proc-vmstat.nr_zone_active_anon
3.267e+09 +12.2% 3.664e+09 proc-vmstat.numa_hit
3.267e+09 +12.2% 3.664e+09 proc-vmstat.numa_local
3.268e+09 +12.2% 3.667e+09 proc-vmstat.pgalloc_normal
3.255e+09 +12.2% 3.653e+09 proc-vmstat.pgfault
3.268e+09 +12.2% 3.667e+09 proc-vmstat.pgfree
8.355e+08 +9.7% 9.168e+08 numa-numastat.node0.local_node
8.355e+08 +9.7% 9.168e+08 numa-numastat.node0.numa_hit
8.106e+08 +13.1% 9.166e+08 numa-numastat.node1.local_node
8.106e+08 +13.1% 9.166e+08 numa-numastat.node1.numa_hit
8.092e+08 +13.4% 9.173e+08 numa-numastat.node2.local_node
8.092e+08 +13.4% 9.174e+08 numa-numastat.node2.numa_hit
8.124e+08 +12.8% 9.163e+08 numa-numastat.node3.local_node
8.124e+08 +12.8% 9.163e+08 numa-numastat.node3.numa_hit
4.171e+08 +10.0% 4.587e+08 numa-vmstat.node0.numa_hit
4.171e+08 +10.0% 4.587e+08 numa-vmstat.node0.numa_local
2026 +43.5% 2908 ± 25% numa-vmstat.node1.nr_mapped
4.054e+08 +13.3% 4.593e+08 numa-vmstat.node1.numa_hit
4.053e+08 +13.3% 4.592e+08 numa-vmstat.node1.numa_local
4.052e+08 +13.5% 4.598e+08 numa-vmstat.node2.numa_hit
4.051e+08 +13.5% 4.597e+08 numa-vmstat.node2.numa_local
10134 ± 17% -25.2% 7584 ± 10% numa-vmstat.node3.nr_slab_reclaimable
13156 ± 15% -21.2% 10362 ± 2% numa-vmstat.node3.nr_slab_unreclaimable
4.07e+08 +12.6% 4.582e+08 numa-vmstat.node3.numa_hit
4.068e+08 +12.6% 4.581e+08 numa-vmstat.node3.numa_local
585.75 ± 9% -21.5% 459.71 ± 3% sched_debug.cfs_rq:/.nr_spread_over.max
77.50 ± 8% -17.3% 64.13 ± 6% sched_debug.cfs_rq:/.nr_spread_over.stddev
608344 ± 41% +64.1% 998173 ± 12% sched_debug.cfs_rq:/.spread0.avg
811885 ± 31% +47.7% 1199394 ± 10% sched_debug.cfs_rq:/.spread0.max
-554391 -65.3% -192546 sched_debug.cfs_rq:/.spread0.min
2652761 ± 24% -28.7% 1890128 ± 22% sched_debug.cpu.avg_idle.max
233778 ± 17% -21.8% 182852 ± 22% sched_debug.cpu.avg_idle.stddev
21294 ± 37% -37.9% 13231 ± 43% sched_debug.cpu.max_idle_balance_cost.stddev
18116 ± 11% -19.0% 14665 ± 8% sched_debug.cpu.nr_switches.max
2627 ± 3% -11.9% 2315 ± 5% sched_debug.cpu.nr_switches.stddev
-21.46 -35.9% -13.75 sched_debug.cpu.nr_uninterruptible.min
14236 ± 14% -21.1% 11229 ± 8% sched_debug.cpu.sched_count.max
2290 ± 3% -13.4% 1983 ± 3% sched_debug.cpu.sched_count.stddev
7163 ± 7% -29.0% 5084 ± 6% sched_debug.cpu.ttwu_count.max
1060 ± 3% -17.4% 876.21 ± 3% sched_debug.cpu.ttwu_count.stddev
799.30 ± 8% -20.7% 634.07 ± 6% sched_debug.cpu.ttwu_local.stddev
56345 ± 20% -32.8% 37847 ± 10% syscalls.sys_close.max
10214940 ±129% -96.2% 386927 ± 29% syscalls.sys_mmap.max
3.599e+09 ± 4% -6.7e+08 2.933e+09 ± 4% syscalls.sys_mmap.noise.100%
3.664e+09 ± 4% -6.5e+08 3.019e+09 ± 4% syscalls.sys_mmap.noise.2%
3.655e+09 ± 4% -6.4e+08 3.011e+09 ± 4% syscalls.sys_mmap.noise.25%
3.664e+09 ± 4% -6.5e+08 3.019e+09 ± 4% syscalls.sys_mmap.noise.5%
3.637e+09 ± 4% -6.5e+08 2.988e+09 ± 4% syscalls.sys_mmap.noise.50%
3.616e+09 ± 4% -6.6e+08 2.96e+09 ± 4% syscalls.sys_mmap.noise.75%
3.562e+08 ± 4% +1.2e+08 4.787e+08 ± 31% syscalls.sys_openat.noise.100%
2119808 ± 57% -60.4% 840502 ± 24% syscalls.sys_write.max
2.979e+09 ± 44% -1.3e+09 1.709e+09 ± 5% syscalls.sys_write.noise.100%
3.002e+09 ± 44% -1.3e+09 1.735e+09 ± 5% syscalls.sys_write.noise.2%
2.999e+09 ± 44% -1.3e+09 1.733e+09 ± 5% syscalls.sys_write.noise.25%
3.002e+09 ± 44% -1.3e+09 1.735e+09 ± 5% syscalls.sys_write.noise.5%
2.995e+09 ± 44% -1.3e+09 1.728e+09 ± 5% syscalls.sys_write.noise.50%
2.989e+09 ± 44% -1.3e+09 1.723e+09 ± 5% syscalls.sys_write.noise.75%
134582 -16.9% 111883 ± 30% interrupts.CAL:Function_call_interrupts
5042 ± 34% +45.0% 7311 ± 18% interrupts.CPU0.NMI:Non-maskable_interrupts
5042 ± 34% +45.0% 7311 ± 18% interrupts.CPU0.PMI:Performance_monitoring_interrupts
276.00 ± 4% +21.8% 336.25 ± 12% interrupts.CPU0.RES:Rescheduling_interrupts
4063 +40.9% 5726 ± 28% interrupts.CPU105.NMI:Non-maskable_interrupts
4063 +40.9% 5726 ± 28% interrupts.CPU105.PMI:Performance_monitoring_interrupts
6629 ± 23% -39.0% 4045 ± 2% interrupts.CPU108.NMI:Non-maskable_interrupts
6629 ± 23% -39.0% 4045 ± 2% interrupts.CPU108.PMI:Performance_monitoring_interrupts
61.25 ± 94% -75.1% 15.25 ± 46% interrupts.CPU116.RES:Rescheduling_interrupts
168.50 ± 13% -43.8% 94.75 ± 18% interrupts.CPU126.RES:Rescheduling_interrupts
1522 ± 29% -58.1% 637.75 ± 30% interrupts.CPU129.CAL:Function_call_interrupts
79.25 ± 41% -60.9% 31.00 ± 31% interrupts.CPU131.RES:Rescheduling_interrupts
5031 ± 33% +53.7% 7733 ± 8% interrupts.CPU132.NMI:Non-maskable_interrupts
5031 ± 33% +53.7% 7733 ± 8% interrupts.CPU132.PMI:Performance_monitoring_interrupts
4054 +52.6% 6187 ± 27% interrupts.CPU137.NMI:Non-maskable_interrupts
4054 +52.6% 6187 ± 27% interrupts.CPU137.PMI:Performance_monitoring_interrupts
8008 -21.8% 6263 ± 21% interrupts.CPU21.NMI:Non-maskable_interrupts
8008 -21.8% 6263 ± 21% interrupts.CPU21.PMI:Performance_monitoring_interrupts
5988 ± 18% -25.2% 4479 ± 17% interrupts.CPU25.NMI:Non-maskable_interrupts
5988 ± 18% -25.2% 4479 ± 17% interrupts.CPU25.PMI:Performance_monitoring_interrupts
167.00 +13.8% 190.00 ± 3% interrupts.CPU30.RES:Rescheduling_interrupts
170.25 ± 8% +16.6% 198.50 ± 10% interrupts.CPU31.RES:Rescheduling_interrupts
908.75 ± 14% -31.4% 623.50 ± 45% interrupts.CPU39.CAL:Function_call_interrupts
4024 +44.1% 5799 ± 27% interrupts.CPU45.NMI:Non-maskable_interrupts
4024 +44.1% 5799 ± 27% interrupts.CPU45.PMI:Performance_monitoring_interrupts
4612 ± 21% +52.6% 7039 ± 23% interrupts.CPU54.NMI:Non-maskable_interrupts
4612 ± 21% +52.6% 7039 ± 23% interrupts.CPU54.PMI:Performance_monitoring_interrupts
175.75 ± 14% -43.5% 99.25 ± 27% interrupts.CPU54.RES:Rescheduling_interrupts
304.25 ± 82% -65.6% 104.75 ± 25% interrupts.CPU55.RES:Rescheduling_interrupts
6694 ± 19% -30.1% 4676 ± 23% interrupts.CPU58.NMI:Non-maskable_interrupts
6694 ± 19% -30.1% 4676 ± 23% interrupts.CPU58.PMI:Performance_monitoring_interrupts
1982 ± 26% -46.9% 1053 ± 19% interrupts.CPU73.CAL:Function_call_interrupts
5503 ± 27% -26.9% 4022 interrupts.CPU80.NMI:Non-maskable_interrupts
5503 ± 27% -26.9% 4022 interrupts.CPU80.PMI:Performance_monitoring_interrupts
6543 ± 23% -23.3% 5020 ± 34% interrupts.CPU88.NMI:Non-maskable_interrupts
6543 ± 23% -23.3% 5020 ± 34% interrupts.CPU88.PMI:Performance_monitoring_interrupts
6059 ± 33% -33.8% 4011 interrupts.CPU90.NMI:Non-maskable_interrupts
6059 ± 33% -33.8% 4011 interrupts.CPU90.PMI:Performance_monitoring_interrupts
16073 +14.0% 18329 ± 4% interrupts.RES:Rescheduling_interrupts
2.276e+10 +5.6% 2.403e+10 perf-stat.i.branch-instructions
57785027 +9.2% 63125578 perf-stat.i.branch-misses
33.59 +1.3 34.86 perf-stat.i.cache-miss-rate%
2.99e+08 +11.5% 3.332e+08 perf-stat.i.cache-misses
8.857e+08 +7.5% 9.52e+08 perf-stat.i.cache-references
4.00 -6.1% 3.76 perf-stat.i.cpi
1425 -10.3% 1277 perf-stat.i.cycles-between-cache-misses
38840111 ± 19% +45.7% 56590709 ± 18% perf-stat.i.dTLB-load-misses
2.805e+10 +6.9% 2.999e+10 perf-stat.i.dTLB-loads
86765075 ± 2% +17.0% 1.015e+08 perf-stat.i.dTLB-store-misses
9.928e+09 +12.1% 1.113e+10 perf-stat.i.dTLB-stores
32364704 +12.3% 36335103 perf-stat.i.iTLB-load-misses
1.029e+11 +6.3% 1.094e+11 perf-stat.i.instructions
3187 -5.5% 3013 perf-stat.i.instructions-per-iTLB-miss
0.25 +6.2% 0.27 perf-stat.i.ipc
429.12 +7.4% 460.92 perf-stat.i.metric.M/sec
10735651 +12.3% 12059205 perf-stat.i.minor-faults
5.53 -1.9 3.59 perf-stat.i.node-load-miss-rate%
12242331 -30.4% 8517490 perf-stat.i.node-load-misses
2.309e+08 +12.7% 2.603e+08 perf-stat.i.node-loads
14.81 -0.7 14.14 perf-stat.i.node-store-miss-rate%
8057655 +10.5% 8902557 perf-stat.i.node-store-misses
46809246 +16.6% 54566564 perf-stat.i.node-stores
10735652 +12.3% 12059206 perf-stat.i.page-faults
0.25 +0.0 0.26 perf-stat.overall.branch-miss-rate%
33.79 +1.2 35.02 perf-stat.overall.cache-miss-rate%
4.00 -6.0% 3.76 perf-stat.overall.cpi
1375 -10.3% 1233 perf-stat.overall.cycles-between-cache-misses
3182 -5.3% 3012 perf-stat.overall.instructions-per-iTLB-miss
0.25 +6.3% 0.27 perf-stat.overall.ipc
5.01 -1.9 3.16 perf-stat.overall.node-load-miss-rate%
14.69 -0.7 14.03 perf-stat.overall.node-store-miss-rate%
2874609 -5.2% 2726260 perf-stat.overall.path-length
2.266e+10 +5.5% 2.391e+10 perf-stat.ps.branch-instructions
57209629 +9.3% 62547538 perf-stat.ps.branch-misses
2.977e+08 +11.4% 3.318e+08 perf-stat.ps.cache-misses
8.813e+08 +7.5% 9.475e+08 perf-stat.ps.cache-references
38567680 ± 19% +45.8% 56225398 ± 18% perf-stat.ps.dTLB-load-misses
2.792e+10 +6.8% 2.983e+10 perf-stat.ps.dTLB-loads
86358963 ± 2% +17.0% 1.01e+08 perf-stat.ps.dTLB-store-misses
9.868e+09 +12.1% 1.106e+10 perf-stat.ps.dTLB-stores
32193889 +12.3% 36145969 perf-stat.ps.iTLB-load-misses
1.025e+11 +6.3% 1.089e+11 perf-stat.ps.instructions
10680325 +12.3% 11997386 perf-stat.ps.minor-faults
12145565 -30.4% 8458562 perf-stat.ps.node-load-misses
2.302e+08 +12.7% 2.593e+08 perf-stat.ps.node-loads
8004247 +10.6% 8849146 perf-stat.ps.node-store-misses
46470686 +16.7% 54225133 perf-stat.ps.node-stores
10680326 +12.3% 11997387 perf-stat.ps.page-faults
3.112e+13 +6.3% 3.308e+13 perf-stat.total.instructions
41.00 -6.5 34.52 ± 5% perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.do_fault.__handle_mm_fault.handle_mm_fault
41.07 -6.5 34.60 ± 5% perf-profile.calltrace.cycles-pp.finish_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
38.25 -6.1 32.14 ± 6% perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.lru_cache_add.alloc_set_pte.finish_fault.do_fault
38.38 -6.1 32.28 ± 6% perf-profile.calltrace.cycles-pp.lru_cache_add.alloc_set_pte.finish_fault.do_fault.__handle_mm_fault
36.53 -6.0 30.50 ± 6% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.pagevec_lru_move_fn.lru_cache_add.alloc_set_pte
36.57 -6.0 30.54 ± 6% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.pagevec_lru_move_fn.lru_cache_add.alloc_set_pte.finish_fault
4.88 -2.8 2.11 ± 3% perf-profile.calltrace.cycles-pp.mem_cgroup_charge.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
64.74 -1.7 63.03 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
65.63 -1.6 63.99 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
65.78 -1.6 64.15 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
72.33 -1.5 70.84 perf-profile.calltrace.cycles-pp.testcase
71.88 -1.3 70.54 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.testcase
63.02 -1.3 61.72 perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
63.45 -1.3 62.17 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
1.83 -1.2 0.58 perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.mem_cgroup_charge.do_fault.__handle_mm_fault.handle_mm_fault
1.17 ± 2% -0.4 0.73 perf-profile.calltrace.cycles-pp.page_add_new_anon_rmap.alloc_set_pte.finish_fault.do_fault.__handle_mm_fault
0.75 ± 4% -0.2 0.58 ± 3% perf-profile.calltrace.cycles-pp.page_remove_rmap.zap_pte_range.unmap_page_range.unmap_vmas.unmap_region
2.23 -0.1 2.15 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.release_pages.tlb_flush_mmu.tlb_finish_mmu.unmap_region
2.22 -0.1 2.15 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.release_pages.tlb_flush_mmu.tlb_finish_mmu
1.21 -0.1 1.14 ± 2% perf-profile.calltrace.cycles-pp.__pagevec_lru_add_fn.pagevec_lru_move_fn.lru_cache_add.alloc_set_pte.finish_fault
0.77 +0.0 0.80 perf-profile.calltrace.cycles-pp.__irqentry_text_end.testcase
0.81 +0.1 0.87 ± 3% perf-profile.calltrace.cycles-pp.try_charge.mem_cgroup_charge.do_fault.__handle_mm_fault.handle_mm_fault
1.28 +0.1 1.37 ± 2% perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.do_fault
1.10 +0.1 1.19 ± 2% perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
1.45 +0.1 1.56 ± 2% perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.do_fault.__handle_mm_fault
1.64 +0.1 1.77 ± 2% perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.do_fault.__handle_mm_fault.handle_mm_fault
1.73 +0.1 1.88 perf-profile.calltrace.cycles-pp.__do_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
2.70 +0.1 2.85 perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap
0.78 +0.2 0.94 ± 4% perf-profile.calltrace.cycles-pp.__list_del_entry_valid.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages_nodemask
2.74 +0.2 2.90 perf-profile.calltrace.cycles-pp.tlb_flush_mmu.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap
2.74 +0.2 2.90 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
0.38 ± 57% +0.2 0.54 ± 4% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge.mem_cgroup_charge.do_fault.__handle_mm_fault
11.34 +0.6 11.92 perf-profile.calltrace.cycles-pp.copy_page.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.76 ± 3% +1.1 1.90 ± 11% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.free_unref_page_list.release_pages
0.77 ± 4% +1.1 1.91 ± 11% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_pcppages_bulk.free_unref_page_list.release_pages.tlb_flush_mmu
21.43 +1.2 22.63 perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.zap_pte_range.unmap_page_range.unmap_vmas
23.92 +1.2 25.14 perf-profile.calltrace.cycles-pp.zap_pte_range.unmap_page_range.unmap_vmas.unmap_region.__do_munmap
23.93 +1.2 25.15 perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
23.93 +1.2 25.15 perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.__do_munmap.__vm_munmap
21.76 +1.2 22.99 perf-profile.calltrace.cycles-pp.tlb_flush_mmu.zap_pte_range.unmap_page_range.unmap_vmas.unmap_region
1.88 ± 2% +1.3 3.16 ± 7% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_list.release_pages.tlb_flush_mmu.zap_pte_range
2.15 +1.3 3.46 ± 7% perf-profile.calltrace.cycles-pp.free_unref_page_list.release_pages.tlb_flush_mmu.zap_pte_range.unmap_page_range
26.70 +1.4 28.07 perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
26.70 +1.4 28.07 perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
26.72 +1.4 28.09 perf-profile.calltrace.cycles-pp.__munmap
26.70 +1.4 28.07 perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
26.70 +1.4 28.07 perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
26.72 +1.4 28.09 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
26.70 +1.4 28.08 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
1.21 ± 9% +6.8 8.03 ± 18% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist
1.21 ± 8% +6.8 8.04 ± 18% perf-profile.calltrace.cycles-pp._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages_nodemask
2.13 ± 4% +7.0 9.12 ± 16% perf-profile.calltrace.cycles-pp.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma
2.58 ± 3% +7.1 9.65 ± 16% perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.do_fault
2.77 ± 3% +7.1 9.84 ± 15% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.do_fault.__handle_mm_fault
3.11 ± 3% +7.1 10.21 ± 15% perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.do_fault.__handle_mm_fault.handle_mm_fault
3.45 ± 3% +7.1 10.55 ± 14% perf-profile.calltrace.cycles-pp.alloc_pages_vma.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
41.03 -6.5 34.56 ± 5% perf-profile.children.cycles-pp.alloc_set_pte
41.08 -6.5 34.61 ± 5% perf-profile.children.cycles-pp.finish_fault
56.64 -6.1 50.52 ± 3% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
38.30 -6.1 32.18 ± 6% perf-profile.children.cycles-pp.pagevec_lru_move_fn
38.39 -6.1 32.30 ± 6% perf-profile.children.cycles-pp.lru_cache_add
4.90 -2.8 2.12 ± 3% perf-profile.children.cycles-pp.mem_cgroup_charge
64.76 -1.7 63.05 perf-profile.children.cycles-pp.handle_mm_fault
65.67 -1.6 64.03 perf-profile.children.cycles-pp.do_user_addr_fault
65.80 -1.6 64.17 perf-profile.children.cycles-pp.exc_page_fault
70.09 -1.5 68.63 perf-profile.children.cycles-pp.asm_exc_page_fault
73.11 -1.4 71.74 perf-profile.children.cycles-pp.testcase
63.05 -1.3 61.74 perf-profile.children.cycles-pp.do_fault
63.47 -1.3 62.19 perf-profile.children.cycles-pp.__handle_mm_fault
1.84 -1.2 0.59 perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
1.18 ± 4% -0.8 0.37 ± 4% perf-profile.children.cycles-pp.__count_memcg_events
1.53 ± 2% -0.8 0.72 ± 5% perf-profile.children.cycles-pp.__mod_memcg_state
2.07 ± 2% -0.7 1.33 ± 4% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
1.17 ± 2% -0.4 0.73 ± 2% perf-profile.children.cycles-pp.page_add_new_anon_rmap
0.56 ± 4% -0.3 0.23 ± 4% perf-profile.children.cycles-pp.mem_cgroup_charge_statistics
0.75 ± 4% -0.2 0.59 ± 3% perf-profile.children.cycles-pp.page_remove_rmap
1.22 -0.1 1.16 ± 2% perf-profile.children.cycles-pp.__pagevec_lru_add_fn
0.13 ± 3% -0.1 0.07 ± 5% perf-profile.children.cycles-pp.uncharge_page
0.29 -0.1 0.23 perf-profile.children.cycles-pp.mem_cgroup_uncharge_list
0.30 ± 3% -0.0 0.27 ± 4% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.05 +0.0 0.06 perf-profile.children.cycles-pp.rcu_all_qs
0.05 ± 8% +0.0 0.07 ± 7% perf-profile.children.cycles-pp.perf_exclude_event
0.05 +0.0 0.06 ± 6% perf-profile.children.cycles-pp.pte_alloc_one
0.14 ± 3% +0.0 0.16 ± 2% perf-profile.children.cycles-pp.scheduler_tick
0.07 ± 5% +0.0 0.09 ± 4% perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
0.19 ± 3% +0.0 0.21 ± 2% perf-profile.children.cycles-pp.update_process_times
0.17 ± 2% +0.0 0.19 ± 3% perf-profile.children.cycles-pp.__mod_zone_page_state
0.12 ± 5% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.cgroup_throttle_swaprate
0.21 ± 4% +0.0 0.23 ± 2% perf-profile.children.cycles-pp.tick_sched_timer
0.19 ± 2% +0.0 0.21 ± 2% perf-profile.children.cycles-pp.tick_sched_handle
0.27 +0.0 0.29 ± 2% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.28 ± 2% +0.0 0.30 perf-profile.children.cycles-pp.__mod_node_page_state
0.78 +0.0 0.80 perf-profile.children.cycles-pp.__irqentry_text_end
0.36 ± 2% +0.0 0.39 perf-profile.children.cycles-pp.__mod_lruvec_state
0.35 ± 2% +0.0 0.38 ± 4% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.42 +0.0 0.46 perf-profile.children.cycles-pp.___perf_sw_event
0.00 +0.1 0.05 perf-profile.children.cycles-pp.__tlb_remove_page_size
0.82 +0.1 0.87 ± 4% perf-profile.children.cycles-pp.try_charge
0.61 +0.1 0.68 perf-profile.children.cycles-pp.__perf_sw_event
2.40 +0.1 2.46 perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
1.11 +0.1 1.19 ± 2% perf-profile.children.cycles-pp.find_get_entry
1.30 +0.1 1.39 ± 2% perf-profile.children.cycles-pp.find_lock_entry
1.47 +0.1 1.58 ± 2% perf-profile.children.cycles-pp.shmem_getpage_gfp
1.64 +0.1 1.77 perf-profile.children.cycles-pp.shmem_fault
1.74 +0.1 1.88 perf-profile.children.cycles-pp.__do_fault
2.75 +0.2 2.91 perf-profile.children.cycles-pp.tlb_finish_mmu
1.54 ± 2% +0.2 1.76 ± 2% perf-profile.children.cycles-pp.__list_del_entry_valid
11.37 +0.6 11.96 perf-profile.children.cycles-pp.copy_page
23.93 +1.2 25.15 perf-profile.children.cycles-pp.unmap_vmas
23.93 +1.2 25.15 perf-profile.children.cycles-pp.unmap_page_range
23.93 +1.2 25.15 perf-profile.children.cycles-pp.zap_pte_range
24.27 +1.4 25.63 perf-profile.children.cycles-pp.release_pages
26.71 +1.4 28.08 perf-profile.children.cycles-pp.__do_munmap
26.71 +1.4 28.08 perf-profile.children.cycles-pp.unmap_region
26.80 +1.4 28.17 perf-profile.children.cycles-pp.do_syscall_64
26.82 +1.4 28.19 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
26.72 +1.4 28.09 perf-profile.children.cycles-pp.__munmap
26.70 +1.4 28.08 perf-profile.children.cycles-pp.__vm_munmap
26.70 +1.4 28.08 perf-profile.children.cycles-pp.__x64_sys_munmap
24.51 +1.4 25.90 perf-profile.children.cycles-pp.tlb_flush_mmu
2.14 +1.5 3.65 ± 7% perf-profile.children.cycles-pp.free_pcppages_bulk
2.45 +1.6 4.01 ± 7% perf-profile.children.cycles-pp.free_unref_page_list
58.69 +2.1 60.75 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
2.14 ± 4% +7.0 9.16 ± 16% perf-profile.children.cycles-pp.rmqueue_bulk
2.61 ± 4% +7.1 9.70 ± 16% perf-profile.children.cycles-pp.rmqueue
2.82 ± 3% +7.1 9.92 ± 15% perf-profile.children.cycles-pp.get_page_from_freelist
3.46 ± 3% +7.1 10.57 ± 14% perf-profile.children.cycles-pp.alloc_pages_vma
3.19 ± 3% +7.1 10.30 ± 15% perf-profile.children.cycles-pp.__alloc_pages_nodemask
3.29 ± 4% +8.2 11.52 ± 13% perf-profile.children.cycles-pp._raw_spin_lock
1.82 -1.2 0.57 perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
1.60 ± 3% -1.2 0.37 ± 4% perf-profile.self.cycles-pp.mem_cgroup_charge
1.18 ± 3% -0.8 0.36 ± 4% perf-profile.self.cycles-pp.__count_memcg_events
1.52 ± 3% -0.8 0.71 ± 5% perf-profile.self.cycles-pp.__mod_memcg_state
0.13 ± 3% -0.1 0.07 ± 5% perf-profile.self.cycles-pp.uncharge_page
0.10 +0.0 0.11 ± 3% perf-profile.self.cycles-pp.perf_swevent_get_recursion_context
0.07 ± 6% +0.0 0.08 perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
0.25 +0.0 0.26 perf-profile.self.cycles-pp.rmqueue
0.12 ± 4% +0.0 0.14 perf-profile.self.cycles-pp.lru_cache_add
0.18 ± 2% +0.0 0.20 ± 2% perf-profile.self.cycles-pp.shmem_fault
0.16 +0.0 0.18 ± 2% perf-profile.self.cycles-pp.shmem_getpage_gfp
0.10 +0.0 0.12 ± 5% perf-profile.self.cycles-pp.free_unref_page_list
0.28 +0.0 0.30 ± 3% perf-profile.self.cycles-pp.handle_mm_fault
0.16 +0.0 0.18 ± 2% perf-profile.self.cycles-pp.page_remove_rmap
0.41 +0.0 0.44 ± 2% perf-profile.self.cycles-pp.__handle_mm_fault
0.30 ± 2% +0.0 0.33 ± 3% perf-profile.self.cycles-pp.try_charge
0.78 +0.0 0.80 perf-profile.self.cycles-pp.__irqentry_text_end
0.26 ± 3% +0.0 0.29 perf-profile.self.cycles-pp.__mod_node_page_state
0.31 ± 2% +0.0 0.34 perf-profile.self.cycles-pp.___perf_sw_event
0.27 ± 3% +0.0 0.30 ± 3% perf-profile.self.cycles-pp.alloc_set_pte
0.34 ± 2% +0.0 0.38 ± 3% perf-profile.self.cycles-pp.free_pages_and_swap_cache
0.38 +0.0 0.41 ± 2% perf-profile.self.cycles-pp.release_pages
0.47 +0.0 0.52 ± 4% perf-profile.self.cycles-pp.page_counter_try_charge
0.74 +0.0 0.78 perf-profile.self.cycles-pp.find_get_entry
0.42 +0.0 0.47 ± 2% perf-profile.self.cycles-pp.__pagevec_lru_add_fn
2.30 +0.1 2.37 perf-profile.self.cycles-pp.irqentry_exit_to_user_mode
0.31 ± 22% +0.1 0.38 ± 21% perf-profile.self.cycles-pp.do_fault
0.54 +0.1 0.62 ± 3% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
2.10 +0.1 2.23 perf-profile.self.cycles-pp.testcase
1.02 +0.1 1.14 ± 3% perf-profile.self.cycles-pp.free_pcppages_bulk
1.30 +0.1 1.43 ± 3% perf-profile.self.cycles-pp.zap_pte_range
1.53 ± 2% +0.2 1.75 ± 2% perf-profile.self.cycles-pp.__list_del_entry_valid
11.31 +0.6 11.90 perf-profile.self.cycles-pp.copy_page
58.69 +2.1 60.75 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Oliver Sang
1 year, 4 months
[locking/qspinlock] 0e8d8f4f12: fsmark.files_per_sec 213.9% improvement
by kernel test robot
Greeting,
FYI, we noticed a 213.9% improvement of fsmark.files_per_sec due to commit:
commit: 0e8d8f4f1214cfbac219d6917b5f6460f818bb7c ("[PATCH v13 3/6] locking/qspinlock: Introduce CNA into the slow path of qspinlock")
url: https://github.com/0day-ci/linux/commits/Alex-Kogan/Add-NUMA-awareness-to...
base: https://git.kernel.org/cgit/linux/kernel/git/tip/tip.git cb262935a166bdef0ccfe6e2adffa00c0f2d038a
in testcase: fsmark
on test machine: 192 threads Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz with 192G memory
with following parameters:
iterations: 1x
nr_threads: 64t
disk: 1BRD_48G
fs: btrfs
filesize: 4M
test_size: 24G
sync_method: NoSync
cpufreq_governor: performance
ucode: 0x5003003
test-description: The fsmark is a file system benchmark to test synchronous write workloads, for example, mail servers workload.
test-url: https://sourceforge.net/projects/fsmark/
In addition to that, the commit also has significant impact on the following tests:
+------------------+------------------------------------------------------------------------+
| testcase: change | reaim: reaim.jobs_per_min 96.1% improvement |
| test machine | 144 threads Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory |
| test parameters | cpufreq_governor=performance |
| | nr_task=100% |
| | runtime=300s |
| | test=new_fserver |
| | ucode=0x5003003 |
+------------------+------------------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/disk/filesize/fs/iterations/kconfig/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase/ucode:
gcc-9/performance/1BRD_48G/4M/btrfs/1x/x86_64-rhel-8.3/64t/debian-10.4-x86_64-20200603.cgz/NoSync/lkp-csl-2ap2/24G/fsmark/0x5003003
commit:
cb45bab007 ("locking/qspinlock: Refactor the qspinlock slow path")
0e8d8f4f12 ("locking/qspinlock: Introduce CNA into the slow path of qspinlock")
cb45bab007ff0cfc 0e8d8f4f1214cfbac219d6917b5
---------------- ---------------------------
%stddev %change %stddev
\ | \
1180 ± 2% +213.9% 3706 ± 11% fsmark.files_per_sec
773.00 ± 4% -30.2% 539.75 ± 7% fsmark.time.percent_of_cpu_this_job_got
137.19 ± 3% -28.5% 98.13 ± 8% fsmark.time.system_time
48917 ± 4% -9.0% 44537 ± 8% meminfo.AnonHugePages
4.83 ± 4% -1.0 3.83 ± 5% mpstat.cpu.all.sys%
93.48 +1.1% 94.51 iostat.cpu.idle
5.76 ± 2% -16.9% 4.79 ± 6% iostat.cpu.system
93.00 +1.1% 94.00 vmstat.cpu.id
20606 ± 5% +97.4% 40682 ± 34% vmstat.system.cs
4237881 ± 29% +321.3% 17854499 ±106% cpuidle.C1.time
44169 ± 32% +396.5% 219318 ±113% cpuidle.C1.usage
72208 ± 6% +191.8% 210719 ± 39% cpuidle.POLL.time
36708 ± 9% +306.0% 149034 ± 48% cpuidle.POLL.usage
2157 ± 2% +13.8% 2454 ± 6% slabinfo.biovec-max.active_objs
2271 ± 2% +12.8% 2561 ± 5% slabinfo.biovec-max.num_objs
6624 +13.4% 7512 ± 5% slabinfo.btrfs_delayed_node.active_objs
6768 +12.1% 7589 ± 5% slabinfo.btrfs_delayed_node.num_objs
47815 ± 86% -89.2% 5149 ± 6% sched_debug.cfs_rq:/.load.stddev
142.69 ± 15% +66.8% 237.99 ± 13% sched_debug.cfs_rq:/.load_avg.avg
1598 ± 30% +92.4% 3075 ± 20% sched_debug.cfs_rq:/.load_avg.max
383.38 ± 21% +90.4% 730.10 ± 13% sched_debug.cfs_rq:/.load_avg.stddev
37712 ± 16% -27.1% 27508 ± 16% sched_debug.cfs_rq:/.min_vruntime.avg
58.89 ± 49% +57.8% 92.96 ± 5% sched_debug.cfs_rq:/.util_est_enqueued.stddev
1901 ± 4% -8.3% 1743 ± 3% proc-vmstat.nr_active_anon
5055884 +3.2% 5219972 proc-vmstat.nr_file_pages
4788961 ± 2% +3.4% 4953360 proc-vmstat.nr_inactive_file
36576 ± 2% -4.1% 35069 proc-vmstat.nr_kernel_stack
9756 -1.6% 9601 proc-vmstat.nr_mapped
5205 ± 2% -5.1% 4939 proc-vmstat.nr_shmem
1901 ± 4% -8.3% 1743 ± 3% proc-vmstat.nr_zone_active_anon
4788961 ± 2% +3.4% 4953360 proc-vmstat.nr_zone_inactive_file
4864 ± 5% -8.2% 4465 proc-vmstat.pgactivate
2.507e+09 -8.7% 2.29e+09 perf-stat.i.branch-instructions
71942409 -5.3% 68125925 ± 2% perf-stat.i.cache-misses
4.372e+10 -18.5% 3.565e+10 ± 3% perf-stat.i.cpu-cycles
1.231e+10 -7.5% 1.138e+10 perf-stat.i.instructions
2515 ± 8% -11.4% 2228 ± 5% perf-stat.i.instructions-per-iTLB-miss
0.23 -17.3% 0.19 ± 3% perf-stat.i.metric.GHz
0.71 ± 19% +37.6% 0.97 ± 17% perf-stat.i.metric.K/sec
35.43 -6.2% 33.24 ± 2% perf-stat.i.metric.M/sec
3.58 -8.5% 3.28 perf-stat.overall.cpi
612.62 -9.0% 557.36 ± 3% perf-stat.overall.cycles-between-cache-misses
2610 ± 6% -11.9% 2299 ± 4% perf-stat.overall.instructions-per-iTLB-miss
0.28 +9.3% 0.31 perf-stat.overall.ipc
2.39e+09 ± 2% -6.3% 2.239e+09 perf-stat.ps.branch-instructions
68534728 ± 2% -4.6% 65359103 ± 2% perf-stat.ps.cache-misses
4.199e+10 ± 2% -13.3% 3.641e+10 perf-stat.ps.cpu-cycles
1.173e+10 ± 2% -5.3% 1.112e+10 perf-stat.ps.instructions
226946 ± 18% +44.8% 328685 ± 19% numa-meminfo.node0.AnonPages
249710 ± 18% +41.9% 354296 ± 20% numa-meminfo.node0.AnonPages.max
233745 ± 17% +42.0% 331984 ± 20% numa-meminfo.node0.Inactive(anon)
8680 ± 73% +135.9% 20474 ± 28% numa-meminfo.node0.PageTables
250653 ± 2% +6.8% 267784 ± 5% numa-meminfo.node0.Unevictable
2890241 ± 9% -58.3% 1206212 ± 10% numa-meminfo.node0.Writeback
338312 ± 83% +111.9% 716979 ± 64% numa-meminfo.node1.Dirty
8444 ± 82% +169.8% 22785 ± 37% numa-meminfo.node2.AnonHugePages
236489 ± 16% +142.5% 573382 ± 36% numa-meminfo.node2.Dirty
2484617 ± 13% +117.6% 5406277 ± 50% numa-meminfo.node2.FilePages
2459002 ± 19% +115.1% 5290461 ± 51% numa-meminfo.node2.Inactive
2216704 ± 16% +131.7% 5135596 ± 53% numa-meminfo.node2.Inactive(file)
871702 ± 16% +52.8% 1331684 ± 6% numa-meminfo.node2.Writeback
3888 ± 16% -24.5% 2937 ± 6% numa-meminfo.node3.Active(anon)
57796 ± 19% -49.7% 29078 ± 34% numa-meminfo.node3.KReclaimable
57796 ± 19% -49.7% 29078 ± 34% numa-meminfo.node3.SReclaimable
130240 ± 7% -30.1% 91037 ± 15% numa-meminfo.node3.Slab
56616 ± 18% +45.2% 82199 ± 19% numa-vmstat.node0.nr_anon_pages
58314 ± 17% +42.4% 83024 ± 20% numa-vmstat.node0.nr_inactive_anon
2172 ± 73% +135.7% 5122 ± 28% numa-vmstat.node0.nr_page_table_pages
62662 ± 2% +6.8% 66945 ± 5% numa-vmstat.node0.nr_unevictable
738434 ± 9% -59.3% 300391 ± 10% numa-vmstat.node0.nr_writeback
58312 ± 17% +42.4% 83023 ± 20% numa-vmstat.node0.nr_zone_inactive_anon
62662 ± 2% +6.8% 66945 ± 5% numa-vmstat.node0.nr_zone_unevictable
926442 ± 6% -53.1% 434282 ± 12% numa-vmstat.node0.nr_zone_write_pending
84324 ± 81% +112.0% 178764 ± 64% numa-vmstat.node1.nr_dirty
427046 ± 84% +74.8% 746464 ± 46% numa-vmstat.node1.nr_zone_write_pending
627042 ± 18% +129.0% 1435752 ± 55% numa-vmstat.node2.nr_dirtied
59963 ± 18% +137.8% 142585 ± 36% numa-vmstat.node2.nr_dirty
634106 ± 15% +112.1% 1345185 ± 50% numa-vmstat.node2.nr_file_pages
567086 ± 17% +125.3% 1277507 ± 53% numa-vmstat.node2.nr_inactive_file
223407 ± 17% +48.3% 331299 ± 7% numa-vmstat.node2.nr_writeback
567087 ± 17% +125.3% 1277515 ± 53% numa-vmstat.node2.nr_zone_inactive_file
283372 ± 17% +67.2% 473906 ± 8% numa-vmstat.node2.nr_zone_write_pending
922.00 ± 19% -20.0% 737.25 ± 6% numa-vmstat.node3.nr_active_anon
14591 ± 20% -50.1% 7276 ± 35% numa-vmstat.node3.nr_slab_reclaimable
922.00 ± 19% -20.0% 737.25 ± 6% numa-vmstat.node3.nr_zone_active_anon
141466 ± 17% +523.4% 881964 ± 85% numa-vmstat.node3.numa_other
282981 ± 14% +46.6% 414731 ± 12% interrupts.CAL:Function_call_interrupts
770.00 ± 51% -56.5% 335.00 ± 59% interrupts.CPU0.NMI:Non-maskable_interrupts
770.00 ± 51% -56.5% 335.00 ± 59% interrupts.CPU0.PMI:Performance_monitoring_interrupts
346.25 ± 15% -46.7% 184.50 ± 23% interrupts.CPU110.NMI:Non-maskable_interrupts
346.25 ± 15% -46.7% 184.50 ± 23% interrupts.CPU110.PMI:Performance_monitoring_interrupts
1359 ± 14% +31.7% 1789 ± 17% interrupts.CPU128.CAL:Function_call_interrupts
1378 ± 14% +2294.9% 33020 ±164% interrupts.CPU153.CAL:Function_call_interrupts
214.50 ± 34% +91.7% 411.25 ± 39% interrupts.CPU16.NMI:Non-maskable_interrupts
214.50 ± 34% +91.7% 411.25 ± 39% interrupts.CPU16.PMI:Performance_monitoring_interrupts
403.50 ± 35% -40.3% 241.00 ± 8% interrupts.CPU169.NMI:Non-maskable_interrupts
403.50 ± 35% -40.3% 241.00 ± 8% interrupts.CPU169.PMI:Performance_monitoring_interrupts
439.00 ± 50% -51.5% 212.75 ± 22% interrupts.CPU170.NMI:Non-maskable_interrupts
439.00 ± 50% -51.5% 212.75 ± 22% interrupts.CPU170.PMI:Performance_monitoring_interrupts
1062 ±116% -77.3% 240.75 ± 23% interrupts.CPU174.NMI:Non-maskable_interrupts
1062 ±116% -77.3% 240.75 ± 23% interrupts.CPU174.PMI:Performance_monitoring_interrupts
291.25 ± 31% -34.2% 191.75 ± 30% interrupts.CPU2.NMI:Non-maskable_interrupts
291.25 ± 31% -34.2% 191.75 ± 30% interrupts.CPU2.PMI:Performance_monitoring_interrupts
342.00 ± 43% -58.3% 142.75 ± 16% interrupts.CPU22.NMI:Non-maskable_interrupts
342.00 ± 43% -58.3% 142.75 ± 16% interrupts.CPU22.PMI:Performance_monitoring_interrupts
241.25 ± 22% -42.6% 138.50 ± 22% interrupts.CPU23.NMI:Non-maskable_interrupts
241.25 ± 22% -42.6% 138.50 ± 22% interrupts.CPU23.PMI:Performance_monitoring_interrupts
1365 ± 14% +49.5% 2041 ± 24% interrupts.CPU35.CAL:Function_call_interrupts
222.00 ± 94% +175.1% 610.75 ± 51% interrupts.CPU63.NMI:Non-maskable_interrupts
222.00 ± 94% +175.1% 610.75 ± 51% interrupts.CPU63.PMI:Performance_monitoring_interrupts
146.75 ± 39% +104.6% 300.25 ± 32% interrupts.CPU64.NMI:Non-maskable_interrupts
146.75 ± 39% +104.6% 300.25 ± 32% interrupts.CPU64.PMI:Performance_monitoring_interrupts
557.00 ± 88% -61.3% 215.75 ± 21% interrupts.CPU74.NMI:Non-maskable_interrupts
557.00 ± 88% -61.3% 215.75 ± 21% interrupts.CPU74.PMI:Performance_monitoring_interrupts
319.25 ± 22% -43.9% 179.25 ± 36% interrupts.CPU75.NMI:Non-maskable_interrupts
319.25 ± 22% -43.9% 179.25 ± 36% interrupts.CPU75.PMI:Performance_monitoring_interrupts
1343 ± 15% +23.0% 1652 ± 4% interrupts.CPU80.CAL:Function_call_interrupts
1095 ± 66% -76.5% 257.00 ± 17% interrupts.CPU96.NMI:Non-maskable_interrupts
1095 ± 66% -76.5% 257.00 ± 17% interrupts.CPU96.PMI:Performance_monitoring_interrupts
817.25 ± 12% +62.6% 1329 ± 17% interrupts.RES:Rescheduling_interrupts
12.39 ± 23% -8.2 4.22 ± 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
12.35 ± 24% -8.1 4.21 ± 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
12.12 ± 24% -8.0 4.08 ± 6% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
12.11 ± 24% -8.0 4.07 ± 6% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
12.10 ± 24% -8.0 4.06 ± 6% perf-profile.calltrace.cycles-pp.btrfs_file_write_iter.new_sync_write.vfs_write.ksys_write.do_syscall_64
12.07 ± 24% -8.0 4.04 ± 6% perf-profile.calltrace.cycles-pp.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write.vfs_write.ksys_write
12.28 ± 24% -7.5 4.83 ± 5% perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.07 ± 31% -6.1 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata
6.36 ± 30% -4.8 1.55 ± 11% perf-profile.calltrace.cycles-pp.btrfs_delalloc_reserve_metadata.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write.vfs_write
6.31 ± 30% -4.8 1.51 ± 12% perf-profile.calltrace.cycles-pp.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.btrfs_buffered_write.btrfs_file_write_iter
6.31 ± 30% -4.8 1.51 ± 12% perf-profile.calltrace.cycles-pp.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write
6.13 ± 31% -4.8 1.38 ± 14% perf-profile.calltrace.cycles-pp._raw_spin_lock.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata.btrfs_buffered_write
4.39 ± 18% -3.2 1.21 ± 15% perf-profile.calltrace.cycles-pp.btrfs_inode_rsv_release.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write.vfs_write
4.38 ± 18% -3.2 1.20 ± 15% perf-profile.calltrace.cycles-pp.btrfs_block_rsv_release.btrfs_inode_rsv_release.btrfs_buffered_write.btrfs_file_write_iter.new_sync_write
4.21 ± 19% -3.1 1.09 ± 20% perf-profile.calltrace.cycles-pp._raw_spin_lock.btrfs_block_rsv_release.btrfs_inode_rsv_release.btrfs_buffered_write.btrfs_file_write_iter
0.13 ±173% +0.5 0.63 ± 6% perf-profile.calltrace.cycles-pp.tick_nohz_irq_exit.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
4.51 ± 6% +0.5 5.04 ± 3% perf-profile.calltrace.cycles-pp.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
4.48 ± 5% +0.5 5.02 ± 3% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
4.37 ± 6% +0.6 4.93 ± 3% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.asm_call_sysvec_on_stack.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.18 ±173% +0.6 0.75 ± 27% perf-profile.calltrace.cycles-pp.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write.cold.new_sync_write
0.18 ±173% +0.6 0.76 ± 27% perf-profile.calltrace.cycles-pp.vprintk_emit.devkmsg_emit.devkmsg_write.cold.new_sync_write.vfs_write
0.18 ±173% +0.6 0.76 ± 27% perf-profile.calltrace.cycles-pp.devkmsg_write.cold.new_sync_write.vfs_write.ksys_write.do_syscall_64
0.18 ±173% +0.6 0.76 ± 27% perf-profile.calltrace.cycles-pp.devkmsg_emit.devkmsg_write.cold.new_sync_write.vfs_write.ksys_write
0.18 ±173% +0.6 0.77 ± 27% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write
0.18 ±173% +0.6 0.77 ± 27% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
0.18 ±173% +0.6 0.77 ± 27% perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
0.18 ±173% +0.6 0.77 ± 27% perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
0.18 ±173% +0.6 0.77 ± 27% perf-profile.calltrace.cycles-pp.write
0.00 +0.6 0.62 ± 7% perf-profile.calltrace.cycles-pp.ktime_get.tick_nohz_irq_exit.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
0.00 +1.1 1.06 ± 21% perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.btrfs_block_rsv_release.btrfs_inode_rsv_release.btrfs_buffered_write
7.48 ± 5% +1.2 8.68 ± 2% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle
0.00 +1.3 1.33 ± 15% perf-profile.calltrace.cycles-pp.__cna_queued_spin_lock_slowpath._raw_spin_lock.__reserve_bytes.btrfs_reserve_metadata_bytes.btrfs_delalloc_reserve_metadata
11.79 ± 8% +2.8 14.63 ± 8% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
9.11 ± 14% +4.4 13.51 ± 18% perf-profile.calltrace.cycles-pp.menu_select.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
10.58 ± 26% -10.6 0.00 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
12.10 ± 24% -8.0 4.06 ± 6% perf-profile.children.cycles-pp.btrfs_file_write_iter
12.07 ± 24% -8.0 4.04 ± 6% perf-profile.children.cycles-pp.btrfs_buffered_write
14.00 ± 18% -7.9 6.09 ± 7% perf-profile.children.cycles-pp._raw_spin_lock
12.64 ± 23% -7.8 4.85 ± 5% perf-profile.children.cycles-pp.ksys_write
12.62 ± 24% -7.8 4.83 ± 5% perf-profile.children.cycles-pp.new_sync_write
12.63 ± 23% -7.8 4.85 ± 5% perf-profile.children.cycles-pp.vfs_write
13.17 ± 23% -7.6 5.60 ± 6% perf-profile.children.cycles-pp.do_syscall_64
13.24 ± 23% -7.3 5.92 ± 6% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
6.38 ± 30% -4.8 1.55 ± 12% perf-profile.children.cycles-pp.__reserve_bytes
6.33 ± 30% -4.8 1.52 ± 12% perf-profile.children.cycles-pp.btrfs_reserve_metadata_bytes
6.36 ± 30% -4.8 1.55 ± 11% perf-profile.children.cycles-pp.btrfs_delalloc_reserve_metadata
4.49 ± 18% -3.3 1.23 ± 15% perf-profile.children.cycles-pp.btrfs_block_rsv_release
4.39 ± 18% -3.2 1.21 ± 15% perf-profile.children.cycles-pp.btrfs_inode_rsv_release
0.16 ± 77% -0.1 0.06 ± 7% perf-profile.children.cycles-pp.do_filp_open
0.16 ± 77% -0.1 0.06 ± 7% perf-profile.children.cycles-pp.path_openat
0.16 ± 75% -0.1 0.06 ± 6% perf-profile.children.cycles-pp.do_sys_open
0.16 ± 75% -0.1 0.06 ± 6% perf-profile.children.cycles-pp.do_sys_openat2
0.21 ± 32% -0.1 0.13 ± 29% perf-profile.children.cycles-pp.rcu_sched_clock_irq
0.22 ± 9% -0.1 0.16 ± 6% perf-profile.children.cycles-pp.update_blocked_averages
0.10 ± 29% -0.0 0.05 ± 70% perf-profile.children.cycles-pp.update_ts_time_stats
0.07 ± 12% -0.0 0.04 ± 58% perf-profile.children.cycles-pp.__pagevec_lru_add_fn
0.16 ± 9% +0.0 0.20 ± 12% perf-profile.children.cycles-pp.brd_lookup_page
0.13 ± 11% +0.0 0.17 ± 18% perf-profile.children.cycles-pp.__radix_tree_lookup
0.14 ± 7% +0.1 0.20 ± 9% perf-profile.children.cycles-pp.__intel_pmu_enable_all
0.48 ± 8% +0.1 0.62 ± 8% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.51 ± 29% +0.3 0.77 ± 27% perf-profile.children.cycles-pp.write
0.39 ± 32% +0.3 0.65 ± 6% perf-profile.children.cycles-pp.tick_nohz_irq_exit
0.50 ± 30% +0.3 0.76 ± 27% perf-profile.children.cycles-pp.devkmsg_write.cold
0.50 ± 30% +0.3 0.76 ± 27% perf-profile.children.cycles-pp.devkmsg_emit
0.03 ±173% +0.3 0.29 ±108% perf-profile.children.cycles-pp.osq_lock
0.03 ±173% +0.3 0.29 ±107% perf-profile.children.cycles-pp.__mutex_lock
0.11 ±130% +0.3 0.45 ± 68% perf-profile.children.cycles-pp.__do_sys_finit_module
0.11 ±130% +0.3 0.45 ± 68% perf-profile.children.cycles-pp.load_module
0.11 ±130% +0.3 0.45 ± 67% perf-profile.children.cycles-pp.syscall
0.14 ± 88% +0.4 0.54 ± 53% perf-profile.children.cycles-pp.wb_workfn
0.14 ± 88% +0.4 0.54 ± 53% perf-profile.children.cycles-pp.wb_writeback
0.14 ± 88% +0.4 0.54 ± 53% perf-profile.children.cycles-pp.writeback_sb_inodes
0.12 ±105% +0.4 0.53 ± 54% perf-profile.children.cycles-pp.__writeback_single_inode
4.76 ± 5% +0.5 5.21 ± 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
4.65 ± 5% +0.5 5.13 ± 2% perf-profile.children.cycles-pp.hrtimer_interrupt
6.15 ± 5% +0.5 6.65 ± 2% perf-profile.children.cycles-pp.asm_call_sysvec_on_stack
2.06 ± 28% +1.0 3.05 ± 9% perf-profile.children.cycles-pp.ktime_get
7.80 ± 4% +1.1 8.94 ± 2% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
10.10 ± 6% +2.0 12.05 ± 6% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
0.00 +2.6 2.56 ± 19% perf-profile.children.cycles-pp.__cna_queued_spin_lock_slowpath
9.16 ± 14% +4.4 13.57 ± 18% perf-profile.children.cycles-pp.menu_select
10.50 ± 26% -10.5 0.00 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.18 ± 30% -0.1 0.11 ± 27% perf-profile.self.cycles-pp.rcu_sched_clock_irq
0.10 ± 11% +0.0 0.13 ± 10% perf-profile.self.cycles-pp.__extent_writepage
0.14 ± 7% +0.1 0.20 ± 9% perf-profile.self.cycles-pp.__intel_pmu_enable_all
0.08 ± 10% +0.1 0.16 ± 22% perf-profile.self.cycles-pp.end_page_writeback
0.03 ±173% +0.2 0.28 ±105% perf-profile.self.cycles-pp.osq_lock
1.77 ± 32% +1.0 2.75 ± 10% perf-profile.self.cycles-pp.ktime_get
0.00 +2.5 2.51 ± 19% perf-profile.self.cycles-pp.__cna_queued_spin_lock_slowpath
7.68 ± 15% +4.1 11.79 ± 21% perf-profile.self.cycles-pp.menu_select
fsmark.files_per_sec
4500 +--------------------------------------------------------------------+
| O O O O O O |
4000 |-+O O |
3500 |-+ O O O O O O |
| O O O O O |
3000 |-+ O |
2500 |-+ |
| |
2000 |-+ |
1500 |-+ |
|.. .+..+..+.+..+..+.+..+..+.+..+..+.+..+.+..+..+.+..+..+.+..+..+.+..|
1000 |-++ |
500 |-+ |
| |
0 +--------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
***************************************************************************************************
lkp-cpl-4sp1: 144 threads Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Oliver Sang
1 year, 5 months
[drm/i915/gt] 8c3b1ba0e7: perf-sanity-tests.Parse_event_definition_strings.fail
by kernel test robot
Greeting,
FYI, we noticed the following commit (built with gcc-9):
commit: 8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813 ("drm/i915/gt: Track the overall awake/busy time")
git://anongit.freedesktop.org/drm-intel drm-intel-gt-next
in testcase: perf-sanity-tests
version: perf-x86_64-fa02fcd94b0c-1_20201126
with following parameters:
perf_compiler: gcc
ucode: 0xe2
on test machine: 4 threads Intel(R) Core(TM) i5-6500 CPU @ 3.20GHz with 32G memory
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang(a)intel.com>
2020-12-24 12:50:16 make ARCH= EXTRA_CFLAGS=-fno-omit-frame-pointer -fsanitize=undefined -fsanitize=address -C /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf
make: Entering directory '/usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf'
BUILD: Doing 'make [33m-j4[m' parallel build
HOSTCC fixdep.o
HOSTLD fixdep-in.o
LINK fixdep
diff -u tools/include/uapi/drm/i915_drm.h include/uapi/drm/i915_drm.h
diff -u tools/arch/x86/lib/memcpy_64.S arch/x86/lib/memcpy_64.S
diff -u tools/arch/x86/lib/memset_64.S arch/x86/lib/memset_64.S
Auto-detecting system features:
... dwarf: [ [32mon[m ]
... dwarf_getlocations: [ [32mon[m ]
... glibc: [ [32mon[m ]
... libbfd: [ [32mon[m ]
... libbfd-buildid: [ [32mon[m ]
... libcap: [ [32mon[m ]
... libelf: [ [32mon[m ]
... libnuma: [ [32mon[m ]
... numa_num_possible_cpus: [ [32mon[m ]
... libperl: [ [32mon[m ]
... libpython: [ [32mon[m ]
... libcrypto: [ [32mon[m ]
... libunwind: [ [32mon[m ]
... libdw-dwarf-unwind: [ [32mon[m ]
... zlib: [ [32mon[m ]
... lzma: [ [32mon[m ]
... get_cpuid: [ [32mon[m ]
... bpf: [ [32mon[m ]
... libaio: [ [32mon[m ]
... libzstd: [ [32mon[m ]
... disassembler-four-args: [ [32mon[m ]
CC fd/array.o
CC exec-cmd.o
CC event-parse.o
CC core.o
LD fd/libapi-in.o
CC fs/fs.o
CC cpumap.o
CC help.o
CC fs/tracing_path.o
CC fs/cgroup.o
LD fs/libapi-in.o
CC cpu.o
CC threadmap.o
CC debug.o
CC pager.o
CC str_error_r.o
LD libapi-in.o
AR libapi.a
CC parse-options.o
CC event-plugin.o
CC evsel.o
CC trace-seq.o
CC parse-filter.o
CC evlist.o
CC mmap.o
CC parse-utils.o
CC kbuffer-parse.o
CC zalloc.o
CC tep_strerror.o
CC event-parse-api.o
CC xyarray.o
LD libtraceevent-in.o
LINK libtraceevent.a
CC lib.o
GEN bpf_helper_defs.h
MKDIR staticobjs/
CC staticobjs/libbpf.o
LD libperf-in.o
AR libperf.a
GEN common-cmds.h
CC run-command.o
CC sigchain.o
CC subcmd-config.o
LD libsubcmd-in.o
MKDIR staticobjs/
AR libsubcmd.a
CC staticobjs/bpf.o
CC staticobjs/nlattr.o
CC staticobjs/btf.o
HOSTCC pmu-events/json.o
HOSTCC pmu-events/jsmn.o
HOSTCC pmu-events/jevents.o
HOSTLD pmu-events/jevents-in.o
CC plugin_jbd2.o
LD plugin_jbd2-in.o
CC plugin_hrtimer.o
LD plugin_hrtimer-in.o
CC plugin_kmem.o
LD plugin_kmem-in.o
CC plugin_kvm.o
LD plugin_kvm-in.o
CC staticobjs/libbpf_errno.o
CC plugin_mac80211.o
CC staticobjs/str_error.o
CC staticobjs/netlink.o
LD plugin_mac80211-in.o
CC plugin_sched_switch.o
LD plugin_sched_switch-in.o
CC plugin_function.o
CC staticobjs/bpf_prog_linfo.o
LD plugin_function-in.o
CC plugin_futex.o
CC staticobjs/libbpf_probes.o
LD plugin_futex-in.o
CC plugin_xen.o
CC staticobjs/xsk.o
LD plugin_xen-in.o
CC plugin_scsi.o
CC staticobjs/hashmap.o
LD plugin_scsi-in.o
CC plugin_tlb.o
CC plugin_cfg80211.o
CC jvmti/libjvmti.o
LD plugin_cfg80211-in.o
GEN perf-archive
LD plugin_tlb-in.o
LINK plugin_jbd2.so
GEN perf-with-kcore
CC jvmti/jvmti_agent.o
CC staticobjs/btf_dump.o
LINK plugin_hrtimer.so
LINK plugin_kmem.so
LINK plugin_kvm.so
LINK plugin_mac80211.so
CC staticobjs/ringbuf.o
LINK plugin_sched_switch.so
LINK plugin_function.so
LINK plugin_futex.so
LINK plugin_xen.so
LD staticobjs/libbpf-in.o
LINK libbpf.a
LINK plugin_scsi.so
CC jvmti/libstring.o
LINK plugin_cfg80211.so
CC builtin-bench.o
LINK plugin_tlb.so
GEN libtraceevent-dynamic-list
CC jvmti/libctype.o
CC builtin-annotate.o
CC builtin-config.o
LD jvmti/jvmti-in.o
CC builtin-diff.o
LINK pmu-events/jevents
CC builtin-evlist.o
CC builtin-ftrace.o
LINK libperf-jvmti.so
CC builtin-help.o
CC builtin-sched.o
CC builtin-buildid-list.o
CC builtin-buildid-cache.o
CC builtin-kallsyms.o
CC builtin-list.o
CC builtin-record.o
CC builtin-report.o
CC builtin-stat.o
CC builtin-timechart.o
CC builtin-top.o
CC builtin-script.o
CC builtin-kmem.o
CC builtin-lock.o
CC builtin-kvm.o
CC builtin-inject.o
CC builtin-mem.o
CC builtin-data.o
CC builtin-version.o
CC builtin-c2c.o
CC builtin-trace.o
CC builtin-probe.o
CC bench/sched-messaging.o
CC bench/sched-pipe.o
CC bench/syscall.o
CC bench/mem-functions.o
CC bench/futex-hash.o
CC bench/futex-wake.o
CC bench/futex-wake-parallel.o
CC bench/futex-requeue.o
CC bench/futex-lock-pi.o
CC tests/builtin-test.o
CC bench/epoll-wait.o
CC bench/epoll-ctl.o
CC tests/parse-events.o
CC util/annotate.o
CC bench/synthesize.o
CC bench/kallsyms-parse.o
CC bench/find-bit-bench.o
CC bench/inject-buildid.o
CC bench/mem-memcpy-x86-64-asm.o
CC bench/mem-memset-x86-64-asm.o
CC bench/numa.o
CC tests/dso-data.o
CC tests/attr.o
CC tests/vmlinux-kallsyms.o
LD bench/perf-in.o
CC arch/common.o
CC tests/openat-syscall.o
CC arch/x86/util/header.o
CC tests/openat-syscall-all-cpus.o
CC arch/x86/util/tsc.o
CC arch/x86/util/pmu.o
CC tests/openat-syscall-tp-fields.o
CC arch/x86/util/kvm-stat.o
CC tests/mmap-basic.o
CC arch/x86/util/perf_regs.o
CC arch/x86/util/topdown.o
CC tests/perf-record.o
CC arch/x86/util/machine.o
CC arch/x86/tests/regs_load.o
CC arch/x86/tests/dwarf-unwind.o
CC util/block-info.o
CC arch/x86/tests/arch-tests.o
CC arch/x86/util/event.o
CC arch/x86/tests/rdpmc.o
CC tests/evsel-roundtrip-name.o
CC arch/x86/util/dwarf-regs.o
CC arch/x86/util/unwind-libunwind.o
CC arch/x86/tests/perf-time-to-tsc.o
CC arch/x86/util/auxtrace.o
CC tests/evsel-tp-sched.o
CC arch/x86/util/archinsn.o
CC tests/fdarray.o
CC arch/x86/tests/insn-x86.o
CC arch/x86/util/intel-pt.o
CC arch/x86/tests/intel-pt-pkt-decoder-test.o
CC tests/pmu.o
CC arch/x86/tests/bp-modify.o
CC util/block-range.o
CC tests/pmu-events.o
LD arch/x86/tests/perf-in.o
CC tests/hists_common.o
CC tests/hists_link.o
CC util/build-id.o
CC ui/setup.o
CC tests/hists_filter.o
CC arch/x86/util/intel-bts.o
CC ui/helpline.o
CC ui/progress.o
CC ui/util.o
CC ui/hist.o
CC tests/hists_output.o
CC util/cacheline.o
CC util/config.o
LD arch/x86/util/perf-in.o
LD arch/x86/perf-in.o
LD arch/perf-in.o
CC tests/hists_cumulate.o
CC util/copyfile.o
CC util/ctype.o
CC util/db-export.o
CC tests/python-use.o
CC tests/bp_signal.o
CC tests/bp_signal_overflow.o
CC tests/bp_account.o
CC tests/wp.o
CC tests/task-exit.o
CC tests/sw-clock.o
CC tests/mmap-thread-lookup.o
CC util/env.o
CC scripts/perl/Perf-Trace-Util/Context.o
CC tests/thread-maps-share.o
CC util/event.o
CC tests/switch-tracking.o
LD scripts/perl/Perf-Trace-Util/perf-in.o
CC scripts/python/Perf-Trace-Util/Context.o
LD scripts/python/Perf-Trace-Util/perf-in.o
LD scripts/perf-in.o
CC trace/beauty/clone.o
CC tests/keep-tracking.o
CC trace/beauty/fcntl.o
CC tests/code-reading.o
CC trace/beauty/flock.o
CC util/evlist.o
CC trace/beauty/fsmount.o
CC trace/beauty/fspick.o
CC trace/beauty/ioctl.o
CC trace/beauty/kcmp.o
CC trace/beauty/mount_flags.o
CC trace/beauty/move_mount.o
CC trace/beauty/pkey_alloc.o
CC tests/sample-parsing.o
CC trace/beauty/arch_prctl.o
CC trace/beauty/prctl.o
CC trace/beauty/renameat.o
CC trace/beauty/sockaddr.o
CC trace/beauty/socket.o
CC trace/beauty/statx.o
CC tests/parse-no-sample-id-all.o
CC trace/beauty/sync_file_range.o
CC tests/kmod-path.o
CC trace/beauty/tracepoints/x86_irq_vectors.o
CC trace/beauty/tracepoints/x86_msr.o
LD trace/beauty/tracepoints/perf-in.o
LD trace/beauty/perf-in.o
CC perf.o
CC tests/thread-map.o
CC tests/llvm.o
CC tests/bpf.o
CC util/sideband_evlist.o
CC util/evsel.o
CC util/evsel_fprintf.o
CC tests/topology.o
CC tests/mem.o
CC tests/cpumap.o
CC tests/stat.o
CC tests/event_update.o
CC ui/stdio/hist.o
CC util/perf_event_attr_fprintf.o
CC tests/event-times.o
CC tests/expr.o
CC util/evswitch.o
CC tests/backward-ring-buffer.o
CC util/find_bit.o
CC tests/sdt.o
CC util/get_current_dir_name.o
CC util/kallsyms.o
CC tests/is_printable_array.o
CC tests/bitmap.o
CC util/levenshtein.o
CC ui/browser.o
CC tests/perf-hooks.o
CC tests/clang.o
CC util/llvm-utils.o
CC tests/unit_number__scnprintf.o
CC tests/mem2node.o
CC tests/maps.o
CC util/mmap.o
CC tests/time-utils-test.o
CC util/memswap.o
CC ui/browsers/annotate.o
BISON util/parse-events-bison.c
CC util/perf_regs.o
CC util/path.o
CC tests/genelf.o
CC ui/browsers/hists.o
CC tests/api-io.o
CC tests/demangle-java-test.o
CC tests/pfm.o
CC tests/parse-metric.o
CC util/print_binary.o
CC util/rlimit.o
CC tests/pe-file-parsing.o
CC util/argv_split.o
CC tests/expand-cgroup.o
CC ui/browsers/map.o
CC util/rbtree.o
CC ui/browsers/scripts.o
CC tests/dwarf-unwind.o
CC ui/browsers/header.o
CC tests/llvm-src-base.o
CC tests/llvm-src-kbuild.o
CC tests/llvm-src-prologue.o
CC tests/llvm-src-relocation.o
CC ui/browsers/res_sample.o
LD tests/perf-in.o
CC util/libstring.o
CC util/bitmap.o
CC util/hweight.o
CC util/smt.o
CC ui/tui/setup.o
CC util/strbuf.o
CC util/string.o
CC ui/tui/util.o
CC util/strlist.o
CC util/strfilter.o
CC ui/tui/helpline.o
CC util/top.o
CC ui/tui/progress.o
LD ui/tui/perf-in.o
CC util/usage.o
CC util/dso.o
CC util/dsos.o
CC util/symbol.o
CC util/symbol_fprintf.o
CC util/color.o
CC util/color_config.o
CC util/metricgroup.o
CC util/header.o
CC util/callchain.o
CC util/values.o
CC util/debug.o
CC util/fncache.o
CC util/machine.o
CC util/map.o
CC util/pstack.o
CC util/session.o
CC util/sample-raw.o
CC util/s390-sample-raw.o
CC util/syscalltbl.o
CC util/ordered-events.o
CC util/namespaces.o
LD ui/browsers/perf-in.o
LD ui/perf-in.o
CC util/comm.o
CC util/thread.o
CC util/thread_map.o
CC util/trace-event-parse.o
CC util/parse-events-bison.o
BISON util/pmu-bison.c
CC util/trace-event-read.o
CC util/trace-event-info.o
GEN python/perf.so
CC util/trace-event-scripting.o
CC util/trace-event.o
CC util/svghelper.o
CC util/sort.o
CC util/hist.o
CC util/util.o
CC util/cpumap.o
CC util/affinity.o
CC util/cputopo.o
CC util/cgroup.o
CC util/target.o
CC util/rblist.o
CC util/intlist.o
CC util/vdso.o
CC util/counts.o
CC util/stat.o
CC util/stat-shadow.o
CC util/stat-display.o
CC util/perf_api_probe.o
CC util/record.o
CC util/srcline.o
CC util/srccode.o
CC util/synthetic-events.o
CC util/data.o
CC util/tsc.o
CC util/cloexec.o
CC util/call-path.o
CC util/rwsem.o
CC util/thread-stack.o
CC util/spark.o
CC util/topdown.o
CC util/stream.o
CC util/auxtrace.o
CC util/intel-pt-decoder/intel-pt-pkt-decoder.o
GEN util/intel-pt-decoder/inat-tables.c
CC util/intel-pt-decoder/intel-pt-log.o
CC util/intel-pt-decoder/intel-pt-decoder.o
CC util/arm-spe-decoder/arm-spe-pkt-decoder.o
CC util/arm-spe-decoder/arm-spe-decoder.o
LD util/arm-spe-decoder/perf-in.o
CC util/scripting-engines/trace-event-perl.o
GEN pmu-events/pmu-events.c
CC pmu-events/pmu-events.o
CC util/scripting-engines/trace-event-python.o
CC util/intel-pt.o
LD pmu-events/pmu-events-in.o
CC util/intel-bts.o
CC util/arm-spe.o
CC util/intel-pt-decoder/intel-pt-insn-decoder.o
LD util/scripting-engines/perf-in.o
CC util/s390-cpumsf.o
CC util/parse-branch-options.o
CC util/dump-insn.o
CC util/parse-regs-options.o
CC util/parse-sublevel-options.o
CC util/term.o
CC util/help-unknown-cmd.o
CC util/mem-events.o
CC util/vsprintf.o
LD util/intel-pt-decoder/perf-in.o
CC util/units.o
CC util/time-utils.o
BISON util/expr-bison.c
CC util/branch.o
CC util/mem2node.o
CC util/clockid.o
CC util/bpf-loader.o
CC util/bpf_map.o
CC util/bpf-prologue.o
CC util/symbol-elf.o
CC util/probe-file.o
CC util/probe-event.o
CC util/probe-finder.o
CC util/dwarf-aux.o
CC util/dwarf-regs.o
CC util/unwind-libunwind-local.o
CC util/unwind-libunwind.o
CC util/data-convert-bt.o
CC util/zlib.o
CC util/lzma.o
CC util/zstd.o
CC util/cap.o
CC util/demangle-java.o
CC util/demangle-rust.o
CC util/jitdump.o
CC util/genelf.o
CC util/genelf_debug.o
CC util/perf-hooks.o
CC util/bpf-event.o
FLEX util/parse-events-flex.c
FLEX util/pmu-flex.c
CC util/pmu-bison.o
FLEX util/expr-flex.c
CC util/expr-bison.o
CC util/expr.o
CC util/parse-events.o
CC util/parse-events-flex.o
CC util/pmu.o
CC util/pmu-flex.o
CC util/expr-flex.o
LD util/perf-in.o
LD perf-in.o
LINK perf
make: Leaving directory '/usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf'
2020-12-24 12:51:30 cd /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf
2020-12-24 12:51:30 mkdir -p /pkg
2020-12-24 12:51:30 mkdir -p /kbuild/obj/consumer/x86_64-rhel-8.3
2020-12-24 12:51:30 cp /pkg/linux/x86_64-rhel-8.3/gcc-9/8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/vmlinux.xz /tmp
2020-12-24 12:51:31 unxz -k /tmp/vmlinux.xz
2020-12-24 12:51:35 cp /tmp/vmlinux /kbuild/obj/consumer/x86_64-rhel-8.3
ignored_by_lkp: BPF filter
ignored_by_lkp: LLVM search and compile
ignored_by_lkp: Add vfs_getname probe to get syscall args filenames
ignored_by_lkp: Use vfs_getname probe to get syscall args filenames
ignored_by_lkp: Check open filename arg using perf trace + vfs_getname
ignored_by_lkp: builtin clang support
2020-12-24 12:51:35 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 1
1: vmlinux symtab matches kallsyms : Ok
2020-12-24 12:51:36 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 2
2: Detect openat syscall event : Ok
2020-12-24 12:51:36 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 3
3: Detect openat syscall event on all cpus : Ok
2020-12-24 12:51:36 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 4
4: Read samples using the mmap interface : FAILED!
2020-12-24 12:51:36 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 5
5: Test data source output : Ok
2020-12-24 12:51:36 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 6
6: Parse event definition strings : FAILED!
2020-12-24 12:51:37 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 7
7: Simple expression parser : Ok
2020-12-24 12:51:38 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 8
8: PERF_RECORD_* events & perf_sample fields : Ok
2020-12-24 12:51:40 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 9
9: Parse perf pmu format : Ok
2020-12-24 12:51:40 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 10
10: PMU events :
10.1: PMU event table sanity : Ok
10.2: PMU event map aliases : Ok
10.3: Parsing of PMU event table metrics : Skip (some metrics failed)
10.4: Parsing of PMU event table metrics with fake PMUs : Ok
2020-12-24 12:51:42 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 11
11: DSO data read : Ok
2020-12-24 12:51:42 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 12
12: DSO data cache : Ok
2020-12-24 12:51:42 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 13
13: DSO data reopen : Ok
2020-12-24 12:51:42 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 14
14: Roundtrip evsel->name : Ok
2020-12-24 12:51:42 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 15
15: Parse sched tracepoints fields : Ok
2020-12-24 12:51:42 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 16
16: syscalls:sys_enter_openat event fields : Ok
2020-12-24 12:51:42 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 17
17: Setup struct perf_event_attr : Ok
2020-12-24 12:51:56 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 18
18: Match and link multiple hists : Ok
2020-12-24 12:51:56 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 19
19: 'import perf' in python : FAILED!
2020-12-24 12:51:56 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 20
20: Breakpoint overflow signal handler : Ok
2020-12-24 12:51:56 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 21
21: Breakpoint overflow sampling : Ok
2020-12-24 12:51:57 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 22
22: Breakpoint accounting : Ok
2020-12-24 12:51:57 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 23
23: Watchpoint :
23.1: Read Only Watchpoint : Skip
23.2: Write Only Watchpoint : Ok
23.3: Read / Write Watchpoint : Ok
23.4: Modify Watchpoint : Ok
2020-12-24 12:51:57 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 24
24: Number of exit events of a simple workload : FAILED!
2020-12-24 12:51:57 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 25
25: Software clock events period values : FAILED!
2020-12-24 12:51:57 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 26
26: Object code reading : FAILED!
2020-12-24 12:52:00 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 27
27: Sample parsing : Ok
2020-12-24 12:52:00 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 28
28: Use a dummy software event to keep tracking : FAILED!
2020-12-24 12:52:00 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 29
29: Parse with no sample_id_all bit set : Ok
2020-12-24 12:52:00 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 30
30: Filter hist entries : Ok
2020-12-24 12:52:00 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 31
31: Lookup mmap thread : FAILED!
2020-12-24 12:52:01 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 32
32: Share thread maps : Ok
2020-12-24 12:52:01 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 33
33: Sort output of hist entries : Ok
2020-12-24 12:52:01 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 34
34: Cumulate child hist entries : Ok
2020-12-24 12:52:01 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 35
35: Track with sched_switch : FAILED!
2020-12-24 12:52:01 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 36
36: Filter fds with revents mask in a fdarray : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 37
37: Add fd to a fdarray, making it autogrow : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 38
38: kmod_path__parse : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 39
39: Thread map : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 41
41: Session topology : FAILED!
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 43
43: Synthesize thread map : FAILED!
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 44
44: Remove thread map : FAILED!
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 45
45: Synthesize cpu map : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 46
46: Synthesize stat config : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 47
47: Synthesize stat : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 48
48: Synthesize stat round : Ok
2020-12-24 12:52:02 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 49
49: Synthesize attr update : FAILED!
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 50
50: Event times : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 51
51: Read backward ring buffer : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 52
52: Print cpu map : FAILED!
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 53
53: Merge cpu map : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 54
54: Probe SDT events : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 55
55: is_printable_array : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 56
56: Print bitmap : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 57
57: perf hooks : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 59
59: unit_number__scnprintf : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 60
60: mem2node : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 61
61: time utils : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 62
62: Test jit_write_elf : Ok
2020-12-24 12:52:03 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 63
63: Test libpfm4 support : Skip (not compiled in)
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 64
64: Test api io : Ok
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 65
65: maps__merge_in : FAILED!
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 66
66: Demangle Java : Ok
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 67
67: Parse and process metrics : Ok
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 68
68: PE file support : FAILED!
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 69
69: Event expansion for cgroups : Ok
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 70
70: x86 rdpmc : Ok
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 71
71: Convert perf time to TSC : FAILED!
2020-12-24 12:52:04 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 72
72: DWARF unwind : FAILED!
2020-12-24 12:52:05 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 73
73: x86 instruction decoder - new instructions : Ok
2020-12-24 12:52:05 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 74
74: Intel PT packet decoder : Ok
2020-12-24 12:52:05 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 75
75: x86 bp modify : Ok
2020-12-24 12:52:05 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 76
76: probe libc's inet_pton & backtrace it with ping : Ok
2020-12-24 12:52:07 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 77
77: Check Arm CoreSight trace data recording and synthesized samples: Skip
2020-12-24 12:52:07 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 78
78: build id cache operations : FAILED!
2020-12-24 12:52:08 sudo /usr/src/perf_selftests-x86_64-rhel-8.3-8c3b1ba0e7ea9a80b0ee4b4445ea59c806787813/tools/perf/perf test 81
81: Zstd perf.data compression/decompression : FAILED!
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Thanks,
Oliver Sang
1 year, 5 months
[workqueue] 6094661b16: WARNING:at_kernel/sched/core.c:#__set_cpus_allowed_ptr
by kernel test robot
Greeting,
FYI, we noticed the following commit (built with gcc-9):
commit: 6094661b1679bd1d59eb620371dfbf327e9feca7 ("[PATCH -tip V2 04/10] workqueue: don't set the worker's cpumask when kthread_bind_mask()")
url: https://github.com/0day-ci/linux/commits/Lai-Jiangshan/workqueue-break-af...
base: https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git a409ed156a90093a03fe6a93721ddf4c591eac87
in testcase: will-it-scale
version: will-it-scale-x86_64-b695a1b-1_20201217
with following parameters:
nr_task: 50%
mode: process
test: context_switch1
cpufreq_governor: performance
ucode: 0x2006a08
test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
test-url: https://github.com/antonblanchard/will-it-scale
on test machine: 104 threads Skylake with 192G memory
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang(a)intel.com>
kern :warn : [ 2.900184] WARNING: CPU: 26 PID: 140 at kernel/sched/core.c:2379 __set_cpus_allowed_ptr+0x1c4/0x1e0
kern :warn : [ 2.900184] Modules linked in:
kern :warn : [ 2.900184] CPU: 26 PID: 140 Comm: cpuhp/26 Not tainted 5.10.0-11829-g6094661b1679 #1
kern :warn : [ 2.900184] RIP: 0010:__set_cpus_allowed_ptr+0x1c4/0x1e0
kern :warn : [ 2.900184] Code: 74 99 8b 15 d2 d6 ad 01 48 c7 c6 a0 30 bf 82 4c 89 f7 e8 8f bd 48 00 85 c0 75 80 41 83 bc 24 c0 03 00 00 01 0f 84 71 ff ff ff <0f> 0b e9 6a ff ff ff e8 b0 5c ac 00 66 66 2e 0f 1f 84 00 00 00 00
kern :warn : [ 2.900184] RSP: 0000:ffffc9000cebfd00 EFLAGS: 00010006
kern :warn : [ 2.900184] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 00000000ffffff98
kern :warn : [ 2.900184] RDX: 00000000ffffff98 RSI: 0000000000000000 RDI: ffff889846721ed0
kern :warn : [ 2.900184] RBP: ffffc9000cebfd40 R08: 0000000000000001 R09: 0000000000000002
kern :warn : [ 2.900184] R10: ffff889846721ed0 R11: ffffc9010cebfcd7 R12: ffff889846730000
kern :warn : [ 2.900184] R13: ffff8897e042b040 R14: ffff889846721ed0 R15: 000000000000001a
kern :warn : [ 2.900184] FS: 0000000000000000(0000) GS:ffff88afbc600000(0000) knlGS:0000000000000000
kern :warn : [ 2.900184] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
kern :warn : [ 2.900184] CR2: 0000000000000000 CR3: 000000303e20a001 CR4: 00000000007706e0
kern :warn : [ 2.900184] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
kern :warn : [ 2.900184] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
kern :warn : [ 2.900184] PKRU: 55555554
kern :warn : [ 2.900184] Call Trace:
kern :warn : [ 2.900184] worker_attach_to_pool+0x33/0xa0
kern :warn : [ 2.900184] create_worker+0xff/0x1a0
kern :warn : [ 2.900184] alloc_unbound_pwq+0x444/0x460
kern :warn : [ 2.900184] wq_update_unbound_numa+0x1a3/0x1e0
kern :warn : [ 2.900184] workqueue_online_cpu+0x1eb/0x240
kern :warn : [ 2.900184] ? workqueue_prepare_cpu+0x80/0x80
kern :warn : [ 2.900184] cpuhp_invoke_callback+0x82/0x440
kern :warn : [ 2.900184] ? smpboot_thread_fn+0x26/0x1e0
kern :warn : [ 2.900184] cpuhp_thread_fun+0xa4/0x100
kern :warn : [ 2.900184] smpboot_thread_fn+0x10b/0x1e0
kern :warn : [ 2.900184] ? sort_range+0x20/0x20
kern :warn : [ 2.900184] kthread+0x116/0x160
kern :warn : [ 2.900184] ? kthread_park+0xa0/0xa0
kern :warn : [ 2.900184] ret_from_fork+0x22/0x30
kern :warn : [ 2.900184] ---[ end trace ee287c8df3029920 ]---
To reproduce:
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Thanks,
Oliver Sang
1 year, 5 months
[kernfs] ca0f27ecb7: BUG:kernel_NULL_pointer_dereference,address
by kernel test robot
Greeting,
FYI, we noticed the following commit (built with gcc-9):
commit: ca0f27ecb7ab7278183680b79f3de35416372fa3 ("[PATCH 6/6] kernfs: add a spinlock to kernfs iattrs for inode updates")
url: https://github.com/0day-ci/linux/commits/Ian-Kent/kernfs-proposed-locking...
base: https://git.kernel.org/cgit/linux/kernel/git/gregkh/driver-core.git accefff5b547a9a1d959c7e76ad539bf2480e78b
in testcase: locktorture
version:
with following parameters:
runtime: 300s
test: default
test-description: This torture test consists of creating a number of kernel threads which acquire the lock and hold it for specific amount of time, thus simulating different critical region behaviors.
test-url: https://www.kernel.org/doc/Documentation/locking/locktorture.txt
on test machine: qemu-system-x86_64 -enable-kvm -cpu SandyBridge -smp 2 -m 8G
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
+---------------------------------------------+------------+------------+
| | 51383f001c | ca0f27ecb7 |
+---------------------------------------------+------------+------------+
| boot_successes | 10 | 0 |
| boot_failures | 0 | 10 |
| BUG:kernel_NULL_pointer_dereference,address | 0 | 10 |
| Oops:#[##] | 0 | 10 |
| RIP:_raw_spin_lock | 0 | 10 |
| Kernel_panic-not_syncing:Fatal_exception | 0 | 10 |
+---------------------------------------------+------------+------------+
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang(a)intel.com>
[ 10.631592] BUG: kernel NULL pointer dereference, address: 0000000000000058
[ 10.633698] #PF: supervisor write access in kernel mode
[ 10.635218] #PF: error_code(0x0002) - not-present page
[ 10.636627] PGD 0 P4D 0
[ 10.637454] Oops: 0002 [#1] SMP PTI
[ 10.638451] CPU: 1 PID: 1 Comm: systemd Not tainted 5.10.0-10901-gca0f27ecb7ab #1
[ 10.640384] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.12.0-1 04/01/2014
[ 10.642512] RIP: 0010:_raw_spin_lock+0x13/0x40
[ 10.643746] Code: 33 b7 a2 58 31 c0 ba ff 00 00 00 f0 0f b1 17 75 01 c3 e9 e0 8d 55 ff 66 66 66 66 90 65 ff 05 14 b7 a2 58 31 c0 ba 01 00 00 00 <f0> 0f b1 17 75 01 c3 89 c6 e8 ff 73 55 ff 66 90 c3 66 66 2e 0f 1f
[ 10.648414] RSP: 0018:ffff967740013da0 EFLAGS: 00010246
[ 10.649743] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000000
[ 10.651555] RDX: 0000000000000001 RSI: ffff8a0581d30708 RDI: 0000000000000058
[ 10.653341] RBP: ffff967740013dc8 R08: ffff8a0540232d48 R09: ffff8a0540232d48
[ 10.655161] R10: ffff8a0581d30000 R11: 0000000000000005 R12: ffff8a0581d30708
[ 10.656903] R13: ffff8a05401d5f80 R14: 0000000000000058 R15: ffff8a058739a9c0
[ 10.658737] FS: 00007f88eae95940(0000) GS:ffff8a067fd00000(0000) knlGS:0000000000000000
[ 10.660820] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 10.662400] CR2: 0000000000000058 CR3: 000000010005e000 CR4: 00000000000406e0
[ 10.664204] Call Trace:
[ 10.665010] kernfs_refresh_inode+0x26/0xc0
[ 10.667380] kernfs_get_inode+0xa0/0x120
[ 10.668556] kernfs_get_tree+0x12b/0x240
[ 10.669670] sysfs_get_tree+0x13/0x40
[ 10.670694] vfs_get_tree+0x25/0xc0
[ 10.671689] path_mount+0x43f/0x9c0
[ 10.672680] do_mount+0x75/0xa0
[ 10.673635] __x64_sys_mount+0x8e/0xe0
[ 10.674686] do_syscall_64+0x33/0x40
[ 10.675718] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[ 10.677137] RIP: 0033:0x7f88ec055fea
[ 10.678181] Code: 48 8b 0d a9 0e 0c 00 f7 d8 64 89 01 48 83 c8 ff c3 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 49 89 ca b8 a5 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 76 0e 0c 00 f7 d8 64 89 01 48
[ 10.682538] RSP: 002b:00007ffcafcda3d8 EFLAGS: 00000246 ORIG_RAX: 00000000000000a5
[ 10.684486] RAX: ffffffffffffffda RBX: 000055e4f51178c0 RCX: 00007f88ec055fea
[ 10.686253] RDX: 000055e4f50973a9 RSI: 000055e4f50941bf RDI: 000055e4f50973a9
[ 10.688136] RBP: 0000000000000003 R08: 0000000000000000 R09: 000055e4f5d2f4e0
[ 10.689919] R10: 000000000000000e R11: 0000000000000246 R12: 0000000000000000
[ 10.691647] R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000001
[ 10.693431] Modules linked in:
[ 10.694351] CR2: 0000000000000058
[ 10.695309] ---[ end trace f3fef33fa8d30916 ]---
To reproduce:
# build kernel
cd linux
cp config-5.10.0-10901-gca0f27ecb7ab .config
make HOSTCC=gcc-9 CC=gcc-9 ARCH=x86_64 olddefconfig prepare modules_prepare bzImage modules
make HOSTCC=gcc-9 CC=gcc-9 ARCH=x86_64 INSTALL_MOD_PATH=<mod-install-dir> modules_install
cd <mod-install-dir>
find lib/ | cpio -o -H newc --quiet | gzip > modules.cgz
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> -m modules.cgz job-script # job-script is attached in this email
Thanks,
Oliver Sang
1 year, 5 months
[rcu] ddc403b8de: WARNING:at_kernel/rcu/tasks.h:#rcu_tasks_verify_self_tests
by kernel test robot
Greeting,
FYI, we noticed the following commit (built with gcc-9):
commit: ddc403b8dedd2df9a95c247b73a534e12e8e40a2 ("rcu-tasks: Add RCU-tasks self tests")
https://git.kernel.org/cgit/linux/kernel/git/paulmck/linux-rcu.git dev.2020.12.09a
in testcase: boot
on test machine: qemu-system-i386 -enable-kvm -cpu SandyBridge -smp 2 -m 8G
caused below changes (please refer to attached dmesg/kmsg for entire log/backtrace):
+------------------------------------------------------------+------------+------------+
| | 936a23496c | ddc403b8de |
+------------------------------------------------------------+------------+------------+
| boot_successes | 32 | 0 |
| boot_failures | 0 | 32 |
| WARNING:at_kernel/rcu/tasks.h:#rcu_tasks_verify_self_tests | 0 | 32 |
| EIP:rcu_tasks_verify_self_tests | 0 | 32 |
+------------------------------------------------------------+------------+------------+
If you fix the issue, kindly add following tag
Reported-by: kernel test robot <oliver.sang(a)intel.com>
[ 10.025923] WARNING: CPU: 1 PID: 1 at kernel/rcu/tasks.h:1278 rcu_tasks_verify_self_tests+0x18/0x20
[ 10.027560] Modules linked in:
[ 10.028138] CPU: 1 PID: 1 Comm: swapper/0 Not tainted 5.10.0-rc1-00218-gddc403b8dedd #1
[ 10.029493] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.12.0-1 04/01/2014
[ 10.030853] EIP: rcu_tasks_verify_self_tests+0x18/0x20
[ 10.031751] Code: 00 00 c3 8d b4 26 00 00 00 00 8d b4 26 00 00 00 00 90 8b 15 fc 87 5c 4d 31 c0 39 15 f8 87 5c 4d 75 08 c3 8d b4 26 00 00 00 00 <0f> 0b b8 ff ff ff ff c3 55 89 e5 83 ec 30 89 5d f4 89 75 f8 89 c6
[ 10.035008] EAX: 00000000 EBX: 00000000 ECX: 00000000 EDX: 00000002
[ 10.036193] ESI: 4aeee4b0 EDI: 413ea568 EBP: 41231f80 ESP: 41231f20
[ 10.037300] DS: 007b ES: 007b FS: 00d8 GS: 00e0 SS: 0068 EFLAGS: 00010202
[ 10.038442] CR0: 80050033 CR2: 37f32080 CR3: 0cee5000 CR4: 00040690
[ 10.039521] Call Trace:
[ 10.039984] ? do_one_initcall+0x65/0x330
[ 10.040656] ? rcu_read_lock_sched_held+0x47/0x80
[ 10.041483] ? trace_initcall_level+0x7c/0xa4
[ 10.042238] ? kernel_init_freeable+0x1ff/0x249
[ 10.043092] ? rest_init+0x23e/0x23e
[ 10.043697] ? kernel_init+0x8/0xe2
[ 10.044267] ? ret_from_fork+0x1c/0x28
[ 10.044796] CPU: 1 PID: 1 Comm: swapper/0 Not tainted 5.10.0-rc1-00218-gddc403b8dedd #1
[ 10.045876] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.12.0-1 04/01/2014
[ 10.047019] Call Trace:
[ 10.047385] ? dump_stack+0x6d/0x8b
[ 10.047883] ? __warn.cold+0x24/0x4d
[ 10.048390] ? rcu_tasks_verify_self_tests+0x18/0x20
[ 10.049227] ? rcu_tasks_verify_self_tests+0x18/0x20
[ 10.050045] ? rcu_tasks_verify_self_tests+0x18/0x20
[ 10.050888] ? report_bug+0xa3/0xe0
[ 10.051501] ? rcu_tasks_be_rude+0x10/0x10
[ 10.052200] ? exc_overflow+0x50/0x50
[ 10.052867] ? handle_bug+0x28/0x50
[ 10.053481] ? exc_invalid_op+0x26/0x80
[ 10.054157] ? handle_exception+0x12b/0x12b
[ 10.054896] ? rcu_tasks_be_rude+0x10/0x10
[ 10.055617] ? serpent_dec_blk_4way+0x1753/0x2518
[ 10.056462] ? validate_chain+0x7fb/0xf00
[ 10.061265] ? rcu_tasks_verify_self_tests+0x18/0x20
[ 10.062151] ? exc_overflow+0x50/0x50
[ 10.062795] ? rcu_tasks_verify_self_tests+0x18/0x20
[ 10.063691] ? do_one_initcall+0x65/0x330
[ 10.064410] ? rcu_read_lock_sched_held+0x47/0x80
[ 10.065264] ? trace_initcall_level+0x7c/0xa4
[ 10.066046] ? kernel_init_freeable+0x1ff/0x249
[ 10.066827] ? rest_init+0x23e/0x23e
[ 10.067450] ? kernel_init+0x8/0xe2
[ 10.068046] ? ret_from_fork+0x1c/0x28
[ 10.068843] irq event stamp: 2154779
[ 10.069513] hardirqs last enabled at (2154789): [<4aedbe95>] console_unlock+0x435/0x5e0
[ 10.070970] hardirqs last disabled at (2154798): [<4aedbe85>] console_unlock+0x425/0x5e0
[ 10.072468] softirqs last enabled at (2154778): [<4c014c4f>] __do_softirq+0x29f/0x455
[ 10.073952] softirqs last disabled at (2154767): [<4ae250f7>] call_on_stack+0x47/0x60
[ 10.075334] ---[ end trace db299d2e6ab9f182 ]---
To reproduce:
# build kernel
cd linux
cp config-5.10.0-rc1-00218-gddc403b8dedd .config
make HOSTCC=gcc-9 CC=gcc-9 ARCH=i386 olddefconfig prepare modules_prepare bzImage
git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> job-script # job-script is attached in this email
Thanks,
Oliver Sang
1 year, 5 months