Re: [LKP] [rcu] kernel BUG at include/linux/pagemap.h:149!
by Frederic Weisbecker
On Fri, Sep 11, 2015 at 10:19:47AM +0800, Boqun Feng wrote:
> Subject: [PATCH 01/27] rcu: Don't disable preemption for Tiny and Tree RCU
> readers
>
> Because preempt_disable() maps to barrier() for non-debug builds,
> it forces the compiler to spill and reload registers. Because Tree
> RCU and Tiny RCU now only appear in CONFIG_PREEMPT=n builds, these
> barrier() instances generate needless extra code for each instance of
> rcu_read_lock() and rcu_read_unlock(). This extra code slows down Tree
> RCU and bloats Tiny RCU.
>
> This commit therefore removes the preempt_disable() and preempt_enable()
> from the non-preemptible implementations of __rcu_read_lock() and
> __rcu_read_unlock(), respectively.
>
> For debug purposes, preempt_disable() and preempt_enable() are still
> kept if CONFIG_PREEMPT_COUNT=y, which makes the detection of sleeping
> inside atomic sections still work in non-preemptible kernels.
>
> Signed-off-by: Boqun Feng <boqun.feng(a)gmail.com>
> Signed-off-by: Paul E. McKenney <paulmck(a)linux.vnet.ibm.com>
> ---
> include/linux/rcupdate.h | 6 ++++--
> include/linux/rcutiny.h | 1 +
> kernel/rcu/tree.c | 9 +++++++++
> 3 files changed, 14 insertions(+), 2 deletions(-)
>
> diff --git a/include/linux/rcupdate.h b/include/linux/rcupdate.h
> index d63bb77..6c3cece 100644
> --- a/include/linux/rcupdate.h
> +++ b/include/linux/rcupdate.h
> @@ -297,12 +297,14 @@ void synchronize_rcu(void);
>
> static inline void __rcu_read_lock(void)
> {
> - preempt_disable();
> + if (IS_ENABLED(CONFIG_PREEMPT_COUNT))
> + preempt_disable();
preempt_disable() is a no-op when !CONFIG_PREEMPT_COUNT, right?
Or rather it's a barrier(), which is anyway implied by rcu_read_lock().
So perhaps we can get rid of the IS_ENABLED() check?
1 year, 7 months
[lkp] [x86/acpi] dc6db24d24: BUG: unable to handle kernel paging request at 0000116007090008
by kernel test robot
FYI, we noticed the following commit:
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit dc6db24d2476cd09c0ecf2b8d80313539f737a89 ("x86/acpi: Set persistent cpuid <-> nodeid mapping when booting")
in testcase: vm-scalability
with following parameters:
runtime: 300
thp_enabled: never
thp_defrag: never
nr_task: 1
nr_pmem: 1
test: swap-w-rand
cpufreq_governor: performance
The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
on test machine: 72 threads Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz with 128G memory
caused below changes:
+------------------------------------------------------------------+------------+------------+
| | 8ad893faf2 | dc6db24d24 |
+------------------------------------------------------------------+------------+------------+
| boot_successes | 7 | 0 |
| boot_failures | 9 | 16 |
| invoked_oom-killer:gfp_mask=0x | 6 | 2 |
| Mem-Info | 6 | 2 |
| Out_of_memory:Kill_process | 6 | |
| page_allocation_failure:order:#,mode:#(GFP_KERNEL|__GFP_NORETRY) | 2 | |
| warn_alloc_failed+0x | 2 | |
| BUG:kernel_hang_in_test_stage | 2 | 2 |
| BUG:kernel_reboot-without-warning_in_test_stage | 1 | |
| BUG:unable_to_handle_kernel | 0 | 12 |
| Oops | 0 | 12 |
| RIP:get_partial_node | 0 | 12 |
| calltrace:devtmpfsd | 0 | 12 |
| RIP:_raw_spin_lock_irqsave | 0 | 9 |
| general_protection_fault:#[##]SMP | 0 | 3 |
| RIP:native_queued_spin_lock_slowpath | 0 | 3 |
| Kernel_panic-not_syncing:Hard_LOCKUP | 0 | 3 |
| RIP:load_balance | 0 | 2 |
| Kernel_panic-not_syncing:Fatal_exception_in_interrupt | 0 | 2 |
| WARNING:at_lib/list_debug.c:#__list_add | 0 | 1 |
| calltrace:_do_fork | 0 | 1 |
| RIP:resched_curr | 0 | 1 |
| Kernel_panic-not_syncing:Fatal_exception | 0 | 1 |
| WARNING:at_include/linux/uaccess.h:#__probe_kernel_read | 0 | 5 |
| Kernel_panic-not_syncing:Out_of_memory_and_no_killable_processes | 0 | 2 |
+------------------------------------------------------------------+------------+------------+
[ 9.531507] pci 0000:80:02.2: bridge window [mem 0x387fffd00000-0x387fffefffff 64bit pref]
[ 9.541378] pci_bus 0000:80: on NUMA node 2
[ 9.546734] ACPI: Enabled 4 GPEs in block 00 to 3F
[ 9.586911] BUG: unable to handle kernel paging request at 0000116007090008
[ 9.595109] IP: [<ffffffff811e50fc>] get_partial_node+0x2c/0x1c0
[ 9.602933] PGD 0
[ 9.605503] Oops: 0000 [#1] SMP
[ 9.609264] Modules linked in:
[ 9.613005] CPU: 24 PID: 585 Comm: kdevtmpfs Not tainted 4.8.0-rc1-00300-gdc6db24d #1
[ 9.622193] Hardware name: Intel Corporation S2600WTT/S2600WTT, BIOS SE5C610.86B.01.01.0008.021120151325 02/11/2015
[ 9.634299] task: ffff880068040000 task.stack: ffff880068024000
[ 9.641168] RIP: 0010:[<ffffffff811e50fc>] [<ffffffff811e50fc>] get_partial_node+0x2c/0x1c0
[ 9.651890] RSP: 0000:ffff8800680279f0 EFLAGS: 00010006
[ 9.658079] RAX: 0000000000000002 RBX: 0000000000000246 RCX: 0000000002098020
[ 9.666308] RDX: ffff882053b9cfc0 RSI: 0000116007090000 RDI: ffff880076804dc0
[ 9.674535] RBP: ffff880068027a90 R08: ffff882053b9cfb0 R09: 0000000000000000
[ 9.682764] R10: ffff880068027c88 R11: 0000000b00000000 R12: ffff880076804dc0
[ 9.690994] R13: 0000000000000000 R14: ffff880076804dc0 R15: ffff882053b9cfb0
[ 9.699224] FS: 0000000000000000(0000) GS:ffff882053b80000(0000) knlGS:0000000000000000
[ 9.708701] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 9.715373] CR2: 0000116007090008 CR3: 0000000001e06000 CR4: 00000000001406e0
[ 9.723602] Stack:
[ 9.726094] ffff88207ffd4080 0000000200000000 0000000000000000 0000000002281220
[ 9.735086] 0000000000000000 0000000000000000 ffffffff82343f68 ffff880068040000
[ 9.744080] ffff880068027a88 ffffffff811d9de5 ffff880068040000 ffffffff82343f70
[ 9.753072] Call Trace:
[ 9.756056] [<ffffffff811d9de5>] ? alloc_pages_current+0x95/0x140
[ 9.763223] [<ffffffff811e551a>] ___slab_alloc+0x28a/0x4b0
[ 9.769696] [<ffffffff813dd477>] ? avc_alloc_node+0x27/0x140
[ 9.776379] [<ffffffff813e2356>] ? selinux_inode_permission+0xc6/0x180
[ 9.784032] [<ffffffff811e4342>] ? new_slab+0x2d2/0x5a0
[ 9.790208] [<ffffffff813dd477>] ? avc_alloc_node+0x27/0x140
[ 9.796881] [<ffffffff811e5760>] __slab_alloc+0x20/0x40
[ 9.803067] [<ffffffff811e6b7f>] kmem_cache_alloc+0x17f/0x1c0
[ 9.809837] [<ffffffff813dd477>] avc_alloc_node+0x27/0x140
[ 9.816317] [<ffffffff813dd87a>] avc_compute_av+0x8a/0x1e0
[ 9.822801] [<ffffffff8121000a>] ? sget_userns+0x4ca/0x4e0
[ 9.829289] [<ffffffff813de596>] avc_has_perm+0x136/0x190
[ 9.835673] [<ffffffff810a4a69>] ? __might_sleep+0x49/0x80
[ 9.842161] [<ffffffff813e0000>] ? inode_doinit_with_dentry+0x530/0x660
[ 9.849901] [<ffffffff813f4c5d>] ? security_transition_sid+0x2d/0x40
[ 9.857351] [<ffffffff813e1379>] may_create+0xb9/0xe0
[ 9.863334] [<ffffffff813e13e2>] selinux_inode_mknod+0x42/0x80
[ 9.870201] [<ffffffff813da552>] security_inode_mknod+0x52/0x80
[ 9.877165] [<ffffffff812197e1>] vfs_mknod+0x131/0x1e0
[ 9.883255] [<ffffffff815b2e65>] handle_create+0x75/0x1e0
[ 9.889639] [<ffffffff8192da66>] ? __schedule+0x2e6/0x790
[ 9.896027] [<ffffffff815b3104>] devtmpfsd+0x134/0x180
[ 9.902117] [<ffffffff815b2fd0>] ? handle_create+0x1e0/0x1e0
[ 9.908792] [<ffffffff8109ded4>] kthread+0xd4/0xf0
[ 9.914503] [<ffffffff81932cbf>] ret_from_fork+0x1f/0x40
[ 9.920788] [<ffffffff8109de00>] ? kthread_create_on_node+0x180/0x180
[ 9.928335] Code: 1f 44 00 00 55 48 89 e5 41 57 41 56 41 55 41 54 53 48 83 e4 f0 48 83 ec 70 48 85 f6 48 c7 44 24 20 00 00 00 00 0f 84 54 01 00 00 <48> 83 7e 08 00 0f 84 49 01 00 00 48 89 f3 49 89 fd 48 89 f7 89
[ 9.954843] RIP [<ffffffff811e50fc>] get_partial_node+0x2c/0x1c0
[ 9.962756] RSP <ffff8800680279f0>
[ 9.966902] CR2: 0000116007090008
[ 9.970871] BUG: unable to handle kernel paging request at 0000000100000048
[ 9.979058] IP: [<ffffffff819329b9>] _raw_spin_lock_irqsave+0x29/0x50
[ 9.986582] PGD 0
[ 9.989147] Oops: 0002 [#2] SMP
[ 9.992891] Modules linked in:
[ 9.996623] CPU: 24 PID: 585 Comm: kdevtmpfs Tainted: G D 4.8.0-rc1-00300-gdc6db24d #1
[ 10.007173] Hardware name: Intel Corporation S2600WTT/S2600WTT, BIOS SE5C610.86B.01.01.0008.021120151325 02/11/2015
[ 10.019279] task: ffff880068040000 task.stack: ffff880068024000
[ 10.026147] RIP: 0010:[<ffffffff819329b9>] [<ffffffff819329b9>] _raw_spin_lock_irqsave+0x29/0x50
[ 10.036577] RSP: 0000:ffff8800680276e0 EFLAGS: 00010046
[ 10.042763] RAX: 0000000000000000 RBX: 0000000000000097 RCX: ffffffff81e5af08
[ 10.050991] RDX: 0000000000000001 RSI: ffff880068027738 RDI: 0000000100000048
[ 10.059221] RBP: ffff8800680276e8 R08: 0000000000000001 R09: 0000000000000001
[ 10.067450] R10: ffff880068027c88 R11: 000000000000048c R12: 0000000100000048
[ 10.075677] R13: 0000000000000008 R14: ffff880068027738 R15: 0000000000000046
[ 10.083906] FS: 0000000000000000(0000) GS:ffff882053b80000(0000) knlGS:0000000000000000
[ 10.093384] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 10.100059] CR2: 0000000100000048 CR3: 0000000001e06000 CR4: 00000000001406e0
[ 10.108288] Stack:
[ 10.110780] 0000000100000000 ffff880068027718 ffffffff81575da0 ffffffff82263b00
[ 10.119773] ffff880068027738 0000000000000008 ffffffff8107e58f ffff880068027728
[ 10.128764] ffffffff81575e4f ffff880068027798 ffffffff8157726f ffff880068027790
[ 10.137756] Call Trace:
[ 10.140741] [<ffffffff81575da0>] _extract_crng+0x40/0xb0
[ 10.151150] [<ffffffff8107e58f>] ? print_oops_end_marker+0x3f/0x60
[ 10.158405] [<ffffffff81575e4f>] extract_crng+0x3f/0x50
[ 10.164591] [<ffffffff8157726f>] get_random_bytes+0x6f/0x1a0
[ 10.171268] [<ffffffff810d811a>] ? console_unlock+0x33a/0x610
[ 10.178048] [<ffffffff8107e58f>] print_oops_end_marker+0x3f/0x60
[ 10.185106] [<ffffffff8107e5cd>] oops_exit+0x1d/0x30
[ 10.191009] [<ffffffff8103091e>] oops_end+0x7e/0xd0
[ 10.196815] [<ffffffff81066592>] no_context+0x112/0x380
[ 10.203002] [<ffffffff81066881>] __bad_area_nosemaphore+0x81/0x1c0
[ 10.210257] [<ffffffff810669d4>] bad_area_nosemaphore+0x14/0x20
[ 10.217219] [<ffffffff81066d6c>] __do_page_fault+0xbc/0x4d0
[ 10.223796] [<ffffffff8146b47d>] ? list_del+0xd/0x30
[ 10.229690] [<ffffffff810671b0>] do_page_fault+0x30/0x80
[ 10.235972] [<ffffffff81933f48>] page_fault+0x28/0x30
[ 10.241965] [<ffffffff811e50fc>] ? get_partial_node+0x2c/0x1c0
[ 10.249610] [<ffffffff811d9de5>] ? alloc_pages_current+0x95/0x140
[ 10.256771] [<ffffffff811e551a>] ___slab_alloc+0x28a/0x4b0
[ 10.263249] [<ffffffff813dd477>] ? avc_alloc_node+0x27/0x140
[ 10.269921] [<ffffffff813e2356>] ? selinux_inode_permission+0xc6/0x180
[ 10.277564] [<ffffffff811e4342>] ? new_slab+0x2d2/0x5a0
[ 10.283749] [<ffffffff813dd477>] ? avc_alloc_node+0x27/0x140
[ 10.290421] [<ffffffff811e5760>] __slab_alloc+0x20/0x40
[ 10.296607] [<ffffffff811e6b7f>] kmem_cache_alloc+0x17f/0x1c0
[ 10.303379] [<ffffffff813dd477>] avc_alloc_node+0x27/0x140
[ 10.309848] [<ffffffff813dd87a>] avc_compute_av+0x8a/0x1e0
[ 10.316326] [<ffffffff8121000a>] ? sget_userns+0x4ca/0x4e0
[ 10.322806] [<ffffffff813de596>] avc_has_perm+0x136/0x190
[ 10.329184] [<ffffffff810a4a69>] ? __might_sleep+0x49/0x80
[ 10.335660] [<ffffffff813e0000>] ? inode_doinit_with_dentry+0x530/0x660
[ 10.343403] [<ffffffff813f4c5d>] ? security_transition_sid+0x2d/0x40
[ 10.350855] [<ffffffff813e1379>] may_create+0xb9/0xe0
[ 10.356849] [<ffffffff813e13e2>] selinux_inode_mknod+0x42/0x80
[ 10.363716] [<ffffffff813da552>] security_inode_mknod+0x52/0x80
[ 10.370680] [<ffffffff812197e1>] vfs_mknod+0x131/0x1e0
[ 10.376770] [<ffffffff815b2e65>] handle_create+0x75/0x1e0
[ 10.383151] [<ffffffff8192da66>] ? __schedule+0x2e6/0x790
[ 10.389533] [<ffffffff815b3104>] devtmpfsd+0x134/0x180
[ 10.395622] [<ffffffff815b2fd0>] ? handle_create+0x1e0/0x1e0
[ 10.402299] [<ffffffff8109ded4>] kthread+0xd4/0xf0
[ 10.408001] [<ffffffff81932cbf>] ret_from_fork+0x1f/0x40
[ 10.414284] [<ffffffff8109de00>] ? kthread_create_on_node+0x180/0x180
[ 10.421829] Code: 00 00 0f 1f 44 00 00 55 48 89 e5 53 9c 58 0f 1f 44 00 00 48 89 c3 fa 66 0f 1f 44 00 00 65 ff 05 9e a8 6d 7e 31 c0 ba 01 00 00 00 <f0> 0f b1 17 85 c0 75 06 48 89 d8 5b 5d c3 89 c6 e8 22 74 79 ff
[ 10.448339] RIP [<ffffffff819329b9>] _raw_spin_lock_irqsave+0x29/0x50
[ 10.455959] RSP <ffff8800680276e0>
[ 10.460101] CR2: 0000000100000048
[ 10.464058] BUG: unable to handle kernel paging request at 0000000100000048
[ 10.472244] IP: [<ffffffff819329b9>] _raw_spin_lock_irqsave+0x29/0x50
[ 10.479768] PGD 0
[ 10.482332] Oops: 0002 [#3] SMP
[ 10.486089] Modules linked in:
[ 10.489822] CPU: 24 PID: 585 Comm: kdevtmpfs Tainted: G D 4.8.0-rc1-00300-gdc6db24d #1
[ 10.500366] Hardware name: Intel Corporation S2600WTT/S2600WTT, BIOS SE5C610.86B.01.01.0008.021120151325 02/11/2015
[ 10.512467] task: ffff880068040000 task.stack: ffff880068024000
[ 10.519334] RIP: 0010:[<ffffffff819329b9>] [<ffffffff819329b9>] _raw_spin_lock_irqsave+0x29/0x50
[ 10.529765] RSP: 0000:ffff8800680273d0 EFLAGS: 00010046
[ 10.535952] RAX: 0000000000000000 RBX: 0000000000000097 RCX: ffffffff81e5af08
[ 10.544183] RDX: 0000000000000001 RSI: ffff880068027428 RDI: 0000000100000048
[ 10.552410] RBP: ffff8800680273d8 R08: 0000000000000001 R09: 0000000000000001
[ 10.560641] R10: ffff880068027c88 R11: 00000000000004d1 R12: 0000000100000048
[ 10.568869] R13: 0000000000000008 R14: ffff880068027428 R15: 0000000000000046
[ 10.577097] FS: 0000000000000000(0000) GS:ffff882053b80000(0000) knlGS:0000000000000000
[ 10.586578] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 10.593250] CR2: 0000000100000048 CR3: 0000000001e06000 CR4: 00000000001406e0
[ 10.601479] Stack:
[ 10.603969] 0000000100000000 ffff880068027408 ffffffff81575da0 ffffffff82263b00
[ 10.612968] ffff880068027428 0000000000000008 ffffffff8107e58f ffff880068027418
[ 10.621966] ffffffff81575e4f ffff880068027488 ffffffff8157726f ffff880068027480
[ 10.630963] Call Trace:
[ 10.633942] [<ffffffff81575da0>] _extract_crng+0x40/0xb0
[ 10.640228] [<ffffffff8107e58f>] ? print_oops_end_marker+0x3f/0x60
[ 10.647484] [<ffffffff81575e4f>] extract_crng+0x3f/0x50
[ 10.653670] [<ffffffff8157726f>] get_random_bytes+0x6f/0x1a0
[ 10.660342] [<ffffffff810d811a>] ? console_unlock+0x33a/0x610
[ 10.667113] [<ffffffff8107e58f>] print_oops_end_marker+0x3f/0x60
[ 10.674173] [<ffffffff8107e5cd>] oops_exit+0x1d/0x30
[ 10.680069] [<ffffffff8103091e>] oops_end+0x7e/0xd0
[ 10.685868] [<ffffffff81066592>] no_context+0x112/0x380
[ 10.692059] [<ffffffff81457b18>] ? put_dec+0x18/0xa0
[ 10.697962] [<ffffffff81066881>] __bad_area_nosemaphore+0x81/0x1c0
[ 10.705218] [<ffffffff810669d4>] bad_area_nosemaphore+0x14/0x20
[ 10.712183] [<ffffffff81066d6c>] __do_page_fault+0xbc/0x4d0
[ 10.718756] [<ffffffff810671b0>] do_page_fault+0x30/0x80
[ 10.725040] [<ffffffff8109f061>] ? atomic_notifier_call_chain+0x21/0x30
[ 10.732783] [<ffffffff81933f48>] page_fault+0x28/0x30
[ 10.738777] [<ffffffff819329b9>] ? _raw_spin_lock_irqsave+0x29/0x50
[ 10.746132] [<ffffffff81575da0>] _extract_crng+0x40/0xb0
[ 10.752415] [<ffffffff8107e58f>] ? print_oops_end_marker+0x3f/0x60
[ 10.759671] [<ffffffff81575e4f>] extract_crng+0x3f/0x50
[ 10.765856] [<ffffffff8157726f>] get_random_bytes+0x6f/0x1a0
[ 10.772530] [<ffffffff810d811a>] ? console_unlock+0x33a/0x610
[ 10.779301] [<ffffffff8107e58f>] print_oops_end_marker+0x3f/0x60
[ 10.786364] [<ffffffff8107e5cd>] oops_exit+0x1d/0x30
[ 10.792257] [<ffffffff8103091e>] oops_end+0x7e/0xd0
[ 10.798057] [<ffffffff81066592>] no_context+0x112/0x380
[ 10.804244] [<ffffffff81066881>] __bad_area_nosemaphore+0x81/0x1c0
[ 10.811498] [<ffffffff810669d4>] bad_area_nosemaphore+0x14/0x20
[ 10.818463] [<ffffffff81066d6c>] __do_page_fault+0xbc/0x4d0
[ 10.825037] [<ffffffff8146b47d>] ? list_del+0xd/0x30
[ 10.830933] [<ffffffff810671b0>] do_page_fault+0x30/0x80
[ 10.837216] [<ffffffff81933f48>] page_fault+0x28/0x30
[ 10.843208] [<ffffffff811e50fc>] ? get_partial_node+0x2c/0x1c0
[ 10.850855] [<ffffffff811d9de5>] ? alloc_pages_current+0x95/0x140
[ 10.858015] [<ffffffff811e551a>] ___slab_alloc+0x28a/0x4b0
[ 10.864491] [<ffffffff813dd477>] ? avc_alloc_node+0x27/0x140
[ 10.871163] [<ffffffff813e2356>] ? selinux_inode_permission+0xc6/0x180
[ 10.878809] [<ffffffff811e4342>] ? new_slab+0x2d2/0x5a0
[ 10.884995] [<ffffffff813dd477>] ? avc_alloc_node+0x27/0x140
[ 10.891667] [<ffffffff811e5760>] __slab_alloc+0x20/0x40
[ 10.897853] [<ffffffff811e6b7f>] kmem_cache_alloc+0x17f/0x1c0
[ 10.904623] [<ffffffff813dd477>] avc_alloc_node+0x27/0x140
[ 10.911103] [<ffffffff813dd87a>] avc_compute_av+0x8a/0x1e0
[ 10.917582] [<ffffffff8121000a>] ? sget_userns+0x4ca/0x4e0
[ 10.924061] [<ffffffff813de596>] avc_has_perm+0x136/0x190
[ 10.930443] [<ffffffff810a4a69>] ? __might_sleep+0x49/0x80
[ 10.936924] [<ffffffff813e0000>] ? inode_doinit_with_dentry+0x530/0x660
[ 10.944666] [<ffffffff813f4c5d>] ? security_transition_sid+0x2d/0x40
[ 10.952120] [<ffffffff813e1379>] may_create+0xb9/0xe0
[ 10.958112] [<ffffffff813e13e2>] selinux_inode_mknod+0x42/0x80
[ 10.964979] [<ffffffff813da552>] security_inode_mknod+0x52/0x80
[ 10.971944] [<ffffffff812197e1>] vfs_mknod+0x131/0x1e0
[ 10.978033] [<ffffffff815b2e65>] handle_create+0x75/0x1e0
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Thanks,
Xiaolong
3 years, 11 months
[lkp] [x86, paravirt] fd6f48529f: aim7.jobs-per-min -26.1% regression
by kernel test robot
Greeting,
FYI, we noticed a -26.1% regression of aim7.jobs-per-min due to commit:
commit: fd6f48529f4067c6bfbead88d3d7d117d51150f6 ("x86,paravirt: Optimize native pv_lock_ops.vcpu_is_preempted")
https://git.kernel.org/pub/scm/linux/kernel/git/peterz/queue.git locking/core
in testcase: aim7
on test machine: qemu-system-x86_64 -enable-kvm -cpu host -smp 4 -m 5G
with following parameters:
test: link_test
load: 2000
test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> job-script # job-script is attached in this email
testcase/path_params/tbox_group/run: aim7/link_test-2000/vm-lkp-a05
8ba24710fb279053 fd6f48529f4067c6bfbead88d3
---------------- --------------------------
%stddev change %stddev
\ | \
9147 ± 3% -26% 6763 ± 3% aim7.jobs-per-min
560854 59% 891815 ± 9% aim7.time.voluntary_context_switches
1329 ± 3% 35% 1792 ± 3% aim7.time.elapsed_time
1329 ± 3% 35% 1792 ± 3% aim7.time.elapsed_time.max
2594 ± 4% 29% 3349 ± 3% aim7.time.user_time
1625 12% 1821 ± 4% aim7.time.system_time
3189 -6% 2988 vmstat.system.cs
5534 -6% 5183 vmstat.system.in
12154 ±118% -1e+04 1682 ± 40% latency_stats.avg.call_rwsem_down_write_failed.copy_process._do_fork.SyS_clone.do_syscall_64.return_from_SYSCALL_64
14029 ± 12% -1e+04 390 ± 59% latency_stats.avg.pipe_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath
17531 ± 84% -2e+04 2189 ± 77% latency_stats.avg.call_rwsem_down_write_failed.__vma_adjust.__split_vma.split_vma.mprotect_fixup.do_mprotect_pkey.SyS_mprotect.entry_SYSCALL_64_fastpath
20760 ± 84% -2e+04 2731 ± 89% latency_stats.avg.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.SyS_execve.do_syscall_64.return_from_SYSCALL_64
143117 ±123% -1e+05 11061 ±173% latency_stats.avg.nfs_wait_on_request.nfs_writepage_setup.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
92032 ± 22% 2e+05 258628 ±138% latency_stats.max.pipe_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
8848 ± 91% -8e+03 931 ± 84% latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_execute.rpc_run_task.nfs4_call_sync_sequence.[nfsv4]._nfs4_proc_getattr.[nfsv4].nfs4_proc_getattr.[nfsv4].__nfs_revalidate_inode.nfs_do_access.nfs_permission.__inode_permission.inode_permission
29448 ±156% -3e+04 379 ±113% latency_stats.max.wb_wait_for_completion.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
145833 ±120% -1e+05 14303 ±173% latency_stats.max.nfs_wait_on_request.nfs_writepage_setup.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
581560 ±137% -5e+05 80796 ± 25% latency_stats.max.pipe_wait.pipe_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath
317533 ± 29% 2e+06 2166387 ± 17% latency_stats.sum.pipe_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
30223 ±152% -3e+04 610 ±107% latency_stats.sum.wb_wait_for_completion.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
162947 ±103% -1e+05 22122 ±173% latency_stats.sum.nfs_wait_on_request.nfs_writepage_setup.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
aim7.jobs-per-min
10000 ++------------------------------------------------------------------+
| *.** *.** *.** * *** * * |
9500 ** ** ** : ** : ** : * :*. : * + * : : *. |
9000 ++ ** :*. : * * **.* : * * * ** : * ** * *
| :+ * ** :** : * * *|
8500 ++* * :: |
| * |
8000 ++ |
| |
7500 ++ |
7000 +O O O O O |
| O O OO O OO OO O O OO O O |
6500 O+ OO OO O O O OOO |
| O O |
6000 ++------------------------------------------------------------------+
aim7.time.elapsed_time
1900 ++----------------------------O----O---------------------------------+
| OO O OO O |
1800 O+O O OOO OO OOO OO OO |
| O O O O O |
1700 +O O O O O |
| |
1600 ++ |
| |
1500 ++ |
| * |
1400 ++* *. *. * :: *|
| :+ * : *** * .* : *: * * *. ** * **** :*
1300 ** * ** ** + ** : *.**** : *. :* : * : * :+ *.* |
| *** *** *** * *.** * * |
1200 ++-------------------------------------------------------------------+
aim7.time.elapsed_time.max
1900 ++----------------------------O----O---------------------------------+
| OO O OO O |
1800 O+O O OOO OO OOO OO OO |
| O O O O O |
1700 +O O O O O |
| |
1600 ++ |
| |
1500 ++ |
| * |
1400 ++* *. *. * :: *|
| :+ * : *** * .* : *: * * *. ** * **** :*
1300 ** * ** ** + ** : *.**** : *. :* : * : * :+ *.* |
| *** *** *** * *.** * * |
1200 ++-------------------------------------------------------------------+
aim7.time.voluntary_context_switches
1e+06 ++---------------------------O----O--------------------------------+
950000 ++ O O O O O O OO O O |
O OO OO O O O O O |
900000 ++ |
850000 ++ O O |
| O O O O O |
800000 +O O O O |
750000 ++ |
700000 ++ |
| |
650000 ++ |
600000 ++ |
* * * ** ** * *.* * ***. * .****
550000 +***. :* **.*** **.** **.* *** * * * .** :* .* ** ** *** |
500000 ++---*------------------------------*-*-*----*-*--*----------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
4 years, 1 month
[lkp] [thermal/powerclamp] ec638db8cb: pm-qa.thermal_02.fail
by kernel test robot
Greeting,
FYI, we noticed that pm-qa.thermal_02.fail due to commit:
commit: ec638db8cb9ddd5ca08b23f2835b6c9c15eb616d ("thermal/powerclamp: add back module device table")
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
in testcase: pm-qa
on test machine: 12 threads Intel(R) Core(TM) i7 CPU X 980 @ 3.33GHz with 6G memory
with following parameters:
test: thermal
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
testcase/path_params/tbox_group/run: pm-qa/thermal/wsm
b58ec8b582471787 ec638db8cb9ddd5ca08b23f283
---------------- --------------------------
fail:runs %reproduction fail:runs
| | |
:4 100% 4:4 pm-qa.thermal_02.fail
4:4 -100% :4 pm-qa.thermal_02.pass
%stddev change %stddev
\ | \
1546 ± 23% 70% 2629 ± 29% pm-qa.time.involuntary_context_switches
33480 3% 34510 pm-qa.time.minor_page_faults
1:4 -25% :4 kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#]
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
4 years, 1 month
[lkp] [sched/autogroup] 8e5bfa8c1f: [No primary change] perf-stat.cpu-migrations +102% increase
by kernel test robot
Greeting,
There is no primary kpi change in this test, below is the data collected through multiple monitors running background just for your information.
commie: 8e5bfa8c1f8471aa4a2d30be631ef2b50e10abaf ("sched/autogroup: Do not use autogroup->tg in zombie threads")
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
in testcase: unixbench
on test machine: 8 threads Intel(R) Core(TM) i7 CPU 870 @ 2.93GHz with 6G memory
with following parameters:
runtime: 300s
nr_task: 100%
test: spawn
test-description: UnixBench is the original BYTE UNIX benchmark suite aims to test performance of Unix-like system.
test-url: https://github.com/kdlucas/byte-unixbench
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
testcase/path_params/tbox_group/run: unixbench/300s-100%-spawn/nhm-white
18f649ef344127ef 8e5bfa8c1f8471aa4a2d30be63
---------------- --------------------------
%stddev change %stddev
\ | \
3494 3457 unixbench.score
4743931 ± 4% -44% 2638522 ± 3% unixbench.time.involuntary_context_switches
26162810 25837723 unixbench.time.voluntary_context_switches
2.915e+08 2.88e+08 unixbench.time.minor_page_faults
1110 -4% 1064 unixbench.time.user_time
583 -5% 552 unixbench.time.percent_of_cpu_this_job_got
820 -7% 763 unixbench.time.system_time
42920 15% 49508 interrupts.CAL:Function_call_interrupts
154558 -8% 141951 vmstat.system.cs
28865 ± 3% -26% 21339 vmstat.system.in
2169 -4% 2083 turbostat.Avg_MHz
79.61 -5% 75.90 turbostat.%Busy
209136 ±123% -2e+05 51277 ±168% latency_stats.max.wait_on_page_bit.__filemap_fdatawait_range.filemap_fdatawait_keep_errors.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
218916 ±114% -2e+05 60379 ±134% latency_stats.max.max
213732 ±121% -2e+05 55750 ±155% latency_stats.sum.wait_on_page_bit.__filemap_fdatawait_range.filemap_fdatawait_keep_errors.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.entry_SYSCALL_64_fastpath
51492031 -8% 47290752 perf-stat.context-switches
3.109e+12 -4% 2.997e+12 perf-stat.iTLB-loads
3.02 8% 3.26 perf-stat.branch-miss-rate%
4134105 102% 8356281 perf-stat.cpu-migrations
2305 2256 perf-stat.instructions-per-iTLB-miss
0.04 0.04 perf-stat.iTLB-load-miss-rate%
1.749e+10 4% 1.826e+10 perf-stat.branch-misses
1.322e+12 1.286e+12 perf-stat.dTLB-loads
5.802e+11 -3% 5.607e+11 perf-stat.branch-instructions
3.108e+12 -4% 2.996e+12 perf-stat.instructions
7.711e+11 -4% 7.414e+11 perf-stat.dTLB-stores
2.919e+08 2.883e+08 perf-stat.page-faults
2.919e+08 2.883e+08 perf-stat.minor-faults
5.687e+12 -4% 5.464e+12 perf-stat.cpu-cycles
1.348e+09 1.328e+09 perf-stat.iTLB-load-misses
6.29 -22% 4.92 perf-stat.cache-miss-rate%
5.619e+09 -22% 4.384e+09 perf-stat.cache-misses
7.872e+08 -17% 6.499e+08 perf-stat.node-loads
4.755e+09 -22% 3.686e+09 perf-stat.node-stores
perf-stat.cpu-cycles
6e+12 ++------------------------------------------------------------------+
O OO.O.O.OO.O.O.OO.O O OO.O.O.OO.O.O.* *.*.**.*.*.**.*.*.**.*.*.**.*
5e+12 ++: : : : : |
| : : : : : |
| : : : : : |
4e+12 ++ : : : : |
|: : : : : |
3e+12 ++ : : : : |
|: : : :: |
2e+12 ++ : : :: |
|: : : :: |
| : : |
1e+12 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.instructions
3.5e+12 ++----------------------------------------------------------------+
| * *. *. |
3e+12 O+OO.O.OO.O.O.OO.O.OO O OO.O.OO.O.O.* : **.*.**.*.**.*.*.**.*.**.*
| : : : : : |
2.5e+12 ++: : : : : |
|: : : : : |
2e+12 ++ : : : : |
|: : : : : |
1.5e+12 ++ :: :: |
|: :: :: |
1e+12 ++ :: :: |
| : : |
5e+11 ++ : : |
| : : |
0 *+------------------*----------------*----------------------------+
perf-stat.cache-misses
6e+09 ++------------------------------------------------------------------+
| **.*.*.**.*.*.**.* **.*.*.**.*.*.* *.*.**.*.*.**.*.*.**.*.*.**.*
5e+09 ++: : : : : |
| : : : : : |
O OO O O OO O O OO O O OO O O OO O O : : |
4e+09 ++ : : : : |
|: : : : : |
3e+09 ++ : : : : |
|: : : :: |
2e+09 ++ : : :: |
|: : : :: |
| : : |
1e+09 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.branch-instructions
6e+11 ++*--*-*--*-*-*-**-*---*--*-*--*-*-*-*--*---**---*-**-*-*-**-*-*-**-*
O OO O O OO O O OO O O OO O O OO O O : : * * |
5e+11 ++: : : : : |
| : : : : : |
|: : : : : |
4e+11 ++ : : : : |
|: : : : : |
3e+11 ++ : : :: |
|: : : :: |
2e+11 ++ : : :: |
|: : : :: |
| : : |
1e+11 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.branch-misses
2e+10 ++----------------------------------------------------------------+
1.8e+10 O+OO O OO O O OO O OO O OO O OO O O |
| **.*.**.*.*.**.*.* *.**.*.**.*.*.* *.**.*.**.*.**.*.*.**.*.**.*
1.6e+10 ++: : : : : |
1.4e+10 ++: : : : : |
| : : : : : |
1.2e+10 ++ : : : : |
1e+10 ++ : : : : |
8e+09 ++ :: :: |
|: :: :: |
6e+09 ++ :: :: |
4e+09 ++ :: :: |
| : : |
2e+09 ++ : : |
0 *+------------------*----------------*----------------------------+
perf-stat.dTLB-loads
1.4e+12 ++----------------------------------------------------------------+
O OO.O.OO.O.O.OO.O.OO O.OO.O.OO.O.O.* *.**.*.**.*.**.*.*.**.*.**.*
1.2e+12 ++: : : : : |
| : : : : : |
1e+12 ++: : : : : |
|: : : : : |
8e+11 ++ : : : : |
|: : : : : |
6e+11 ++ :: :: |
|: :: :: |
4e+11 ++ :: :: |
| : : |
2e+11 ++ : : |
| : : |
0 *+------------------*----------------*----------------------------+
perf-stat.dTLB-stores
8e+11 ++-----*--*-*---------------*--*-*-----------*---*-*------*--*----*-*
O OO.O O OO O O.OO.O O OO.O O OO O O.* *.*.* * *.*.* * *.* |
7e+11 ++: : : : : |
6e+11 ++: : : : : |
| : : : : : |
5e+11 ++ : : : : |
|: : : : : |
4e+11 ++ : : : : |
|: : : :: |
3e+11 ++ : : :: |
2e+11 ++ : : :: |
| : : |
1e+11 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.iTLB-loads
3.5e+12 ++----------------------------------------------------------------+
| |
3e+12 O+OO.O.OO.O.O.OO.O.OO O.OO.O.OO.O.O.* *.**.*.**.*.**.*.*.**.*.**.*
| : : : : : |
2.5e+12 ++: : : : : |
| : : : : : |
2e+12 ++ : : : : |
|: : : : : |
1.5e+12 ++ :: :: |
|: :: :: |
1e+12 ++ :: :: |
|: :: :: |
5e+11 ++ : : |
| : : |
0 *+------------------*----------------*----------------------------+
perf-stat.node-loads
8e+08 ++**-*-*-**-*-*-**-*---**-*-*-**-*-*-*--*-*-**-*-*-**-*-*-**-*-*-**-*
| : : : : : |
7e+08 ++: : : : : |
6e+08 O+OO O O OO O O OO O O OO O O OO O O : : |
|: : : : : |
5e+08 ++ : : : : |
|: : : : : |
4e+08 ++ : : :: |
|: : : :: |
3e+08 ++ : : :: |
2e+08 ++ : : :: |
| : : |
1e+08 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.node-stores
5e+09 ++--------*-------------------*----------------*--------*---------+
4.5e+09 ++**.*.* + *.**.*.* *.**.*. : *.*.* *.**.*. : *.**.* **.*.**.*
| : * : : * : : * |
4e+09 ++: O : : : : |
3.5e+09 O+OO OO O O OO O OO O OO O OO O O : : |
|: : : : : |
3e+09 ++ : : : : |
2.5e+09 ++ : : : : |
2e+09 ++ :: :: |
|: :: :: |
1.5e+09 ++ :: :: |
1e+09 ++ : : |
| : : |
5e+08 ++ : : |
0 *+------------------*----------------*----------------------------+
perf-stat.context-switches
6e+07 ++------------------------------------------------------------------+
| |
5e+07 ++**.*.*.**.*.*.**.* **.*.*.**.*.*.* *.*.**.*.*.**.*.*.**.*.*.**.*
O OO O O OO O O OO O O OO O O OO O O : : |
| : : : : : |
4e+07 ++: : : : : |
|: : : : : |
3e+07 ++ : : : : |
|: : : :: |
2e+07 ++ : : :: |
|: : : :: |
|: :: :: |
1e+07 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.cpu-migrations
9e+06 ++------------------------------------------------------------------+
O OO O O OO O O OO O O OO O O OO O O |
8e+06 ++ |
7e+06 ++ |
| |
6e+06 ++ |
5e+06 ++ |
| |
4e+06 ++**.*.*.**.*.*.**.* **.*.*.**.*.*.* *.*.**.*.*.**.*.*.**.*.*.**.*
3e+06 ++: : : : : |
|: : : : : |
2e+06 ++ : : :: |
1e+06 ++ : : :: |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.cache-miss-rate_
7 ++----------------------------------------------------------------------+
| .*. *. .*.*.*.**. .*. .* .*.*.*.* .*. .*. .*.**.*. .*.*.**.*.*.*
6 ++* * * * * * * : * * * * |
| : : : : : |
5 O+O O OO O O O O OO O O O O O OO O O O : : |
| : : : : : |
4 ++ : : : : |
|: : : : : |
3 ++ : : : : |
|: : : :: |
2 ++ : : :: |
|: :: : |
1 ++ : : |
| : : |
0 *+--------------------*------------------*------------------------------+
perf-stat.branch-miss-rate_
3.5 ++--------------------------------------------------------------------+
O O OO O O O OO O O O OO O O O OO O O |
3 ++*.**.*.*.*.**.*.*.* *.*.*.*.**.*.*.* *.*.*.**.*.*.*.**.*.*.*.**.*.*
| : : : : : |
2.5 ++: : : : : |
| : : : : : |
2 ++ : : : : |
|: : : : : |
1.5 ++ : : :: |
|: :: :: |
1 ++ :: :: |
|: : :: |
0.5 ++ : : |
| : : |
0 *+--------------------*----------------*------------------------------+
perf-stat.iTLB-load-miss-rate_
0.045 O+OO-O-O-OO-O-O-OO-O-O-OO-O-O-OO-O-O----*----*------*-----*------**-+
| :*.* **.* **.* :*.* **.* * : *.* *.*.* *.* *.*.* *
0.04 ++: : : : : |
0.035 ++: : : : : |
|: : : : : |
0.03 ++ : : : : |
0.025 ++ : : : : |
|: : : : : |
0.02 ++ : : :: |
0.015 ++ : : :: |
|: : : :: |
0.01 ++ : : |
0.005 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
perf-stat.instructions-per-iTLB-miss
2500 ++-------------------------------------------------------------------+
O O.OO.O.O.OO.O.O.OO O O.O.OO.O.O.OO.* **.*.*.*.**.*.*.**.*.*.**.*.*
| : : : : : |
2000 ++: : : : : |
| : : : : : |
|: : : : : |
1500 ++ : : : : |
|: : : : : |
1000 ++ : : : : |
|: : : : : |
|: : : : : |
500 ++ : : |
| : : |
| : : |
0 *+-------------------*-----------------*-----------------------------+
turbostat.Avg_MHz
2500 ++-------------------------------------------------------------------+
| |
O O.OO.O.O.OO.O.O.OO O O.O.OO.O.O.OO.* **.*.*.*.**.*.*.**.*.*.**.*.*
2000 ++: : : : : |
| : : : : : |
| : : : : : |
1500 ++ : : : : |
|: : : : : |
1000 ++ : : : : |
|: : : : : |
|: : : : : |
500 ++ :: :: |
| : : |
| : : |
0 *+-------------------*-----------------*-----------------------------+
turbostat._Busy
80 ++*-**-*-*-*-*-**-*-*---**-*-*-*-*-**-*---*-*-**-*-*-*-**-*-*-*-*-**-*-*
O O OO O O O O OO O O O OO O O O O OO : : |
70 ++: : : : : |
60 ++: : : : : |
|: : : : : |
50 ++ : : : : |
|: : : : : |
40 ++ : : : : |
|: : : : : |
30 ++ : : : : |
20 ++ : : : : |
| : : |
10 ++ : : |
| : : |
0 *+--------------------*-----------------*------------------------------+
interrupts.CAL:Function_call_interrupts
60000 ++------------------------------------------------------------------+
| |
50000 O+ O O O O O O O O O |
| O O OO O OO O O O O |
| **.*.*.**.*.*.**.* **.*.*.**.*.*.* *.*.**.*.*.**.*.*.**.*.*.**.*
40000 ++: : : : : |
| : : : : : |
30000 ++ : : : : |
|: : : : : |
20000 ++ : : :: |
|: : : :: |
|: : : :: |
10000 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
unixbench.time.user_time
1200 ++-------------------------------------------------------------------+
| *.**.*.*.**.*.*.** *.*.**.*.*.**.* **.*.*.*.**.*.*.**.*.*.**.*.*
1000 O+O OO O O OO O O OO O O O OO O O OO : : |
| : : : : : |
| : : : : : |
800 ++ : : : : |
|: : : : : |
600 ++ : : : : |
|: : : : : |
400 ++ : : : : |
|: : : : : |
| : : |
200 ++ : : |
| : : |
0 *+-------------------*-----------------*-----------------------------+
unixbench.time.system_time
900 ++--------------------------------------------------------------------+
| *.**.*.*.*.**.*.*.* *.*.*.*.**.*.*.* *.*.*.**.*.*.*.**.*.*.*.**.*.*
800 O+O OO O O O OO O O O OO O O O OO O O : : |
700 ++: : : : : |
| : : : : : |
600 ++ : : : : |
500 ++ : : : : |
|: : : : : |
400 ++ :: :: |
300 ++ :: :: |
|: :: :: |
200 ++ : : |
100 ++ : : |
| : : |
0 *+--------------------*----------------*------------------------------+
unixbench.time.percent_of_cpu_this_job_got
600 ++*-**-*-*-*-**-*-*-*--*-*-*-*-**-*-*-*--*-*-*-**-*-*-*-**-*-*-*-**-*-*
O O OO O O O OO O O O OO O O O OO O O : : |
500 ++: : : : : |
| : : : : : |
|: : : : : |
400 ++ : : : : |
|: : : : : |
300 ++ : : :: |
|: :: :: |
200 ++ :: :: |
|: :: :: |
| : : |
100 ++ : : |
| : : |
0 *+--------------------*----------------*------------------------------+
unixbench.time.involuntary_context_switches
5e+06 ++------------*-------------------*----------------*----*-*-----*-*
4.5e+06 ++**. .**.*.* *.*. *.* .*.**.* * *.* .*.**.* *.* *.*.* |
| : * * : * : : * |
4e+06 ++: : : : : |
3.5e+06 ++: : : : : |
|: : : : : |
3e+06 ++ O O O O O OO:O O OO : : |
2.5e+06 O+O O O O O : : O O O O : : |
2e+06 ++ :: :: |
|: :: :: |
1.5e+06 ++ :: :: |
1e+06 ++ : : |
| : : |
500000 ++ : : |
0 *+------------------*----------------*----------------------------+
vmstat.system.in
35000 ++------------------------------------------------------------------+
| |
30000 ++ .*. .*. *. .**. *.*
| **.*.*.**.* **.* **.*.*.**.* * *.*.**.*.*.* *.* *.*.* |
25000 ++: : : : : |
O :O OO O O O O OO OO : : |
20000 ++O O O OO : : O O O O : : |
|: : : : : |
15000 ++ : : : : |
|: : : :: |
10000 ++ : : :: |
|: : : :: |
5000 ++ : : |
| : : |
0 *+-------------------*----------------*-----------------------------+
vmstat.system.cs
160000 ++------------*-------------------*-----------------*----**------*-*
| **.*.*.OO.O O*.*.* OO.*.**.O.*.* * **.*.**.*.* *.* *.*.* |
140000 O+OO O O O O O : O OO O OO : : |
120000 ++: : : : : |
| : : : : : |
100000 ++ : : : : |
|: : : : : |
80000 ++ : : : : |
|: :: : : |
60000 ++ :: : : |
40000 ++ :: : : |
| : : |
20000 ++ : : |
| : : |
0 *+-------------------*----------------*----------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
4 years, 1 month
[lkp] [mm] 47c2d84f2d: [No primary change] interrupts.CAL:Function_call_interrupts -64% decrease
by kernel test robot
Greeting,
There is no primary kpi change in this test, below is the data collected through multiple monitors running background just for your information.
commit: 47c2d84f2dc7d29d10713beef3ec1239ec60f883 ("mm: remove the page size change check in tlb_remove_page")
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git master
in testcase: pbzip2
on test machine: 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz with 64G memory
with following parameters:
nr_threads: 25%
blocksize: 900K
cpufreq_governor: performance
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
testcase/path_params/tbox_group/run: pbzip2/25%-900K-performance/ivb42
7a8e9a729179b749 47c2d84f2dc7d29d10713beef3
---------------- --------------------------
%stddev change %stddev
\ | \
2249503 -64% 812158 interrupts.CAL:Function_call_interrupts
1778 ± 4% 8% 1925 vmstat.system.cs
57359 -8% 52680 vmstat.system.in
530318 ± 4% 9% 575717 perf-stat.context-switches
4.804e+09 4.746e+09 perf-stat.node-stores
60.89 60.12 perf-stat.iTLB-load-miss-rate%
86802648 85115436 perf-stat.iTLB-loads
1.351e+08 -5% 1.283e+08 perf-stat.iTLB-load-misses
0.04 -7% 0.04 perf-stat.dTLB-store-miss-rate%
7.374e+08 -21% 5.844e+08 ± 3% perf-stat.dTLB-load-misses
interrupts.CAL:Function_call_interrupts
2.4e+06 ++----------------------------------------------------------------+
** **.** *.* * *** *.** *****. **** .** **** * ** **.****|
2.2e+06 ++* ***** * * * * * *** * **.* * * *
2e+06 ++ |
| |
1.8e+06 ++ |
1.6e+06 ++ |
| |
1.4e+06 ++ |
1.2e+06 ++ |
| |
1e+06 ++ |
800000 ++ OO OOOOOO OO OO |
OOO OO O |
600000 ++----------------------------------------------------------------+
vmstat.system.in
58000 ++------------------------------------------------------------------+
* *. ** .* ** * ** * ** * * ** * * * *.** *
57000 +** ** ** ** **.*** :*.* ** *.* ** *.* ** *.* ** :+ ** * *|
| * * * |
56000 ++ |
| |
55000 ++ |
| |
54000 ++ |
| |
53000 ++ O O O O O |
| O OO O O O O O |
52000 OOO OO |
| |
51000 ++------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
4 years, 1 month
[lkp] [x86/tsc] 8b223bc7ab: BUG:unable_to_handle_kernel
by kernel test robot
FYI, we noticed the following commit:
commit: 8b223bc7abe0e30e8d297a24ee6c6c07ef8d0bb9 ("x86/tsc: Store and check TSC ADJUST MSR")
https://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git x86/timers
in testcase: boot
on test machine: qemu-system-x86_64 -enable-kvm -cpu host -smp 2 -m 1G
caused below changes:
+------------------------------------------------------------------+------------+------------+
| | bec8520dca | 8b223bc7ab |
+------------------------------------------------------------------+------------+------------+
| boot_successes | 2 | 0 |
| boot_failures | 4 | 4 |
| WARNING:at_fs/sysfs/dir.c:#sysfs_warn_dup | 1 | |
| calltrace:parport_pc_init | 1 | |
| calltrace:SyS_finit_module | 1 | |
| WARNING:at_lib/kobject.c:#kobject_add_internal | 1 | |
| BUG:kernel_reboot-without-warning_in_test_stage | 1 | |
| invoked_oom-killer:gfp_mask=0x | 2 | |
| Mem-Info | 2 | |
| Kernel_panic-not_syncing:Out_of_memory_and_no_killable_processes | 2 | |
| BUG:unable_to_handle_kernel | 0 | 4 |
| Oops | 0 | 4 |
| RIP:_find_next_bit | 0 | 4 |
| Kernel_panic-not_syncing:Fatal_exception | 0 | 4 |
+------------------------------------------------------------------+------------+------------+
[ 0.000000] clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns
[ 0.000000] hpet clockevent registered
[ 0.000000] tsc: Detected 2593.992 MHz processor
[ 0.000000] BUG: unable to handle kernel NULL pointer dereference at (null)
[ 0.000000] IP: [<ffffffff8147c1a5>] _find_next_bit+0x15/0x70
[ 0.000000] PGD 0
[ 0.000000]
[ 0.000000] Oops: 0000 [#1] SMP
[ 0.000000] Modules linked in:
[ 0.000000] CPU: 0 PID: 0 Comm: swapper/0 Not tainted 4.9.0-rc5-00008-g8b223bc #1
[ 0.000000] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS Debian-1.8.2-1 04/01/2014
[ 0.000000] task: ffffffff81e0e4c0 task.stack: ffffffff81e00000
[ 0.000000] RIP: 0010:[<ffffffff8147c1a5>] [<ffffffff8147c1a5>] _find_next_bit+0x15/0x70
[ 0.000000] RSP: 0000:ffffffff81e03eb0 EFLAGS: 00010246
[ 0.000000] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000000
[ 0.000000] RDX: 0000000000000000 RSI: 0000000000000002 RDI: 0000000000000000
[ 0.000000] RBP: ffffffff81e03ec0 R08: ffffffffffffffff R09: 0000000000000000
[ 0.000000] R10: 0000000000000020 R11: 000000000013ca64 R12: 0000000000000000
[ 0.000000] R13: 0000000000000000 R14: 0000000000000000 R15: 000000000000008e
[ 0.000000] FS: 0000000000000000(0000) GS:ffff880029a00000(0000) knlGS:0000000000000000
[ 0.000000] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 0.000000] CR2: 0000000000000000 CR3: 0000000001e07000 CR4: 00000000000406b0
[ 0.000000] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[ 0.000000] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
[ 0.000000] Stack:
[ 0.000000] ffffffff81e03ec0 ffffffff8147c218 ffffffff81e03ee0 ffffffff81465a46
[ 0.000000] ffff880029a0d580 000000000000d580 ffffffff81e03f18 ffffffff81053005
[ 0.000000] 0000000000000002 0000000000016a00 ffffffff820b4920 ffffffff820c12e0
[ 0.000000] Call Trace:
[ 0.000000] [<ffffffff8147c218>] ? find_next_bit+0x18/0x20
[ 0.000000] [<ffffffff81465a46>] cpumask_any_but+0x26/0x50
[ 0.000000] [<ffffffff81053005>] tsc_store_and_check_tsc_adjust+0x75/0x120
[ 0.000000] [<ffffffff8200a518>] tsc_init+0x19f/0x24d
[ 0.000000] [<ffffffff82006df7>] x86_late_time_init+0xf/0x11
[ 0.000000] [<ffffffff81ffbeb0>] start_kernel+0x35b/0x414
[ 0.000000] [<ffffffff81ffb120>] ? early_idt_handler_array+0x120/0x120
[ 0.000000] [<ffffffff81ffb2d6>] x86_64_start_reservations+0x2a/0x2c
[ 0.000000] [<ffffffff81ffb415>] x86_64_start_kernel+0x13d/0x14c
[ 0.000000] Code: d0 48 0f 47 c2 5d c3 0f 1f 44 00 00 66 2e 0f 1f 84 00 00 00 00 00 48 89 c8 48 89 d1 49 c7 c0 ff ff ff ff 48 c1 e9 06 49 89 c1 55 <4c> 33 0c cf 89 d1 48 83 e2 c0 49 d3 e0 48 89 e5 4c 89 c1 4c 21
[ 0.000000] RIP [<ffffffff8147c1a5>] _find_next_bit+0x15/0x70
[ 0.000000] RSP <ffffffff81e03eb0>
[ 0.000000] CR2: 0000000000000000
[ 0.000000] ---[ end trace a444a5b30abeb907 ]---
[ 0.000000] Kernel panic - not syncing: Fatal exception
Elapsed time: 40
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> job-script # job-script is attached in this email
Thanks,
Xiaolong
4 years, 1 month
[lkp] [mm] e7c1db75fe: BUG:sleeping_function_called_from_invalid_context_at_mm/page_alloc.c
by kernel test robot
FYI, we noticed the following commit:
commit e7c1db75fed821a961ce1ca2b602b08e75de0cd8 ("mm: Prevent __alloc_pages_nodemask() RCU CPU stall warnings")
https://git.kernel.org/pub/scm/linux/kernel/git/paulmck/linux-rcu.git rcu/next
in testcase: boot
on test machine: qemu-system-x86_64 -enable-kvm -cpu Nehalem -smp 2 -m 1G
caused below changes:
+----------------------------------------------------------------------+------------+------------+
| | 7cebc6b63b | e7c1db75fe |
+----------------------------------------------------------------------+------------+------------+
| boot_successes | 0 | 0 |
| boot_failures | 14 | 14 |
| invoked_oom-killer:gfp_mask=0x | 2 | |
| Mem-Info | 2 | |
| Kernel_panic-not_syncing:Out_of_memory_and_no_killable_processes | 2 | |
| BUG:sleeping_function_called_from_invalid_context_at_mm/swap.c | 12 | 14 |
| calltrace:run_init_process | 12 | 14 |
| calltrace:SyS_execve | 7 | 5 |
| calltrace:SyS_exit_group | 12 | 13 |
| calltrace:SyS_read | 1 | |
| WARNING:at_fs/sysfs/dir.c:#sysfs_warn_dup | 2 | 1 |
| calltrace:parport_pc_init | 2 | 1 |
| calltrace:SyS_finit_module | 2 | 1 |
| WARNING:at_lib/kobject.c:#kobject_add_internal | 2 | 1 |
| calltrace:SyS_munmap | 2 | |
| BUG:sleeping_function_called_from_invalid_context_at_mm/page_alloc.c | 0 | 8 |
| calltrace:SyS_clone | 0 | 1 |
+----------------------------------------------------------------------+------------+------------+
[ 8.953192] BUG: sleeping function called from invalid context at mm/page_alloc.c:3746
[ 8.956353] in_atomic(): 1, irqs_disabled(): 1, pid: 0, name: swapper/0
[ 8.958262] CPU: 0 PID: 0 Comm: swapper/0 Tainted: G W 4.9.0-rc1-00048-ge7c1db7 #1
[ 8.961309] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS Debian-1.8.2-1 04/01/2014
[ 8.964341] ffff880029a03720 ffffffff81472549 ffffffff81e0d4c0 0000000002281220
[ 8.967460] ffff880029a03738 ffffffff810a8583 ffff8800364be080 ffff880029a037c0
[ 8.970582] ffffffff81190be5 ffffffff81e0d4c0 ffffffff81e0d4c0 ffffffff81e0d4c0
[ 8.973694] Call Trace:
[ 8.974871] <IRQ>
[ 8.975317] [<ffffffff81472549>] dump_stack+0x63/0x8a
[ 8.977116] [<ffffffff810a8583>] ___might_sleep+0xd3/0x120
[ 8.978852] [<ffffffff81190be5>] __alloc_pages_nodemask+0xd5/0x2f0
[ 8.980663] [<ffffffff811e5335>] alloc_pages_current+0x95/0x140
[ 8.982401] [<ffffffff811efd70>] new_slab+0x3c0/0x5a0
[ 8.984001] [<ffffffff811f1b4c>] ? kfree+0x15c/0x170
[ 8.985648] [<ffffffff811f1110>] ___slab_alloc+0x3a0/0x4b0
[ 8.987316] [<ffffffff818734a1>] ? inet_twsk_alloc+0x41/0x160
[ 8.989031] [<ffffffff818734a1>] ? inet_twsk_alloc+0x41/0x160
[ 8.990743] [<ffffffff811f1240>] __slab_alloc+0x20/0x40
[ 8.992365] [<ffffffff811f264f>] kmem_cache_alloc+0x17f/0x1c0
[ 8.994079] [<ffffffff818734a1>] inet_twsk_alloc+0x41/0x160
[ 8.995763] [<ffffffff8189103e>] tcp_time_wait+0x4e/0x290
[ 8.997414] [<ffffffff81884dbb>] tcp_rcv_state_process+0xd6b/0xd70
[ 8.999199] [<ffffffff813fd0fb>] ? security_sock_rcv_skb+0x3b/0x50
[ 9.000989] [<ffffffff8188e532>] tcp_v4_do_rcv+0xb2/0x200
[ 9.002649] [<ffffffff8188fde0>] tcp_v4_rcv+0x910/0xa10
[ 9.004270] [<ffffffff81905000>] ? ipv6_chk_mcast_addr+0x120/0x140
[ 9.006068] [<ffffffff818e03a8>] ? ip6_input_finish+0xf8/0x460
[ 9.007792] [<ffffffff81868a71>] ip_local_deliver_finish+0xa1/0x1d0
[ 9.009590] [<ffffffff81868d50>] ip_local_deliver+0x60/0xe0
[ 9.011266] [<ffffffff8188f4ba>] ? tcp_v4_early_demux+0x13a/0x150
[ 9.013038] [<ffffffff818686fd>] ip_rcv_finish+0xdd/0x3b0
[ 9.014693] [<ffffffff81869033>] ip_rcv+0x263/0x390
[ 9.016257] [<ffffffff81664fe0>] ? e1000_intr+0xa0/0x120
[ 9.017904] [<ffffffff81824993>] __netif_receive_skb_core+0x513/0xa60
[ 9.019733] [<ffffffff8105a52b>] ? native_apic_msr_write+0x2b/0x30
[ 9.021513] [<ffffffff8189521c>] ? tcp4_gro_receive+0x11c/0x1c0
[ 9.023267] [<ffffffff81824ef8>] __netif_receive_skb+0x18/0x60
[ 9.024998] [<ffffffff81824f78>] netif_receive_skb_internal+0x38/0xb0
[ 9.026822] [<ffffffff81826054>] napi_gro_receive+0xd4/0x130
[ 9.028514] [<ffffffff816679aa>] e1000_clean_rx_irq+0x19a/0x510
[ 9.030252] [<ffffffff81666e00>] e1000_clean+0x260/0x8c0
[ 9.031892] [<ffffffff810b7798>] ? account_entity_enqueue+0xd8/0x100
[ 9.033707] [<ffffffff818257f9>] net_rx_action+0x229/0x390
[ 9.035367] [<ffffffff81971934>] __do_softirq+0x104/0x2ab
[ 9.037032] [<ffffffff81086bf1>] irq_exit+0xf1/0x100
[ 9.038616] [<ffffffff81971674>] do_IRQ+0x54/0xd0
[ 9.040149] [<ffffffff8196f50c>] common_interrupt+0x8c/0x8c
[ 9.041828] <EOI>
[ 9.042256] [<ffffffff8196e2e6>] ? native_safe_halt+0x6/0x10
[ 9.044674] [<ffffffff8196dfe0>] default_idle+0x20/0xf0
[ 9.046295] [<ffffffff81038d1f>] arch_cpu_idle+0xf/0x20
[ 9.047923] [<ffffffff8196e453>] default_idle_call+0x23/0x30
[ 9.049628] [<ffffffff810c75d9>] cpu_startup_entry+0x189/0x200
[ 9.051355] [<ffffffff81960785>] rest_init+0x85/0x90
[ 9.053096] [<ffffffff81ffaf7a>] start_kernel+0x407/0x414
[ 9.054860] [<ffffffff81ffa120>] ? early_idt_handler_array+0x120/0x120
[ 9.056765] [<ffffffff81ffa2d6>] x86_64_start_reservations+0x2a/0x2c
[ 9.058658] [<ffffffff81ffa415>] x86_64_start_kernel+0x13d/0x14c
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp qemu -k <bzImage> job-script # job-script is attached in this email
Thanks,
Xiaolong
4 years, 1 month
[lkp] [mm] 79404c5a5c: netperf.Throughput_Mbps +23.0% improvement
by kernel test robot
Greeting,
FYI, we noticed a +23.0% improvement of netperf.Throughput_Mbps due to commit:
commit 79404c5a5c66481aa55c0cae685e49e0f44a0479 ("mm: page_alloc: High-order per-cpu page allocator")
https://git.kernel.org/pub/scm/linux/kernel/git/mel/linux.git mm-pagealloc-highorder-percpu-v3r1
+------------------+-----------------------------------------------------------------+
| testcase: change | netperf: netperf.Throughput_Mbps 23.0% improvement |
| test machine | 8 threads Intel(R) Atom(TM) CPU C2750 @ 2.40GHz with 16G memory |
| test parameters | cluster=cs-localhost |
| | cpufreq_governor=performance |
| | ip=ipv4 |
| | nr_threads=25% |
| | runtime=300s |
| | send_size=10K |
| | test=SCTP_STREAM_MANY |
+------------------+-----------------------------------------------------------------+
test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/
In addition to that, the commit also has significant impact on the following tests:
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
***************************************************************************************************
lkp-avoton3: 8 threads Intel(R) Atom(TM) CPU C2750 @ 2.40GHz with 16G memory
=========================================================================================
cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase:
cs-localhost/gcc-6/performance/ipv4/x86_64-rhel-7.2/25%/debian-x86_64-2016-08-31.cgz/300s/10K/lkp-avoton3/SCTP_STREAM_MANY/netperf
commit:
v4.9-rc5
79404c5a5c ("mm: page_alloc: High-order per-cpu page allocator")
v4.9-rc5 79404c5a5c66481aa55c0cae68
---------------- --------------------------
fail:runs %reproduction fail:runs
| | |
%stddev %change %stddev
\ | \
4124 ± 6% +23.0% 5074 ± 1% netperf.Throughput_Mbps
165.20 ± 5% +8.2% 178.75 ± 1% netperf.time.percent_of_cpu_this_job_got
490.26 ± 5% +8.0% 529.39 ± 0% netperf.time.system_time
56253 ± 8% +20.0% 67530 ± 12% meminfo.DirectMap4k
4332800 ± 6% +22.8% 5319765 ± 1% softirqs.NET_RX
57290 ± 6% +22.7% 70278 ± 1% vmstat.system.cs
8725309 ± 7% +23.5% 10773756 ± 1% cpuidle.C1-AVN.usage
18968 ± 6% +22.3% 23200 ± 1% cpuidle.POLL.usage
323.15 ± 0% -100.0% 0.00 ± -1% uptime.boot
1899 ± 1% -100.0% 0.00 ± -1% uptime.idle
31.90 ± 5% +10.7% 35.31 ± 0% turbostat.%Busy
817.80 ± 7% +12.3% 918.25 ± 0% turbostat.Avg_MHz
38.00 ± 8% +20.4% 45.75 ± 3% turbostat.CoreTmp
39068988 ± 6% +22.9% 48019605 ± 1% proc-vmstat.numa_hit
39068960 ± 6% +22.9% 48019576 ± 1% proc-vmstat.numa_local
2.243e+08 ± 6% +23.1% 2.761e+08 ± 1% proc-vmstat.pgalloc_normal
2.243e+08 ± 6% -43.1% 1.276e+08 ± 1% proc-vmstat.pgfree
87247 ± 7% +11.9% 97636 ± 0% sched_debug.cfs_rq:/.min_vruntime.avg
46310 ± 11% +50.1% 69504 ± 2% sched_debug.cfs_rq:/.min_vruntime.min
1.47 ± 9% -18.8% 1.19 ± 16% sched_debug.cpu.clock.stddev
1.47 ± 9% -18.9% 1.19 ± 16% sched_debug.cpu.clock_task.stddev
55358 ± 9% +30.4% 72172 ± 3% sched_debug.cpu.nr_load_updates.min
12494 ± 9% -32.1% 8488 ± 17% sched_debug.cpu.nr_load_updates.stddev
1074035 ± 7% +23.3% 1324227 ± 0% sched_debug.cpu.nr_switches.avg
1577603 ± 10% +19.3% 1882212 ± 5% sched_debug.cpu.nr_switches.max
8.859e+10 ± 6% +17.2% 1.039e+11 ± 0% perf-stat.branch-instructions
4.18 ± 0% -4.5% 3.99 ± 0% perf-stat.branch-miss-rate%
3.699e+09 ± 5% +12.0% 4.145e+09 ± 0% perf-stat.branch-misses
1.183e+11 ± 8% +11.9% 1.324e+11 ± 0% perf-stat.cache-references
17412166 ± 6% +22.7% 21360182 ± 1% perf-stat.context-switches
1.94e+12 ± 7% +13.3% 2.198e+12 ± 0% perf-stat.cpu-cycles
18000 ± 19% -26.1% 13294 ± 2% perf-stat.cpu-migrations
0.91 ± 0% -4.2% 0.87 ± 0% perf-stat.iTLB-load-miss-rate%
4.303e+09 ± 5% +12.8% 4.853e+09 ± 1% perf-stat.iTLB-load-misses
4.698e+11 ± 6% +17.7% 5.528e+11 ± 1% perf-stat.iTLB-loads
4.609e+11 ± 5% +19.9% 5.526e+11 ± 1% perf-stat.instructions
107.14 ± 1% +6.3% 113.88 ± 0% perf-stat.instructions-per-iTLB-miss
0.24 ± 2% +5.6% 0.25 ± 0% perf-stat.ipc
7.80 ± 7% +32.7% 10.34 ± 4% perf-profile.calltrace.cycles-pp.___sys_recvmsg.__sys_recvmsg.sys_recvmsg.entry_SYSCALL_64_fastpath
2.02 ± 12% -20.0% 1.61 ± 6% perf-profile.calltrace.cycles-pp.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg
1.53 ± 14% -35.6% 0.98 ± 6% perf-profile.calltrace.cycles-pp.__kmalloc_node_track_caller.__kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty
1.59 ± 13% -33.2% 1.06 ± 6% perf-profile.calltrace.cycles-pp.__kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user
7.17 ± 12% +29.0% 9.26 ± 4% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.ip_finish_output.ip_output.ip_local_out
6.76 ± 12% +28.6% 8.69 ± 3% perf-profile.calltrace.cycles-pp.__netif_receive_skb.process_backlog.net_rx_action.__softirqentry_text_start.do_softirq_own_stack
6.73 ± 12% +28.3% 8.64 ± 3% perf-profile.calltrace.cycles-pp.__netif_receive_skb_core.__netif_receive_skb.process_backlog.net_rx_action.__softirqentry_text_start
8.83 ± 8% +32.2% 11.67 ± 4% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.sctp_sendmsg.inet_sendmsg.sock_sendmsg
7.09 ± 12% +28.6% 9.11 ± 4% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq_own_stack.do_softirq.__local_bh_enable_ip.ip_finish_output2
7.91 ± 7% +32.8% 10.51 ± 4% perf-profile.calltrace.cycles-pp.__sys_recvmsg.sys_recvmsg.entry_SYSCALL_64_fastpath
0.70 ± 9% +123.3% 1.57 ± 4% perf-profile.calltrace.cycles-pp._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
1.14 ± 17% -22.2% 0.89 ± 3% perf-profile.calltrace.cycles-pp.consume_skb.sctp_chunk_put.sctp_chunk_free.sctp_outq_sack.sctp_cmd_interpreter
3.41 ± 16% -41.3% 2.00 ± 2% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg.inet_sendmsg
2.82 ± 8% +25.4% 3.54 ± 9% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.sock_recvmsg
7.16 ± 12% +28.7% 9.21 ± 4% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.ip_finish_output.ip_output
7.12 ± 12% +28.6% 9.16 ± 4% perf-profile.calltrace.cycles-pp.do_softirq_own_stack.do_softirq.__local_bh_enable_ip.ip_finish_output2.ip_finish_output
36.95 ± 10% +17.6% 43.43 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_fastpath
1.00 ± 15% -59.8% 0.40 ± 58% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.kmalloc_large_node.__kmalloc_node_track_caller.__kmalloc_reserve
7.28 ± 7% +33.2% 9.69 ± 4% perf-profile.calltrace.cycles-pp.inet_recvmsg.sock_recvmsg.___sys_recvmsg.__sys_recvmsg.sys_recvmsg
7.59 ± 12% +29.7% 9.85 ± 4% perf-profile.calltrace.cycles-pp.ip_finish_output.ip_output.ip_local_out.ip_queue_xmit.sctp_v4_xmit
7.55 ± 12% +29.5% 9.78 ± 4% perf-profile.calltrace.cycles-pp.ip_finish_output2.ip_finish_output.ip_output.ip_local_out.ip_queue_xmit
6.34 ± 12% +28.4% 8.14 ± 4% perf-profile.calltrace.cycles-pp.ip_local_deliver.ip_rcv_finish.ip_rcv.__netif_receive_skb_core.__netif_receive_skb
6.29 ± 12% +28.4% 8.08 ± 4% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.ip_local_deliver.ip_rcv_finish.ip_rcv.__netif_receive_skb_core
8.19 ± 14% +32.0% 10.82 ± 4% perf-profile.calltrace.cycles-pp.ip_local_out.ip_queue_xmit.sctp_v4_xmit.sctp_packet_transmit.sctp_outq_flush
7.88 ± 14% +34.6% 10.61 ± 4% perf-profile.calltrace.cycles-pp.ip_output.ip_local_out.ip_queue_xmit.sctp_v4_xmit.sctp_packet_transmit
8.48 ± 12% +29.3% 10.96 ± 4% perf-profile.calltrace.cycles-pp.ip_queue_xmit.sctp_v4_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_outq_uncork
6.54 ± 12% +28.7% 8.41 ± 4% perf-profile.calltrace.cycles-pp.ip_rcv.__netif_receive_skb_core.__netif_receive_skb.process_backlog.net_rx_action
6.39 ± 12% +28.7% 8.23 ± 4% perf-profile.calltrace.cycles-pp.ip_rcv_finish.ip_rcv.__netif_receive_skb_core.__netif_receive_skb.process_backlog
1.08 ± 8% +29.2% 1.39 ± 5% perf-profile.calltrace.cycles-pp.kfree_skb.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.sock_recvmsg
1.48 ± 14% -36.6% 0.94 ± 6% perf-profile.calltrace.cycles-pp.kmalloc_large_node.__kmalloc_node_track_caller.__kmalloc_reserve.__alloc_skb._sctp_make_chunk
0.76 ± 9% +115.4% 1.63 ± 3% perf-profile.calltrace.cycles-pp.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.sock_recvmsg
3.76 ± 6% +45.9% 5.49 ± 8% perf-profile.calltrace.cycles-pp.memcpy_erms.sctp_outq_flush.sctp_outq_uncork.sctp_cmd_interpreter.sctp_do_sm
0.70 ± 9% +123.1% 1.56 ± 4% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg
6.99 ± 12% +28.3% 8.96 ± 4% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq_own_stack.do_softirq.__local_bh_enable_ip
6.86 ± 12% +28.4% 8.81 ± 3% perf-profile.calltrace.cycles-pp.process_backlog.net_rx_action.__softirqentry_text_start.do_softirq_own_stack.do_softirq
8.97 ± 8% +31.6% 11.81 ± 4% perf-profile.calltrace.cycles-pp.release_sock.sctp_sendmsg.inet_sendmsg.sock_sendmsg.___sys_sendmsg
6.95 ± 7% +35.3% 9.40 ± 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_inq_push.sctp_backlog_rcv.__release_sock.release_sock
4.80 ± 13% +25.3% 6.01 ± 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_inq_push.sctp_rcv.ip_local_deliver_finish.ip_local_deliver
8.75 ± 8% +32.3% 11.58 ± 4% perf-profile.calltrace.cycles-pp.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg.inet_sendmsg
1.27 ± 16% -17.3% 1.05 ± 2% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_chunk_free.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm
7.03 ± 8% +35.7% 9.53 ± 4% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_inq_push.sctp_backlog_rcv
3.99 ± 13% +25.2% 5.00 ± 4% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_inq_push.sctp_rcv
6.11 ± 11% +22.1% 7.46 ± 6% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg.inet_sendmsg
7.09 ± 13% -21.8% 5.55 ± 5% perf-profile.calltrace.cycles-pp.sctp_datamsg_from_user.sctp_sendmsg.inet_sendmsg.sock_sendmsg.___sys_sendmsg
6.86 ± 7% +35.9% 9.31 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_inq_push.sctp_backlog_rcv.__release_sock
4.44 ± 13% +25.1% 5.56 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_inq_push.sctp_rcv.ip_local_deliver_finish
6.08 ± 11% +21.7% 7.40 ± 6% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg.inet_sendmsg.sock_sendmsg
6.94 ± 7% +35.0% 9.37 ± 4% perf-profile.calltrace.cycles-pp.sctp_inq_push.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg
4.90 ± 13% +25.4% 6.15 ± 4% perf-profile.calltrace.cycles-pp.sctp_inq_push.sctp_rcv.ip_local_deliver_finish.ip_local_deliver.ip_rcv_finish
7.48 ± 7% +40.6% 10.52 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_outq_uncork.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
7.32 ± 12% +23.7% 9.05 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_outq_uncork.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
7.55 ± 7% +40.5% 10.61 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_uncork.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_inq_push
7.42 ± 12% +23.5% 9.17 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_uncork.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg
9.64 ± 11% +29.3% 12.47 ± 3% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_outq_uncork.sctp_cmd_interpreter.sctp_do_sm
9.07 ± 12% +23.2% 11.17 ± 5% perf-profile.calltrace.cycles-pp.sctp_primitive_SEND.sctp_sendmsg.inet_sendmsg.sock_sendmsg.___sys_sendmsg
6.01 ± 12% +27.9% 7.69 ± 4% perf-profile.calltrace.cycles-pp.sctp_rcv.ip_local_deliver_finish.ip_local_deliver.ip_rcv_finish.ip_rcv
7.17 ± 7% +33.6% 9.58 ± 4% perf-profile.calltrace.cycles-pp.sctp_recvmsg.inet_recvmsg.sock_recvmsg.___sys_recvmsg.__sys_recvmsg
1.96 ± 9% +57.6% 3.08 ± 3% perf-profile.calltrace.cycles-pp.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.sock_recvmsg.___sys_recvmsg
1.44 ± 8% +31.1% 1.89 ± 4% perf-profile.calltrace.cycles-pp.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.sock_recvmsg.___sys_recvmsg
3.68 ± 15% -37.4% 2.30 ± 2% perf-profile.calltrace.cycles-pp.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg.inet_sendmsg.sock_sendmsg
8.53 ± 12% +29.5% 11.04 ± 3% perf-profile.calltrace.cycles-pp.sctp_v4_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_outq_uncork.sctp_cmd_interpreter
3.01 ± 7% +24.6% 3.75 ± 9% perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.sock_recvmsg.___sys_recvmsg
0.96 ± 17% -28.3% 0.69 ± 1% perf-profile.calltrace.cycles-pp.skb_release_all.consume_skb.sctp_chunk_put.sctp_chunk_free.sctp_outq_sack
0.86 ± 9% +28.2% 1.11 ± 5% perf-profile.calltrace.cycles-pp.skb_release_all.kfree_skb.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg
7.42 ± 7% +33.2% 9.89 ± 4% perf-profile.calltrace.cycles-pp.sock_recvmsg.___sys_recvmsg.__sys_recvmsg.sys_recvmsg.entry_SYSCALL_64_fastpath
7.92 ± 7% +32.8% 10.52 ± 4% perf-profile.calltrace.cycles-pp.sys_recvmsg.entry_SYSCALL_64_fastpath
7.80 ± 7% +32.7% 10.35 ± 4% perf-profile.children.cycles-pp.___sys_recvmsg
1.99 ± 12% -18.1% 1.63 ± 2% perf-profile.children.cycles-pp.__alloc_pages_nodemask
0.77 ± 12% +39.6% 1.07 ± 5% perf-profile.children.cycles-pp.__dev_queue_xmit
1.08 ± 16% -74.7% 0.27 ± 14% perf-profile.children.cycles-pp.__free_pages_ok
7.43 ± 12% +28.9% 9.58 ± 4% perf-profile.children.cycles-pp.__local_bh_enable_ip
6.78 ± 12% +28.5% 8.71 ± 3% perf-profile.children.cycles-pp.__netif_receive_skb
6.76 ± 12% +28.3% 8.67 ± 4% perf-profile.children.cycles-pp.__netif_receive_skb_core
8.85 ± 8% +32.0% 11.68 ± 4% perf-profile.children.cycles-pp.__release_sock
1.37 ± 16% +20.6% 1.65 ± 4% perf-profile.children.cycles-pp.__schedule
7.18 ± 12% +28.3% 9.21 ± 4% perf-profile.children.cycles-pp.__softirqentry_text_start
7.92 ± 7% +32.8% 10.51 ± 4% perf-profile.children.cycles-pp.__sys_recvmsg
0.87 ± 10% +102.3% 1.76 ± 3% perf-profile.children.cycles-pp._raw_spin_lock_bh
0.81 ± 11% +38.9% 1.12 ± 5% perf-profile.children.cycles-pp.dev_queue_xmit
7.17 ± 12% +28.8% 9.23 ± 4% perf-profile.children.cycles-pp.do_softirq
7.16 ± 12% +28.7% 9.21 ± 4% perf-profile.children.cycles-pp.do_softirq_own_stack
37.15 ± 10% +17.6% 43.68 ± 2% perf-profile.children.cycles-pp.entry_SYSCALL_64_fastpath
1.51 ± 12% -25.1% 1.13 ± 3% perf-profile.children.cycles-pp.get_page_from_freelist
7.28 ± 7% +33.2% 9.70 ± 4% perf-profile.children.cycles-pp.inet_recvmsg
7.92 ± 12% +29.6% 10.27 ± 4% perf-profile.children.cycles-pp.ip_finish_output
7.85 ± 12% +29.6% 10.17 ± 4% perf-profile.children.cycles-pp.ip_finish_output2
6.35 ± 12% +28.5% 8.16 ± 4% perf-profile.children.cycles-pp.ip_local_deliver
6.31 ± 12% +28.6% 8.11 ± 4% perf-profile.children.cycles-pp.ip_local_deliver_finish
8.40 ± 12% +29.2% 10.85 ± 4% perf-profile.children.cycles-pp.ip_local_out
8.23 ± 12% +29.1% 10.62 ± 4% perf-profile.children.cycles-pp.ip_output
8.49 ± 12% +29.3% 10.98 ± 3% perf-profile.children.cycles-pp.ip_queue_xmit
6.57 ± 12% +28.4% 8.44 ± 4% perf-profile.children.cycles-pp.ip_rcv
6.42 ± 12% +28.5% 8.25 ± 4% perf-profile.children.cycles-pp.ip_rcv_finish
1.08 ± 8% +29.2% 1.40 ± 5% perf-profile.children.cycles-pp.kfree_skb
1.00 ± 8% +90.4% 1.90 ± 3% perf-profile.children.cycles-pp.lock_sock_nested
3.92 ± 5% +43.4% 5.62 ± 7% perf-profile.children.cycles-pp.memcpy_erms
0.83 ± 11% +89.7% 1.58 ± 3% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
7.02 ± 12% +28.2% 9.00 ± 3% perf-profile.children.cycles-pp.net_rx_action
6.88 ± 12% +28.3% 8.83 ± 3% perf-profile.children.cycles-pp.process_backlog
9.10 ± 8% +31.4% 11.95 ± 4% perf-profile.children.cycles-pp.release_sock
1.47 ± 15% +21.1% 1.78 ± 4% perf-profile.children.cycles-pp.schedule
11.77 ± 9% +31.0% 15.43 ± 4% perf-profile.children.cycles-pp.sctp_assoc_bh_rcv
8.77 ± 8% +32.2% 11.59 ± 4% perf-profile.children.cycles-pp.sctp_backlog_rcv
17.20 ± 10% +28.2% 22.05 ± 4% perf-profile.children.cycles-pp.sctp_cmd_interpreter
7.10 ± 13% -21.5% 5.57 ± 5% perf-profile.children.cycles-pp.sctp_datamsg_from_user
17.44 ± 10% +28.1% 22.33 ± 4% perf-profile.children.cycles-pp.sctp_do_sm
11.88 ± 9% +30.9% 15.54 ± 4% perf-profile.children.cycles-pp.sctp_inq_push
14.87 ± 9% +32.3% 19.66 ± 5% perf-profile.children.cycles-pp.sctp_outq_flush
15.05 ± 9% +32.1% 19.88 ± 5% perf-profile.children.cycles-pp.sctp_outq_uncork
9.66 ± 11% +29.2% 12.48 ± 3% perf-profile.children.cycles-pp.sctp_packet_transmit
9.09 ± 12% +23.2% 11.20 ± 5% perf-profile.children.cycles-pp.sctp_primitive_SEND
6.03 ± 12% +27.9% 7.71 ± 4% perf-profile.children.cycles-pp.sctp_rcv
7.17 ± 7% +33.7% 9.59 ± 4% perf-profile.children.cycles-pp.sctp_recvmsg
1.96 ± 9% +57.5% 3.09 ± 3% perf-profile.children.cycles-pp.sctp_skb_recv_datagram
1.45 ± 8% +31.1% 1.89 ± 4% perf-profile.children.cycles-pp.sctp_ulpevent_free
3.70 ± 15% -37.1% 2.33 ± 2% perf-profile.children.cycles-pp.sctp_user_addto_chunk
8.56 ± 12% +29.4% 11.07 ± 3% perf-profile.children.cycles-pp.sctp_v4_xmit
3.02 ± 7% +24.5% 3.75 ± 9% perf-profile.children.cycles-pp.skb_copy_datagram_iter
7.42 ± 7% +33.3% 9.89 ± 4% perf-profile.children.cycles-pp.sock_recvmsg
7.93 ± 7% +32.8% 10.53 ± 4% perf-profile.children.cycles-pp.sys_recvmsg
3.92 ± 5% +43.4% 5.62 ± 7% perf-profile.self.cycles-pp.memcpy_erms
0.83 ± 11% +89.7% 1.58 ± 3% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
4 years, 1 month
[lkp] [sched] d455080958: [No primary change] hackbench.time.involuntary_context_switches -8.0% decrease
by kernel test robot
Greeting,
There is no primary kpi change in this test, below is the data collected through multiple monitors running background just for your information.
commit d4550809586dc011ef8c5a9179e6a4325e7b98e1 ("sched: look for idle cpu at wake up")
https://git.linaro.org/people/vincent.guittot/kernel.git sched/pelt
in testcase: hackbench
on test machine: 8 threads Ivy Bridge with 16G memory
with following parameters:
nr_threads: 50%
mode: process
ipc: socket
cpufreq_governor: performance
test-description: Hackbench is both a benchmark and a stress test for the Linux kernel scheduler.
test-url: https://github.com/linux-test-project/ltp/blob/master/testcases/kernel/sc...
Details are as below:
-------------------------------------------------------------------------------------------------->
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
=========================================================================================
compiler/cpufreq_governor/ipc/kconfig/mode/nr_threads/rootfs/tbox_group/testcase:
gcc-6/performance/socket/x86_64-rhel-7.2/process/50%/debian-x86_64-2016-08-31.cgz/lkp-ivb-d01/hackbench
commit:
fba9fb5f5f ("sched: use load_avg for selecting idlest group")
d455080958 ("sched: look for idle cpu at wake up")
fba9fb5f5f79866e d4550809586dc011ef8c5a9179
---------------- --------------------------
%stddev %change %stddev
\ | \
34162896 ± 1% -8.0% 31437510 ± 1% hackbench.time.involuntary_context_switches
733.00 ± 0% +1.3% 742.25 ± 0% hackbench.time.percent_of_cpu_this_job_got
1.547e+08 ± 1% -9.7% 1.397e+08 ± 0% hackbench.time.voluntary_context_switches
20736 ± 20% +254.5% 73507 ± 25% softirqs.NET_RX
432890 ± 15% +89.1% 818743 ± 18% softirqs.RCU
528670 ± 0% +18.9% 628410 ± 1% softirqs.SCHED
75.25 ± 1% -11.3% 66.75 ± 0% vmstat.procs.r
363821 ± 0% -6.3% 341078 ± 0% vmstat.system.cs
56805 ± 0% -7.4% 52594 ± 0% vmstat.system.in
6352414 ± 2% -17.8% 5221427 ± 1% proc-vmstat.numa_hit
6352382 ± 2% -17.8% 5221393 ± 1% proc-vmstat.numa_local
18746436 ± 1% -18.5% 15281493 ± 1% proc-vmstat.pgalloc_normal
18734627 ± 1% -18.5% 15269275 ± 1% proc-vmstat.pgfree
92.92 ± 0% +1.3% 94.16 ± 0% turbostat.%Busy
3431 ± 0% +1.3% 3476 ± 0% turbostat.Avg_MHz
6.95 ± 0% -17.6% 5.72 ± 0% turbostat.CPU%c1
0.03 ± 0% -50.0% 0.02 ± 33% turbostat.CPU%c3
49520562 ± 3% -47.6% 25951758 ± 1% cpuidle.C1E-IVB.time
874280 ± 1% -17.1% 725099 ± 1% cpuidle.C1E-IVB.usage
24613560 ± 5% -78.2% 5355682 ± 1% cpuidle.C3-IVB.time
196895 ± 3% -63.3% 72279 ± 1% cpuidle.C3-IVB.usage
12706889 ± 5% -45.6% 6918345 ± 1% cpuidle.C6-IVB.time
37427 ± 8% -75.0% 9356 ± 1% cpuidle.C6-IVB.usage
181016 ± 1% +21.0% 218998 ± 2% cpuidle.POLL.usage
0.50 ± 0% +1.7% 0.51 ± 0% perf-stat.branch-miss-rate%
10.45 ± 3% -16.3% 8.75 ± 1% perf-stat.cache-miss-rate%
1.051e+11 ± 1% +13.1% 1.189e+11 ± 1% perf-stat.cache-references
2.227e+08 ± 1% -6.8% 2.075e+08 ± 0% perf-stat.context-switches
15827085 ± 0% +55.5% 24608395 ± 1% perf-stat.cpu-migrations
2.82e+12 ± 1% -2.5% 2.749e+12 ± 1% perf-stat.dTLB-stores
9.469e+08 ± 1% -3.6% 9.126e+08 ± 1% perf-stat.iTLB-load-misses
0.74 ± 0% -2.7% 0.72 ± 1% perf-stat.ipc
16738 ± 11% -31.2% 11515 ± 16% sched_debug.cfs_rq:/.min_vruntime.stddev
16747 ± 11% -31.3% 11512 ± 16% sched_debug.cfs_rq:/.spread0.stddev
1107 ± 4% +11.0% 1229 ± 3% sched_debug.cfs_rq:/.util_avg.avg
1497 ± 6% +12.4% 1683 ± 3% sched_debug.cfs_rq:/.util_avg.max
239.41 ± 11% +19.2% 285.40 ± 5% sched_debug.cfs_rq:/.util_avg.stddev
226803 ± 7% -34.6% 148369 ± 8% sched_debug.cpu.avg_idle.max
69458 ± 8% -41.4% 40697 ± 35% sched_debug.cpu.avg_idle.stddev
1518435 ± 2% +16.3% 1766592 ± 6% sched_debug.cpu.nr_switches.stddev
3.33 ± 2% -8.0% 3.06 ± 1% perf-profile.calltrace.cycles-pp.__kmalloc_reserve.__alloc_skb.alloc_skb_with_frags.sock_alloc_send_pskb.unix_stream_sendmsg
3.02 ± 5% +16.7% 3.52 ± 2% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_sync_key.sock_def_readable.unix_stream_sendmsg.sock_sendmsg
4.20 ± 6% +26.3% 5.30 ± 2% perf-profile.calltrace.cycles-pp.__wake_up_sync_key.sock_def_readable.unix_stream_sendmsg.sock_sendmsg.sock_write_iter
1.01 ± 12% +38.0% 1.39 ± 9% perf-profile.calltrace.cycles-pp._raw_spin_lock.sock_sendmsg.sock_write_iter.__vfs_write.vfs_write
0.77 ± 21% +114.2% 1.66 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__wake_up_sync_key.sock_def_readable.unix_stream_sendmsg.sock_sendmsg
2.92 ± 5% +17.5% 3.43 ± 2% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_sync_key.sock_def_readable.unix_stream_sendmsg
4.46 ± 2% -18.0% 3.65 ± 1% perf-profile.calltrace.cycles-pp.call_cpuidle.cpu_startup_entry.start_secondary
1.05 ± 9% -22.4% 0.82 ± 7% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.skb_copy_datagram_from_iter.unix_stream_sendmsg.sock_sendmsg.sock_write_iter
5.43 ± 2% -13.8% 4.68 ± 1% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary
4.45 ± 2% -18.1% 3.65 ± 1% perf-profile.calltrace.cycles-pp.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
4.42 ± 2% -18.4% 3.61 ± 1% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
2.88 ± 5% +17.2% 3.38 ± 2% perf-profile.calltrace.cycles-pp.default_wake_function.autoremove_wake_function.__wake_up_common.__wake_up_sync_key.sock_def_readable
4.78 ± 3% -25.7% 3.55 ± 1% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
0.71 ± 22% +116.1% 1.54 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__wake_up_sync_key.sock_def_readable.unix_stream_sendmsg
0.00 ± -1% +Inf% 0.82 ± 22% perf-profile.calltrace.cycles-pp.select_task_rq_fair.try_to_wake_up.default_wake_function.autoremove_wake_function.__wake_up_common
1.77 ± 7% +15.5% 2.04 ± 2% perf-profile.calltrace.cycles-pp.skb_queue_tail.unix_stream_sendmsg.sock_sendmsg.sock_write_iter.__vfs_write
5.76 ± 4% +22.9% 7.07 ± 1% perf-profile.calltrace.cycles-pp.sock_def_readable.unix_stream_sendmsg.sock_sendmsg.sock_write_iter.__vfs_write
5.47 ± 2% -13.8% 4.71 ± 1% perf-profile.calltrace.cycles-pp.start_secondary
2.79 ± 5% +18.4% 3.31 ± 3% perf-profile.calltrace.cycles-pp.try_to_wake_up.default_wake_function.autoremove_wake_function.__wake_up_common.__wake_up_sync_key
1.63 ± 2% -13.6% 1.41 ± 1% perf-profile.children.cycles-pp.___slab_alloc
1.79 ± 2% -14.0% 1.54 ± 1% perf-profile.children.cycles-pp.__slab_alloc
3.47 ± 5% +18.1% 4.10 ± 1% perf-profile.children.cycles-pp.__wake_up_common
4.66 ± 5% +26.3% 5.89 ± 1% perf-profile.children.cycles-pp.__wake_up_sync_key
3.78 ± 4% +34.6% 5.08 ± 1% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
3.37 ± 4% +18.9% 4.00 ± 2% perf-profile.children.cycles-pp.autoremove_wake_function
5.01 ± 1% -17.8% 4.12 ± 1% perf-profile.children.cycles-pp.call_cpuidle
6.11 ± 1% -13.4% 5.29 ± 0% perf-profile.children.cycles-pp.cpu_startup_entry
5.00 ± 1% -17.7% 4.11 ± 1% perf-profile.children.cycles-pp.cpuidle_enter
4.96 ± 1% -18.0% 4.07 ± 1% perf-profile.children.cycles-pp.cpuidle_enter_state
3.32 ± 5% +18.5% 3.93 ± 2% perf-profile.children.cycles-pp.default_wake_function
4.89 ± 1% -18.2% 4.00 ± 1% perf-profile.children.cycles-pp.intel_idle
2.36 ± 5% +47.3% 3.48 ± 4% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
0.76 ± 6% +54.4% 1.18 ± 4% perf-profile.children.cycles-pp.prepare_to_wait
0.69 ± 6% +89.2% 1.31 ± 4% perf-profile.children.cycles-pp.select_task_rq_fair
5.80 ± 4% +22.7% 7.11 ± 1% perf-profile.children.cycles-pp.sock_def_readable
5.47 ± 2% -13.8% 4.71 ± 1% perf-profile.children.cycles-pp.start_secondary
3.21 ± 4% +19.8% 3.85 ± 2% perf-profile.children.cycles-pp.try_to_wake_up
1.48 ± 2% +10.7% 1.64 ± 3% perf-profile.children.cycles-pp.unix_write_space
2.44 ± 1% +16.1% 2.83 ± 1% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
4.89 ± 1% -18.2% 4.00 ± 1% perf-profile.self.cycles-pp.intel_idle
2.36 ± 5% +47.3% 3.48 ± 4% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
1.58 ± 0% +13.3% 1.79 ± 3% perf-profile.self.cycles-pp.sock_def_readable
perf-stat.cpu-migrations
2.8e+07 ++----------------------------------------------------------------+
| O |
2.6e+07 O+ O O O O O O O O O O O O O O O O O O |
| O O O O
2.4e+07 ++ O |
| |
2.2e+07 ++ |
| |
2e+07 ++ |
| |
1.8e+07 ++ |
| |
1.6e+07 *+.*..*.*..*..*..*.*..*..*..*.*..*..*..*.*..*..*..*.*..*..*..* |
| |
1.4e+07 ++----------------------------------------------------------------+
turbostat.Avg_MHz
3490 ++-------------------------------------------------------------------+
| O O O O O O O |
3480 O+ O O O O O O O O O O |
| O O O O O O O
| |
3470 ++ |
| |
3460 ++ |
| |
3450 ++ |
| |
| |
3440 ++ .*..*.*..*..*.. .*..*..*..*..*.. |
*..*..*. *..*..*..*..* *.. |
3430 ++-----------------------------------------------------*--*-*--*-----+
turbostat._Busy
94.6 ++-------------------------------------------------------------------+
| O |
94.4 O+ O O O O O O O O O O O O O O O O |
94.2 ++ O O O O O |
| O O
94 ++ |
93.8 ++ |
| |
93.6 ++ |
93.4 ++ |
| |
93.2 ++ .*.. .*..*.. .*.. |
93 *+.*..*..*..*.*. *..*..*..*..*..* *. *..*.. |
| *..*.*..* |
92.8 ++-------------------------------------------------------------------+
hackbench.time.percent_of_cpu_this_job_got
746 ++--------------------------------------------------------------------+
| O |
744 O+ O O O O O O O O O O O O O O O |
| O O O O O |
742 ++ O O O
| |
740 ++ |
| |
738 ++ |
| |
736 ++ |
*.. .*..*.. .*..*.*..*.. .*..*..*..*..*.. |
734 ++ *. *. *..*..*..*. *. |
| *..*..*..* |
732 ++--------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Xiaolong
4 years, 1 month