Re: [LKP] [ext4] d3b6f23f71: stress-ng.fiemap.ops_per_sec -60.5% regression

From: Xing Zhengjun
Date: Mon Apr 13 2020 - 04:43:28 EST


Hi Harjani,

Do you have time to take a look at this? Thanks.

On 4/7/2020 4:00 PM, kernel test robot wrote:
Greeting,

FYI, we noticed a -60.5% regression of stress-ng.fiemap.ops_per_sec due to commit:


commit: d3b6f23f71670007817a5d59f3fbafab2b794e8c ("ext4: move ext4_fiemap to use iomap framework")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: stress-ng
on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
with following parameters:

nr_threads: 10%
disk: 1HDD
testtime: 1s
class: os
cpufreq_governor: performance
ucode: 0x500002c
fs: ext4






Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml

=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
os/gcc-7/performance/1HDD/ext4/x86_64-rhel-7.6/10%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c

commit:
b2c5764262 ("ext4: make ext4_ind_map_blocks work with fiemap")
d3b6f23f71 ("ext4: move ext4_fiemap to use iomap framework")

b2c5764262edded1 d3b6f23f71670007817a5d59f3f
---------------- ---------------------------
fail:runs %reproduction fail:runs
| | |
:4 25% 1:4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
2:4 5% 2:4 perf-profile.calltrace.cycles-pp.sync_regs.error_entry
2:4 6% 3:4 perf-profile.calltrace.cycles-pp.error_entry
3:4 9% 3:4 perf-profile.children.cycles-pp.error_entry
0:4 1% 0:4 perf-profile.self.cycles-pp.error_entry
%stddev %change %stddev
\ | \
28623 +28.2% 36703 ± 12% stress-ng.daemon.ops
28632 +28.2% 36704 ± 12% stress-ng.daemon.ops_per_sec
566.00 ± 22% -53.2% 265.00 ± 53% stress-ng.dev.ops
278.81 ± 22% -53.0% 131.00 ± 54% stress-ng.dev.ops_per_sec
73160 -60.6% 28849 ± 3% stress-ng.fiemap.ops
72471 -60.5% 28612 ± 3% stress-ng.fiemap.ops_per_sec
23421 ± 12% +21.2% 28388 ± 6% stress-ng.filename.ops
22638 ± 12% +20.3% 27241 ± 6% stress-ng.filename.ops_per_sec
21.25 ± 7% -10.6% 19.00 ± 3% stress-ng.iomix.ops
38.75 ± 49% -47.7% 20.25 ± 96% stress-ng.memhotplug.ops
34.45 ± 52% -51.8% 16.62 ±106% stress-ng.memhotplug.ops_per_sec
1734 ± 10% +31.4% 2278 ± 10% stress-ng.resources.ops
807.56 ± 5% +35.2% 1091 ± 8% stress-ng.resources.ops_per_sec
1007356 ± 3% -16.5% 840642 ± 9% stress-ng.revio.ops
1007692 ± 3% -16.6% 840711 ± 9% stress-ng.revio.ops_per_sec
21812 ± 3% +16.0% 25294 ± 5% stress-ng.sysbadaddr.ops
21821 ± 3% +15.9% 25294 ± 5% stress-ng.sysbadaddr.ops_per_sec
440.75 ± 4% +21.9% 537.25 ± 9% stress-ng.sysfs.ops
440.53 ± 4% +21.9% 536.86 ± 9% stress-ng.sysfs.ops_per_sec
13286582 -11.1% 11805520 ± 6% stress-ng.time.file_system_outputs
68253896 +2.4% 69860122 stress-ng.time.minor_page_faults
197.00 ± 4% -15.9% 165.75 ± 12% stress-ng.xattr.ops
192.45 ± 5% -16.1% 161.46 ± 11% stress-ng.xattr.ops_per_sec
15310 +62.5% 24875 ± 22% stress-ng.zombie.ops
15310 +62.5% 24874 ± 22% stress-ng.zombie.ops_per_sec
203.50 ± 12% -47.3% 107.25 ± 49% vmstat.io.bi
861318 ± 18% -29.7% 605884 ± 5% meminfo.AnonHugePages
1062742 ± 14% -20.2% 847853 ± 3% meminfo.AnonPages
31093 ± 6% +9.6% 34090 ± 3% meminfo.KernelStack
7151 ± 34% +55.8% 11145 ± 9% meminfo.Mlocked
1.082e+08 ± 5% -40.2% 64705429 ± 31% numa-numastat.node0.local_node
1.082e+08 ± 5% -40.2% 64739883 ± 31% numa-numastat.node0.numa_hit
46032662 ± 21% +104.3% 94042918 ± 20% numa-numastat.node1.local_node
46074205 ± 21% +104.2% 94072810 ± 20% numa-numastat.node1.numa_hit
3942 ± 3% +14.2% 4501 ± 4% slabinfo.pool_workqueue.active_objs
4098 ± 3% +14.3% 4683 ± 4% slabinfo.pool_workqueue.num_objs
4817 ± 7% +13.3% 5456 ± 8% slabinfo.proc_dir_entry.active_objs
5153 ± 6% +12.5% 5797 ± 8% slabinfo.proc_dir_entry.num_objs
18598 ± 13% -33.1% 12437 ± 20% sched_debug.cfs_rq:/.load.avg
452595 ± 56% -71.4% 129637 ± 76% sched_debug.cfs_rq:/.load.max
67675 ± 35% -55.1% 30377 ± 42% sched_debug.cfs_rq:/.load.stddev
18114 ± 12% -33.7% 12011 ± 20% sched_debug.cfs_rq:/.runnable_weight.avg
448215 ± 58% -72.8% 121789 ± 82% sched_debug.cfs_rq:/.runnable_weight.max
67083 ± 37% -56.3% 29305 ± 43% sched_debug.cfs_rq:/.runnable_weight.stddev
-38032 +434.3% -203212 sched_debug.cfs_rq:/.spread0.avg
-204466 +95.8% -400301 sched_debug.cfs_rq:/.spread0.min
90.02 ± 25% -58.1% 37.69 ± 52% sched_debug.cfs_rq:/.util_est_enqueued.avg
677.54 ± 6% -39.3% 411.50 ± 22% sched_debug.cfs_rq:/.util_est_enqueued.max
196.57 ± 8% -47.6% 103.05 ± 36% sched_debug.cfs_rq:/.util_est_enqueued.stddev
3.34 ± 23% +34.1% 4.48 ± 4% sched_debug.cpu.clock.stddev
3.34 ± 23% +34.1% 4.48 ± 4% sched_debug.cpu.clock_task.stddev
402872 ± 7% -11.9% 354819 ± 2% proc-vmstat.nr_active_anon
1730331 -9.5% 1566418 ± 5% proc-vmstat.nr_dirtied
31042 ± 6% +9.3% 33915 ± 3% proc-vmstat.nr_kernel_stack
229047 -2.4% 223615 proc-vmstat.nr_mapped
74008 ± 7% +20.5% 89163 ± 8% proc-vmstat.nr_written
402872 ± 7% -11.9% 354819 ± 2% proc-vmstat.nr_zone_active_anon
50587 ± 11% -25.2% 37829 ± 14% proc-vmstat.numa_pages_migrated
457500 -23.1% 351918 ± 31% proc-vmstat.numa_pte_updates
81382485 +1.9% 82907822 proc-vmstat.pgfault
2.885e+08 ± 5% -13.3% 2.502e+08 ± 6% proc-vmstat.pgfree
42206 ± 12% -46.9% 22399 ± 49% proc-vmstat.pgpgin
431233 ± 13% -64.8% 151736 ±109% proc-vmstat.pgrotated
176754 ± 7% -40.2% 105637 ± 31% proc-vmstat.thp_fault_alloc
314.50 ± 82% +341.5% 1388 ± 44% proc-vmstat.unevictable_pgs_stranded
1075269 ± 14% -41.3% 631388 ± 17% numa-meminfo.node0.Active
976056 ± 12% -39.7% 588727 ± 19% numa-meminfo.node0.Active(anon)
426857 ± 22% -36.4% 271375 ± 13% numa-meminfo.node0.AnonHugePages
558590 ± 19% -36.4% 355402 ± 14% numa-meminfo.node0.AnonPages
1794824 ± 9% -28.8% 1277157 ± 20% numa-meminfo.node0.FilePages
8517 ± 92% -82.7% 1473 ± 89% numa-meminfo.node0.Inactive(file)
633118 ± 2% -41.7% 368920 ± 36% numa-meminfo.node0.Mapped
2958038 ± 12% -27.7% 2139271 ± 12% numa-meminfo.node0.MemUsed
181401 ± 5% -13.7% 156561 ± 4% numa-meminfo.node0.SUnreclaim
258124 ± 6% -13.0% 224535 ± 5% numa-meminfo.node0.Slab
702083 ± 16% +31.0% 919406 ± 11% numa-meminfo.node1.Active
38663 ±107% +137.8% 91951 ± 31% numa-meminfo.node1.Active(file)
1154975 ± 7% +41.6% 1635593 ± 12% numa-meminfo.node1.FilePages
395813 ± 25% +62.8% 644533 ± 16% numa-meminfo.node1.Inactive
394313 ± 25% +62.5% 640686 ± 16% numa-meminfo.node1.Inactive(anon)
273317 +88.8% 515976 ± 25% numa-meminfo.node1.Mapped
2279237 ± 6% +25.7% 2865582 ± 7% numa-meminfo.node1.MemUsed
10830 ± 18% +29.6% 14033 ± 9% numa-meminfo.node1.PageTables
149390 ± 3% +23.2% 184085 ± 3% numa-meminfo.node1.SUnreclaim
569542 ± 16% +74.8% 995336 ± 21% numa-meminfo.node1.Shmem
220774 ± 5% +20.3% 265656 ± 3% numa-meminfo.node1.Slab
35623587 ± 5% -11.7% 31444514 ± 3% perf-stat.i.cache-misses
2.576e+08 ± 5% -6.8% 2.4e+08 ± 2% perf-stat.i.cache-references
3585 -7.3% 3323 ± 5% perf-stat.i.cpu-migrations
180139 ± 2% +4.2% 187668 perf-stat.i.minor-faults
69.13 +2.6 71.75 perf-stat.i.node-load-miss-rate%
4313695 ± 2% -7.4% 3994957 ± 2% perf-stat.i.node-load-misses
5466253 ± 11% -17.3% 4521173 ± 6% perf-stat.i.node-loads
2818674 ± 6% -15.8% 2372542 ± 5% perf-stat.i.node-stores
227810 +4.6% 238290 perf-stat.i.page-faults
12.67 ± 4% -7.2% 11.76 ± 2% perf-stat.overall.MPKI
1.01 ± 4% -0.0 0.97 ± 3% perf-stat.overall.branch-miss-rate%
1044 +13.1% 1181 ± 4% perf-stat.overall.cycles-between-cache-misses
40.37 ± 4% +3.6 44.00 ± 2% perf-stat.overall.node-store-miss-rate%
36139526 ± 5% -12.5% 31625519 ± 3% perf-stat.ps.cache-misses
2.566e+08 ± 5% -6.9% 2.389e+08 ± 2% perf-stat.ps.cache-references
3562 -7.2% 3306 ± 5% perf-stat.ps.cpu-migrations
179088 +4.2% 186579 perf-stat.ps.minor-faults
4323383 ± 2% -7.5% 3999214 perf-stat.ps.node-load-misses
5607721 ± 10% -18.5% 4568664 ± 6% perf-stat.ps.node-loads
2855134 ± 7% -16.4% 2387345 ± 5% perf-stat.ps.node-stores
226270 +4.6% 236709 perf-stat.ps.page-faults
242305 ± 10% -42.4% 139551 ± 18% numa-vmstat.node0.nr_active_anon
135983 ± 17% -37.4% 85189 ± 10% numa-vmstat.node0.nr_anon_pages
209.25 ± 16% -38.1% 129.50 ± 10% numa-vmstat.node0.nr_anon_transparent_hugepages
449367 ± 9% -29.7% 315804 ± 20% numa-vmstat.node0.nr_file_pages
2167 ± 90% -80.6% 419.75 ± 98% numa-vmstat.node0.nr_inactive_file
157405 ± 3% -41.4% 92206 ± 35% numa-vmstat.node0.nr_mapped
2022 ± 30% -73.3% 539.25 ± 91% numa-vmstat.node0.nr_mlock
3336 ± 10% -24.3% 2524 ± 25% numa-vmstat.node0.nr_page_table_pages
286158 ± 10% -41.2% 168337 ± 37% numa-vmstat.node0.nr_shmem
45493 ± 5% -14.1% 39094 ± 4% numa-vmstat.node0.nr_slab_unreclaimable
242294 ± 10% -42.4% 139547 ± 18% numa-vmstat.node0.nr_zone_active_anon
2167 ± 90% -80.6% 419.75 ± 98% numa-vmstat.node0.nr_zone_inactive_file
54053924 ± 8% -39.3% 32786242 ± 34% numa-vmstat.node0.numa_hit
53929628 ± 8% -39.5% 32619715 ± 34% numa-vmstat.node0.numa_local
9701 ±107% +136.9% 22985 ± 31% numa-vmstat.node1.nr_active_file
202.50 ± 16% -25.1% 151.75 ± 23% numa-vmstat.node1.nr_anon_transparent_hugepages
284922 ± 7% +43.3% 408195 ± 13% numa-vmstat.node1.nr_file_pages
96002 ± 26% +67.5% 160850 ± 17% numa-vmstat.node1.nr_inactive_anon
68077 ± 2% +90.3% 129533 ± 25% numa-vmstat.node1.nr_mapped
138482 ± 15% +79.2% 248100 ± 22% numa-vmstat.node1.nr_shmem
37396 ± 3% +23.3% 46094 ± 3% numa-vmstat.node1.nr_slab_unreclaimable
9701 ±107% +136.9% 22985 ± 31% numa-vmstat.node1.nr_zone_active_file
96005 ± 26% +67.5% 160846 ± 17% numa-vmstat.node1.nr_zone_inactive_anon
23343661 ± 17% +99.9% 46664267 ± 23% numa-vmstat.node1.numa_hit
23248487 ± 17% +100.5% 46610447 ± 23% numa-vmstat.node1.numa_local
105745 ± 23% +112.6% 224805 ± 24% softirqs.CPU0.NET_RX
133310 ± 36% -45.3% 72987 ± 52% softirqs.CPU1.NET_RX
170110 ± 55% -66.8% 56407 ±147% softirqs.CPU11.NET_RX
91465 ± 36% -65.2% 31858 ±112% softirqs.CPU13.NET_RX
164491 ± 57% -77.7% 36641 ±121% softirqs.CPU15.NET_RX
121069 ± 55% -99.3% 816.75 ± 96% softirqs.CPU17.NET_RX
81019 ± 4% -8.7% 73967 ± 4% softirqs.CPU20.RCU
72143 ± 63% -89.8% 7360 ±172% softirqs.CPU22.NET_RX
270663 ± 17% -57.9% 113915 ± 45% softirqs.CPU24.NET_RX
20149 ± 76% +474.1% 115680 ± 62% softirqs.CPU26.NET_RX
14033 ± 70% +977.5% 151211 ± 75% softirqs.CPU27.NET_RX
27834 ± 94% +476.1% 160357 ± 28% softirqs.CPU28.NET_RX
35346 ± 68% +212.0% 110290 ± 30% softirqs.CPU29.NET_RX
34347 ±103% +336.5% 149941 ± 32% softirqs.CPU32.NET_RX
70077 ± 3% +10.8% 77624 ± 3% softirqs.CPU34.RCU
36453 ± 84% +339.6% 160253 ± 42% softirqs.CPU36.NET_RX
72367 ± 2% +10.6% 80043 softirqs.CPU37.RCU
25239 ±118% +267.7% 92799 ± 45% softirqs.CPU38.NET_RX
4995 ±170% +1155.8% 62734 ± 62% softirqs.CPU39.NET_RX
4641 ±145% +1611.3% 79432 ± 90% softirqs.CPU42.NET_RX
7192 ± 65% +918.0% 73225 ± 66% softirqs.CPU45.NET_RX
1772 ±166% +1837.4% 34344 ± 63% softirqs.CPU46.NET_RX
13149 ± 81% +874.7% 128170 ± 58% softirqs.CPU47.NET_RX
86484 ± 94% -92.6% 6357 ±172% softirqs.CPU48.NET_RX
129128 ± 27% -95.8% 5434 ±172% softirqs.CPU55.NET_RX
82772 ± 59% -91.7% 6891 ±164% softirqs.CPU56.NET_RX
145313 ± 57% -87.8% 17796 ± 88% softirqs.CPU57.NET_RX
118160 ± 33% -86.3% 16226 ±109% softirqs.CPU58.NET_RX
94576 ± 56% -94.1% 5557 ±173% softirqs.CPU6.NET_RX
82900 ± 77% -66.8% 27508 ±171% softirqs.CPU62.NET_RX
157291 ± 30% -81.1% 29656 ±111% softirqs.CPU64.NET_RX
135101 ± 28% -80.2% 26748 ± 90% softirqs.CPU67.NET_RX
146574 ± 56% -100.0% 69.75 ± 98% softirqs.CPU68.NET_RX
81347 ± 2% -9.0% 74024 ± 2% softirqs.CPU68.RCU
201729 ± 37% -99.6% 887.50 ±107% softirqs.CPU69.NET_RX
108454 ± 78% -97.9% 2254 ±169% softirqs.CPU70.NET_RX
55289 ±104% -89.3% 5942 ±172% softirqs.CPU71.NET_RX
10112 ±172% +964.6% 107651 ± 89% softirqs.CPU72.NET_RX
3136 ±171% +1522.2% 50879 ± 66% softirqs.CPU73.NET_RX
13353 ± 79% +809.2% 121407 ±101% softirqs.CPU74.NET_RX
75194 ± 3% +10.3% 82957 ± 5% softirqs.CPU75.RCU
11002 ±173% +1040.8% 125512 ± 61% softirqs.CPU76.NET_RX
2463 ±173% +2567.3% 65708 ± 77% softirqs.CPU78.NET_RX
25956 ± 3% -7.8% 23932 ± 3% softirqs.CPU78.SCHED
16366 ±150% +340.7% 72125 ± 91% softirqs.CPU82.NET_RX
14553 ±130% +1513.4% 234809 ± 27% softirqs.CPU93.NET_RX
26314 -9.2% 23884 ± 3% softirqs.CPU93.SCHED
4582 ± 88% +4903.4% 229268 ± 23% softirqs.CPU94.NET_RX
11214 ±111% +1762.5% 208867 ± 18% softirqs.CPU95.NET_RX
1.53 ± 27% -0.5 0.99 ± 17% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.52 ± 27% -0.5 0.99 ± 17% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.39 ± 29% -0.5 0.88 ± 21% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64
1.39 ± 29% -0.5 0.88 ± 21% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.50 ± 59% +0.3 0.81 ± 13% perf-profile.calltrace.cycles-pp.filemap_map_pages.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault
5.70 ± 9% +0.8 6.47 ± 7% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
5.48 ± 9% +0.8 6.27 ± 7% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal
5.49 ± 9% +0.8 6.28 ± 7% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal
4.30 ± 4% +1.3 5.60 ± 7% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode
4.40 ± 4% +1.3 5.69 ± 7% perf-profile.calltrace.cycles-pp.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
4.37 ± 4% +1.3 5.66 ± 7% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
4.36 ± 4% +1.3 5.66 ± 7% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
4.33 ± 4% +1.3 5.62 ± 7% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
4.44 ± 4% +1.3 5.74 ± 7% perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode
3.20 ± 10% -2.4 0.78 ±156% perf-profile.children.cycles-pp.copy_page
0.16 ± 9% -0.1 0.08 ± 64% perf-profile.children.cycles-pp.irq_work_interrupt
0.16 ± 9% -0.1 0.08 ± 64% perf-profile.children.cycles-pp.smp_irq_work_interrupt
0.24 ± 5% -0.1 0.17 ± 18% perf-profile.children.cycles-pp.irq_work_run_list
0.16 ± 9% -0.1 0.10 ± 24% perf-profile.children.cycles-pp.irq_work_run
0.16 ± 9% -0.1 0.10 ± 24% perf-profile.children.cycles-pp.printk
0.23 ± 6% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.__do_execve_file
0.08 ± 14% -0.1 0.03 ±100% perf-profile.children.cycles-pp.delay_tsc
0.16 ± 6% -0.1 0.11 ± 9% perf-profile.children.cycles-pp.load_elf_binary
0.16 ± 7% -0.0 0.12 ± 13% perf-profile.children.cycles-pp.search_binary_handler
0.20 ± 7% -0.0 0.15 ± 10% perf-profile.children.cycles-pp.call_usermodehelper_exec_async
0.19 ± 6% -0.0 0.15 ± 11% perf-profile.children.cycles-pp.do_execve
0.08 ± 10% -0.0 0.04 ± 59% perf-profile.children.cycles-pp.__vunmap
0.15 ± 3% -0.0 0.11 ± 7% perf-profile.children.cycles-pp.rcu_idle_exit
0.12 ± 10% -0.0 0.09 ± 14% perf-profile.children.cycles-pp.__switch_to_asm
0.09 ± 13% -0.0 0.07 ± 5% perf-profile.children.cycles-pp.des3_ede_encrypt
0.06 ± 11% +0.0 0.09 ± 13% perf-profile.children.cycles-pp.mark_page_accessed
0.15 ± 5% +0.0 0.19 ± 12% perf-profile.children.cycles-pp.apparmor_cred_prepare
0.22 ± 8% +0.0 0.27 ± 11% perf-profile.children.cycles-pp.mem_cgroup_throttle_swaprate
0.17 ± 2% +0.0 0.22 ± 12% perf-profile.children.cycles-pp.security_prepare_creds
0.95 ± 17% +0.3 1.22 ± 14% perf-profile.children.cycles-pp.filemap_map_pages
5.92 ± 8% +0.7 6.65 ± 7% perf-profile.children.cycles-pp.get_signal
5.66 ± 9% +0.8 6.44 ± 7% perf-profile.children.cycles-pp.mmput
5.65 ± 9% +0.8 6.43 ± 7% perf-profile.children.cycles-pp.exit_mmap
4.40 ± 4% +1.3 5.70 ± 7% perf-profile.children.cycles-pp.prepare_exit_to_usermode
4.45 ± 4% +1.3 5.75 ± 7% perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
3.16 ± 10% -2.4 0.77 ±155% perf-profile.self.cycles-pp.copy_page
0.08 ± 14% -0.1 0.03 ±100% perf-profile.self.cycles-pp.delay_tsc
0.12 ± 10% -0.0 0.09 ± 14% perf-profile.self.cycles-pp.__switch_to_asm
0.08 ± 12% -0.0 0.06 ± 17% perf-profile.self.cycles-pp.enqueue_task_fair
0.09 ± 13% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.des3_ede_encrypt
0.07 ± 13% +0.0 0.08 ± 19% perf-profile.self.cycles-pp.__lru_cache_add
0.19 ± 9% +0.0 0.22 ± 10% perf-profile.self.cycles-pp.mem_cgroup_throttle_swaprate
0.15 ± 5% +0.0 0.19 ± 11% perf-profile.self.cycles-pp.apparmor_cred_prepare
0.05 ± 58% +0.0 0.09 ± 13% perf-profile.self.cycles-pp.mark_page_accessed
0.58 ± 10% +0.2 0.80 ± 20% perf-profile.self.cycles-pp.release_pages
0.75 ±173% +1.3e+05% 1005 ±100% interrupts.127:PCI-MSI.31981660-edge.i40e-eth0-TxRx-91
820.75 ±111% -99.9% 0.50 ±173% interrupts.47:PCI-MSI.31981580-edge.i40e-eth0-TxRx-11
449.25 ± 86% -100.0% 0.00 interrupts.53:PCI-MSI.31981586-edge.i40e-eth0-TxRx-17
33.25 ±157% -100.0% 0.00 interrupts.57:PCI-MSI.31981590-edge.i40e-eth0-TxRx-21
0.75 ±110% +63533.3% 477.25 ±162% interrupts.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
561.50 ±160% -100.0% 0.00 interrupts.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
82921 ± 8% -11.1% 73748 ± 6% interrupts.CPU11.CAL:Function_call_interrupts
66509 ± 30% -32.6% 44828 ± 8% interrupts.CPU14.TLB:TLB_shootdowns
43105 ± 98% -90.3% 4183 ± 21% interrupts.CPU17.RES:Rescheduling_interrupts
148719 ± 70% -69.4% 45471 ± 16% interrupts.CPU17.TLB:TLB_shootdowns
85589 ± 42% -52.2% 40884 ± 5% interrupts.CPU20.TLB:TLB_shootdowns
222472 ± 41% -98.0% 4360 ± 45% interrupts.CPU22.RES:Rescheduling_interrupts
0.50 ±173% +95350.0% 477.25 ±162% interrupts.CPU25.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
76029 ± 10% +14.9% 87389 ± 5% interrupts.CPU25.CAL:Function_call_interrupts
399042 ± 6% +13.4% 452479 ± 8% interrupts.CPU27.LOC:Local_timer_interrupts
561.00 ±161% -100.0% 0.00 interrupts.CPU29.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
7034 ± 46% +1083.8% 83279 ±138% interrupts.CPU29.RES:Rescheduling_interrupts
17829 ± 99% -71.0% 5172 ± 16% interrupts.CPU30.RES:Rescheduling_interrupts
5569 ± 15% +2414.7% 140059 ± 94% interrupts.CPU31.RES:Rescheduling_interrupts
37674 ± 16% +36.6% 51473 ± 25% interrupts.CPU31.TLB:TLB_shootdowns
47905 ± 39% +76.6% 84583 ± 38% interrupts.CPU34.TLB:TLB_shootdowns
568.75 ±140% +224.8% 1847 ± 90% interrupts.CPU36.NMI:Non-maskable_interrupts
568.75 ±140% +224.8% 1847 ± 90% interrupts.CPU36.PMI:Performance_monitoring_interrupts
4236 ± 25% +2168.5% 96092 ± 90% interrupts.CPU36.RES:Rescheduling_interrupts
52717 ± 27% +43.3% 75565 ± 28% interrupts.CPU37.TLB:TLB_shootdowns
41418 ± 9% +136.6% 98010 ± 50% interrupts.CPU39.TLB:TLB_shootdowns
5551 ± 8% +847.8% 52615 ± 66% interrupts.CPU40.RES:Rescheduling_interrupts
4746 ± 25% +865.9% 45841 ± 91% interrupts.CPU42.RES:Rescheduling_interrupts
37556 ± 11% +24.6% 46808 ± 6% interrupts.CPU42.TLB:TLB_shootdowns
21846 ±124% -84.4% 3415 ± 46% interrupts.CPU48.RES:Rescheduling_interrupts
891.50 ± 22% -35.2% 577.25 ± 40% interrupts.CPU49.NMI:Non-maskable_interrupts
891.50 ± 22% -35.2% 577.25 ± 40% interrupts.CPU49.PMI:Performance_monitoring_interrupts
20459 ±120% -79.2% 4263 ± 14% interrupts.CPU49.RES:Rescheduling_interrupts
59840 ± 21% -23.1% 46042 ± 16% interrupts.CPU5.TLB:TLB_shootdowns
65200 ± 19% -34.5% 42678 ± 9% interrupts.CPU51.TLB:TLB_shootdowns
70923 ±153% -94.0% 4270 ± 29% interrupts.CPU53.RES:Rescheduling_interrupts
65312 ± 22% -28.7% 46578 ± 14% interrupts.CPU56.TLB:TLB_shootdowns
65828 ± 24% -33.4% 43846 ± 4% interrupts.CPU59.TLB:TLB_shootdowns
72558 ±156% -93.2% 4906 ± 9% interrupts.CPU6.RES:Rescheduling_interrupts
68698 ± 34% -32.6% 46327 ± 18% interrupts.CPU61.TLB:TLB_shootdowns
109745 ± 44% -57.4% 46711 ± 16% interrupts.CPU62.TLB:TLB_shootdowns
89714 ± 44% -48.5% 46198 ± 7% interrupts.CPU63.TLB:TLB_shootdowns
59380 ±136% -91.5% 5066 ± 13% interrupts.CPU69.RES:Rescheduling_interrupts
40094 ± 18% +133.9% 93798 ± 44% interrupts.CPU78.TLB:TLB_shootdowns
129884 ± 72% -55.3% 58034 ±157% interrupts.CPU8.RES:Rescheduling_interrupts
69984 ± 11% +51.4% 105957 ± 20% interrupts.CPU80.CAL:Function_call_interrupts
32857 ± 10% +128.7% 75131 ± 36% interrupts.CPU80.TLB:TLB_shootdowns
35726 ± 16% +34.6% 48081 ± 12% interrupts.CPU82.TLB:TLB_shootdowns
73820 ± 17% +28.2% 94643 ± 8% interrupts.CPU84.CAL:Function_call_interrupts
38829 ± 28% +190.3% 112736 ± 42% interrupts.CPU84.TLB:TLB_shootdowns
36129 ± 4% +47.6% 53329 ± 13% interrupts.CPU85.TLB:TLB_shootdowns
4693 ± 7% +1323.0% 66793 ±145% interrupts.CPU86.RES:Rescheduling_interrupts
38003 ± 11% +94.8% 74031 ± 43% interrupts.CPU86.TLB:TLB_shootdowns
78022 ± 3% +7.9% 84210 ± 3% interrupts.CPU87.CAL:Function_call_interrupts
36359 ± 6% +54.9% 56304 ± 48% interrupts.CPU88.TLB:TLB_shootdowns
89031 ±105% -95.0% 4475 ± 40% interrupts.CPU9.RES:Rescheduling_interrupts
40085 ± 11% +60.6% 64368 ± 27% interrupts.CPU91.TLB:TLB_shootdowns
42244 ± 10% +44.8% 61162 ± 35% interrupts.CPU94.TLB:TLB_shootdowns
40959 ± 15% +109.4% 85780 ± 41% interrupts.CPU95.TLB:TLB_shootdowns


stress-ng.fiemap.ops
80000 +-------------------------------------------------------------------+
75000 |..+. .+.. .+..+.. .+. .+.. |
| +..+..+..+.+. .+..+.. .+ +. +. +.+..+..+..+.+..|
70000 |-+ + +. |
65000 |-+ |
60000 |-+ |
55000 |-+ |
| |
50000 |-+ |
45000 |-+ |
40000 |-+ |
35000 |-+ O |
| O O O O O |
30000 |-+ O O O O O O O O O O O O O O O O O |
25000 +-------------------------------------------------------------------+
stress-ng.fiemap.ops_per_sec
80000 +-------------------------------------------------------------------+
75000 |.. .+.. .+.. |
| +. .+..+..+.+. .+..+.. .+.+. +..+.+..+..+.+..+..+..+.+..|
70000 |-+ +. + +. |
65000 |-+ |
60000 |-+ |
55000 |-+ |
| |
50000 |-+ |
45000 |-+ |
40000 |-+ |
35000 |-+ O |
| O O O O |
30000 |-+ O O O O O O O O O O O O O O O O |
25000 +-------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Rong Chen


_______________________________________________
LKP mailing list -- lkp@xxxxxxxxxxxx
To unsubscribe send an email to lkp-leave@xxxxxxxxxxxx


--
Zhengjun Xing