Re: [LKP] [ext4] d3b6f23f71: stress-ng.fiemap.ops_per_sec -60.5% regression

From: Ritesh Harjani
Date: Mon Apr 13 2020 - 06:57:10 EST




On 4/13/20 2:07 PM, Xing Zhengjun wrote:
Hi Harjani,

   Do you have time to take a look at this? Thanks.

Hello Xing,

I do want to look into this. But as of now I am stuck with another
mballoc failure issue. I will get back at this once I have some handle
over that one.

BTW, are you planning to take look at this?

-ritesh



On 4/7/2020 4:00 PM, kernel test robot wrote:
Greeting,

FYI, we noticed a -60.5% regression of stress-ng.fiemap.ops_per_sec due to commit:


commit: d3b6f23f71670007817a5d59f3fbafab2b794e8c ("ext4: move ext4_fiemap to use iomap framework")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: stress-ng
on test machine: 96 threads Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 192G memory
with following parameters:

    nr_threads: 10%
    disk: 1HDD
    testtime: 1s
    class: os
    cpufreq_governor: performance
    ucode: 0x500002c
    fs: ext4






Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

         git clone https://github.com/intel/lkp-tests.git
         cd lkp-tests
         bin/lkp install job.yaml  # job file is attached in this email
         bin/lkp run     job.yaml

=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode:
os/gcc-7/performance/1HDD/ext4/x86_64-rhel-7.6/10%/debian-x86_64-20191114.cgz/lkp-csl-2sp5/stress-ng/1s/0x500002c

commit:
   b2c5764262 ("ext4: make ext4_ind_map_blocks work with fiemap")
   d3b6f23f71 ("ext4: move ext4_fiemap to use iomap framework")

b2c5764262edded1 d3b6f23f71670007817a5d59f3f
---------------- ---------------------------
        fail:runs  %reproduction    fail:runs
            |             |             |
            :4           25%           1:4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
           2:4            5%           2:4 perf-profile.calltrace.cycles-pp.sync_regs.error_entry
           2:4            6%           3:4 perf-profile.calltrace.cycles-pp.error_entry
           3:4            9%           3:4 perf-profile.children.cycles-pp.error_entry
           0:4            1%           0:4 perf-profile.self.cycles-pp.error_entry
          %stddev     %change         %stddev
              \          |                \
      28623           +28.2%      36703 ± 12%  stress-ng.daemon.ops
      28632           +28.2%      36704 ± 12% stress-ng.daemon.ops_per_sec
     566.00 ± 22%     -53.2%     265.00 ± 53%  stress-ng.dev.ops
     278.81 ± 22%     -53.0%     131.00 ± 54%  stress-ng.dev.ops_per_sec
      73160           -60.6%      28849 ±  3%  stress-ng.fiemap.ops
      72471           -60.5%      28612 ±  3% stress-ng.fiemap.ops_per_sec
      23421 ± 12%     +21.2%      28388 ±  6%  stress-ng.filename.ops
      22638 ± 12%     +20.3%      27241 ±  6% stress-ng.filename.ops_per_sec
      21.25 ±  7%     -10.6%      19.00 ±  3%  stress-ng.iomix.ops
      38.75 ± 49%     -47.7%      20.25 ± 96%  stress-ng.memhotplug.ops
      34.45 ± 52%     -51.8%      16.62 ±106% stress-ng.memhotplug.ops_per_sec
       1734 ± 10%     +31.4%       2278 ± 10%  stress-ng.resources.ops
     807.56 ±  5%     +35.2%       1091 ±  8% stress-ng.resources.ops_per_sec
    1007356 ±  3%     -16.5%     840642 ±  9%  stress-ng.revio.ops
    1007692 ±  3%     -16.6%     840711 ±  9% stress-ng.revio.ops_per_sec
      21812 ±  3%     +16.0%      25294 ±  5%  stress-ng.sysbadaddr.ops
      21821 ±  3%     +15.9%      25294 ±  5% stress-ng.sysbadaddr.ops_per_sec
     440.75 ±  4%     +21.9%     537.25 ±  9%  stress-ng.sysfs.ops
     440.53 ±  4%     +21.9%     536.86 ±  9% stress-ng.sysfs.ops_per_sec
   13286582           -11.1%   11805520 ±  6% stress-ng.time.file_system_outputs
   68253896            +2.4%   69860122 stress-ng.time.minor_page_faults
     197.00 ±  4%     -15.9%     165.75 ± 12%  stress-ng.xattr.ops
     192.45 ±  5%     -16.1%     161.46 ± 11% stress-ng.xattr.ops_per_sec
      15310           +62.5%      24875 ± 22%  stress-ng.zombie.ops
      15310           +62.5%      24874 ± 22% stress-ng.zombie.ops_per_sec
     203.50 ± 12%     -47.3%     107.25 ± 49%  vmstat.io.bi
     861318 ± 18%     -29.7%     605884 ±  5%  meminfo.AnonHugePages
    1062742 ± 14%     -20.2%     847853 ±  3%  meminfo.AnonPages
      31093 ±  6%      +9.6%      34090 ±  3%  meminfo.KernelStack
       7151 ± 34%     +55.8%      11145 ±  9%  meminfo.Mlocked
  1.082e+08 ±  5%     -40.2%   64705429 ± 31% numa-numastat.node0.local_node
  1.082e+08 ±  5%     -40.2%   64739883 ± 31% numa-numastat.node0.numa_hit
   46032662 ± 21%    +104.3%   94042918 ± 20% numa-numastat.node1.local_node
   46074205 ± 21%    +104.2%   94072810 ± 20% numa-numastat.node1.numa_hit
       3942 ±  3%     +14.2%       4501 ±  4% slabinfo.pool_workqueue.active_objs
       4098 ±  3%     +14.3%       4683 ±  4% slabinfo.pool_workqueue.num_objs
       4817 ±  7%     +13.3%       5456 ±  8% slabinfo.proc_dir_entry.active_objs
       5153 ±  6%     +12.5%       5797 ±  8% slabinfo.proc_dir_entry.num_objs
      18598 ± 13%     -33.1%      12437 ± 20% sched_debug.cfs_rq:/.load.avg
     452595 ± 56%     -71.4%     129637 ± 76% sched_debug.cfs_rq:/.load.max
      67675 ± 35%     -55.1%      30377 ± 42% sched_debug.cfs_rq:/.load.stddev
      18114 ± 12%     -33.7%      12011 ± 20% sched_debug.cfs_rq:/.runnable_weight.avg
     448215 ± 58%     -72.8%     121789 ± 82% sched_debug.cfs_rq:/.runnable_weight.max
      67083 ± 37%     -56.3%      29305 ± 43% sched_debug.cfs_rq:/.runnable_weight.stddev
     -38032          +434.3%    -203212 sched_debug.cfs_rq:/.spread0.avg
    -204466           +95.8%    -400301 sched_debug.cfs_rq:/.spread0.min
      90.02 ± 25%     -58.1%      37.69 ± 52% sched_debug.cfs_rq:/.util_est_enqueued.avg
     677.54 ±  6%     -39.3%     411.50 ± 22% sched_debug.cfs_rq:/.util_est_enqueued.max
     196.57 ±  8%     -47.6%     103.05 ± 36% sched_debug.cfs_rq:/.util_est_enqueued.stddev
       3.34 ± 23%     +34.1%       4.48 ±  4% sched_debug.cpu.clock.stddev
       3.34 ± 23%     +34.1%       4.48 ±  4% sched_debug.cpu.clock_task.stddev
     402872 ±  7%     -11.9%     354819 ±  2%  proc-vmstat.nr_active_anon
    1730331            -9.5%    1566418 ±  5%  proc-vmstat.nr_dirtied
      31042 ±  6%      +9.3%      33915 ±  3% proc-vmstat.nr_kernel_stack
     229047            -2.4%     223615        proc-vmstat.nr_mapped
      74008 ±  7%     +20.5%      89163 ±  8%  proc-vmstat.nr_written
     402872 ±  7%     -11.9%     354819 ±  2% proc-vmstat.nr_zone_active_anon
      50587 ± 11%     -25.2%      37829 ± 14% proc-vmstat.numa_pages_migrated
     457500           -23.1%     351918 ± 31% proc-vmstat.numa_pte_updates
   81382485            +1.9%   82907822        proc-vmstat.pgfault
  2.885e+08 ±  5%     -13.3%  2.502e+08 ±  6%  proc-vmstat.pgfree
      42206 ± 12%     -46.9%      22399 ± 49%  proc-vmstat.pgpgin
     431233 ± 13%     -64.8%     151736 ±109%  proc-vmstat.pgrotated
     176754 ±  7%     -40.2%     105637 ± 31% proc-vmstat.thp_fault_alloc
     314.50 ± 82%    +341.5%       1388 ± 44% proc-vmstat.unevictable_pgs_stranded
    1075269 ± 14%     -41.3%     631388 ± 17%  numa-meminfo.node0.Active
     976056 ± 12%     -39.7%     588727 ± 19% numa-meminfo.node0.Active(anon)
     426857 ± 22%     -36.4%     271375 ± 13% numa-meminfo.node0.AnonHugePages
     558590 ± 19%     -36.4%     355402 ± 14% numa-meminfo.node0.AnonPages
    1794824 ±  9%     -28.8%    1277157 ± 20% numa-meminfo.node0.FilePages
       8517 ± 92%     -82.7%       1473 ± 89% numa-meminfo.node0.Inactive(file)
     633118 ±  2%     -41.7%     368920 ± 36%  numa-meminfo.node0.Mapped
    2958038 ± 12%     -27.7%    2139271 ± 12%  numa-meminfo.node0.MemUsed
     181401 ±  5%     -13.7%     156561 ±  4% numa-meminfo.node0.SUnreclaim
     258124 ±  6%     -13.0%     224535 ±  5%  numa-meminfo.node0.Slab
     702083 ± 16%     +31.0%     919406 ± 11%  numa-meminfo.node1.Active
      38663 ±107%    +137.8%      91951 ± 31% numa-meminfo.node1.Active(file)
    1154975 ±  7%     +41.6%    1635593 ± 12% numa-meminfo.node1.FilePages
     395813 ± 25%     +62.8%     644533 ± 16% numa-meminfo.node1.Inactive
     394313 ± 25%     +62.5%     640686 ± 16% numa-meminfo.node1.Inactive(anon)
     273317           +88.8%     515976 ± 25%  numa-meminfo.node1.Mapped
    2279237 ±  6%     +25.7%    2865582 ±  7%  numa-meminfo.node1.MemUsed
      10830 ± 18%     +29.6%      14033 ±  9% numa-meminfo.node1.PageTables
     149390 ±  3%     +23.2%     184085 ±  3% numa-meminfo.node1.SUnreclaim
     569542 ± 16%     +74.8%     995336 ± 21%  numa-meminfo.node1.Shmem
     220774 ±  5%     +20.3%     265656 ±  3%  numa-meminfo.node1.Slab
   35623587 ±  5%     -11.7%   31444514 ±  3%  perf-stat.i.cache-misses
  2.576e+08 ±  5%      -6.8%    2.4e+08 ±  2% perf-stat.i.cache-references
       3585            -7.3%       3323 ±  5%  perf-stat.i.cpu-migrations
     180139 ±  2%      +4.2%     187668        perf-stat.i.minor-faults
      69.13            +2.6       71.75 perf-stat.i.node-load-miss-rate%
    4313695 ±  2%      -7.4%    3994957 ±  2% perf-stat.i.node-load-misses
    5466253 ± 11%     -17.3%    4521173 ±  6%  perf-stat.i.node-loads
    2818674 ±  6%     -15.8%    2372542 ±  5%  perf-stat.i.node-stores
     227810            +4.6%     238290        perf-stat.i.page-faults
      12.67 ±  4%      -7.2%      11.76 ±  2%  perf-stat.overall.MPKI
       1.01 ±  4%      -0.0        0.97 ±  3% perf-stat.overall.branch-miss-rate%
       1044           +13.1%       1181 ±  4% perf-stat.overall.cycles-between-cache-misses
      40.37 ±  4%      +3.6       44.00 ±  2% perf-stat.overall.node-store-miss-rate%
   36139526 ±  5%     -12.5%   31625519 ±  3%  perf-stat.ps.cache-misses
  2.566e+08 ±  5%      -6.9%  2.389e+08 ±  2% perf-stat.ps.cache-references
       3562            -7.2%       3306 ±  5% perf-stat.ps.cpu-migrations
     179088            +4.2%     186579        perf-stat.ps.minor-faults
    4323383 ±  2%      -7.5%    3999214 perf-stat.ps.node-load-misses
    5607721 ± 10%     -18.5%    4568664 ±  6%  perf-stat.ps.node-loads
    2855134 ±  7%     -16.4%    2387345 ±  5%  perf-stat.ps.node-stores
     226270            +4.6%     236709        perf-stat.ps.page-faults
     242305 ± 10%     -42.4%     139551 ± 18% numa-vmstat.node0.nr_active_anon
     135983 ± 17%     -37.4%      85189 ± 10% numa-vmstat.node0.nr_anon_pages
     209.25 ± 16%     -38.1%     129.50 ± 10% numa-vmstat.node0.nr_anon_transparent_hugepages
     449367 ±  9%     -29.7%     315804 ± 20% numa-vmstat.node0.nr_file_pages
       2167 ± 90%     -80.6%     419.75 ± 98% numa-vmstat.node0.nr_inactive_file
     157405 ±  3%     -41.4%      92206 ± 35% numa-vmstat.node0.nr_mapped
       2022 ± 30%     -73.3%     539.25 ± 91%  numa-vmstat.node0.nr_mlock
       3336 ± 10%     -24.3%       2524 ± 25% numa-vmstat.node0.nr_page_table_pages
     286158 ± 10%     -41.2%     168337 ± 37%  numa-vmstat.node0.nr_shmem
      45493 ±  5%     -14.1%      39094 ±  4% numa-vmstat.node0.nr_slab_unreclaimable
     242294 ± 10%     -42.4%     139547 ± 18% numa-vmstat.node0.nr_zone_active_anon
       2167 ± 90%     -80.6%     419.75 ± 98% numa-vmstat.node0.nr_zone_inactive_file
   54053924 ±  8%     -39.3%   32786242 ± 34%  numa-vmstat.node0.numa_hit
   53929628 ±  8%     -39.5%   32619715 ± 34% numa-vmstat.node0.numa_local
       9701 ±107%    +136.9%      22985 ± 31% numa-vmstat.node1.nr_active_file
     202.50 ± 16%     -25.1%     151.75 ± 23% numa-vmstat.node1.nr_anon_transparent_hugepages
     284922 ±  7%     +43.3%     408195 ± 13% numa-vmstat.node1.nr_file_pages
      96002 ± 26%     +67.5%     160850 ± 17% numa-vmstat.node1.nr_inactive_anon
      68077 ±  2%     +90.3%     129533 ± 25% numa-vmstat.node1.nr_mapped
     138482 ± 15%     +79.2%     248100 ± 22%  numa-vmstat.node1.nr_shmem
      37396 ±  3%     +23.3%      46094 ±  3% numa-vmstat.node1.nr_slab_unreclaimable
       9701 ±107%    +136.9%      22985 ± 31% numa-vmstat.node1.nr_zone_active_file
      96005 ± 26%     +67.5%     160846 ± 17% numa-vmstat.node1.nr_zone_inactive_anon
   23343661 ± 17%     +99.9%   46664267 ± 23%  numa-vmstat.node1.numa_hit
   23248487 ± 17%    +100.5%   46610447 ± 23% numa-vmstat.node1.numa_local
     105745 ± 23%    +112.6%     224805 ± 24%  softirqs.CPU0.NET_RX
     133310 ± 36%     -45.3%      72987 ± 52%  softirqs.CPU1.NET_RX
     170110 ± 55%     -66.8%      56407 ±147%  softirqs.CPU11.NET_RX
      91465 ± 36%     -65.2%      31858 ±112%  softirqs.CPU13.NET_RX
     164491 ± 57%     -77.7%      36641 ±121%  softirqs.CPU15.NET_RX
     121069 ± 55%     -99.3%     816.75 ± 96%  softirqs.CPU17.NET_RX
      81019 ±  4%      -8.7%      73967 ±  4%  softirqs.CPU20.RCU
      72143 ± 63%     -89.8%       7360 ±172%  softirqs.CPU22.NET_RX
     270663 ± 17%     -57.9%     113915 ± 45%  softirqs.CPU24.NET_RX
      20149 ± 76%    +474.1%     115680 ± 62%  softirqs.CPU26.NET_RX
      14033 ± 70%    +977.5%     151211 ± 75%  softirqs.CPU27.NET_RX
      27834 ± 94%    +476.1%     160357 ± 28%  softirqs.CPU28.NET_RX
      35346 ± 68%    +212.0%     110290 ± 30%  softirqs.CPU29.NET_RX
      34347 ±103%    +336.5%     149941 ± 32%  softirqs.CPU32.NET_RX
      70077 ±  3%     +10.8%      77624 ±  3%  softirqs.CPU34.RCU
      36453 ± 84%    +339.6%     160253 ± 42%  softirqs.CPU36.NET_RX
      72367 ±  2%     +10.6%      80043        softirqs.CPU37.RCU
      25239 ±118%    +267.7%      92799 ± 45%  softirqs.CPU38.NET_RX
       4995 ±170%   +1155.8%      62734 ± 62%  softirqs.CPU39.NET_RX
       4641 ±145%   +1611.3%      79432 ± 90%  softirqs.CPU42.NET_RX
       7192 ± 65%    +918.0%      73225 ± 66%  softirqs.CPU45.NET_RX
       1772 ±166%   +1837.4%      34344 ± 63%  softirqs.CPU46.NET_RX
      13149 ± 81%    +874.7%     128170 ± 58%  softirqs.CPU47.NET_RX
      86484 ± 94%     -92.6%       6357 ±172%  softirqs.CPU48.NET_RX
     129128 ± 27%     -95.8%       5434 ±172%  softirqs.CPU55.NET_RX
      82772 ± 59%     -91.7%       6891 ±164%  softirqs.CPU56.NET_RX
     145313 ± 57%     -87.8%      17796 ± 88%  softirqs.CPU57.NET_RX
     118160 ± 33%     -86.3%      16226 ±109%  softirqs.CPU58.NET_RX
      94576 ± 56%     -94.1%       5557 ±173%  softirqs.CPU6.NET_RX
      82900 ± 77%     -66.8%      27508 ±171%  softirqs.CPU62.NET_RX
     157291 ± 30%     -81.1%      29656 ±111%  softirqs.CPU64.NET_RX
     135101 ± 28%     -80.2%      26748 ± 90%  softirqs.CPU67.NET_RX
     146574 ± 56%    -100.0%      69.75 ± 98%  softirqs.CPU68.NET_RX
      81347 ±  2%      -9.0%      74024 ±  2%  softirqs.CPU68.RCU
     201729 ± 37%     -99.6%     887.50 ±107%  softirqs.CPU69.NET_RX
     108454 ± 78%     -97.9%       2254 ±169%  softirqs.CPU70.NET_RX
      55289 ±104%     -89.3%       5942 ±172%  softirqs.CPU71.NET_RX
      10112 ±172%    +964.6%     107651 ± 89%  softirqs.CPU72.NET_RX
       3136 ±171%   +1522.2%      50879 ± 66%  softirqs.CPU73.NET_RX
      13353 ± 79%    +809.2%     121407 ±101%  softirqs.CPU74.NET_RX
      75194 ±  3%     +10.3%      82957 ±  5%  softirqs.CPU75.RCU
      11002 ±173%   +1040.8%     125512 ± 61%  softirqs.CPU76.NET_RX
       2463 ±173%   +2567.3%      65708 ± 77%  softirqs.CPU78.NET_RX
      25956 ±  3%      -7.8%      23932 ±  3%  softirqs.CPU78.SCHED
      16366 ±150%    +340.7%      72125 ± 91%  softirqs.CPU82.NET_RX
      14553 ±130%   +1513.4%     234809 ± 27%  softirqs.CPU93.NET_RX
      26314            -9.2%      23884 ±  3%  softirqs.CPU93.SCHED
       4582 ± 88%   +4903.4%     229268 ± 23%  softirqs.CPU94.NET_RX
      11214 ±111%   +1762.5%     208867 ± 18%  softirqs.CPU95.NET_RX
       1.53 ± 27%      -0.5        0.99 ± 17% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
       1.52 ± 27%      -0.5        0.99 ± 17% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
       1.39 ± 29%      -0.5        0.88 ± 21% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64
       1.39 ± 29%      -0.5        0.88 ± 21% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
       0.50 ± 59%      +0.3        0.81 ± 13% perf-profile.calltrace.cycles-pp.filemap_map_pages.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault
       5.70 ±  9%      +0.8        6.47 ±  7% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
       5.48 ±  9%      +0.8        6.27 ±  7% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal
       5.49 ±  9%      +0.8        6.28 ±  7% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal
       4.30 ±  4%      +1.3        5.60 ±  7% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode
       4.40 ±  4%      +1.3        5.69 ±  7% perf-profile.calltrace.cycles-pp.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
       4.37 ±  4%      +1.3        5.66 ±  7% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
       4.36 ±  4%      +1.3        5.66 ±  7% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
       4.33 ±  4%      +1.3        5.62 ±  7% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.prepare_exit_to_usermode.swapgs_restore_regs_and_return_to_usermode
       4.44 ±  4%      +1.3        5.74 ±  7% perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode
       3.20 ± 10%      -2.4        0.78 ±156% perf-profile.children.cycles-pp.copy_page
       0.16 ±  9%      -0.1        0.08 ± 64% perf-profile.children.cycles-pp.irq_work_interrupt
       0.16 ±  9%      -0.1        0.08 ± 64% perf-profile.children.cycles-pp.smp_irq_work_interrupt
       0.24 ±  5%      -0.1        0.17 ± 18% perf-profile.children.cycles-pp.irq_work_run_list
       0.16 ±  9%      -0.1        0.10 ± 24% perf-profile.children.cycles-pp.irq_work_run
       0.16 ±  9%      -0.1        0.10 ± 24% perf-profile.children.cycles-pp.printk
       0.23 ±  6%      -0.1        0.17 ±  9% perf-profile.children.cycles-pp.__do_execve_file
       0.08 ± 14%      -0.1        0.03 ±100% perf-profile.children.cycles-pp.delay_tsc
       0.16 ±  6%      -0.1        0.11 ±  9% perf-profile.children.cycles-pp.load_elf_binary
       0.16 ±  7%      -0.0        0.12 ± 13% perf-profile.children.cycles-pp.search_binary_handler
       0.20 ±  7%      -0.0        0.15 ± 10% perf-profile.children.cycles-pp.call_usermodehelper_exec_async
       0.19 ±  6%      -0.0        0.15 ± 11% perf-profile.children.cycles-pp.do_execve
       0.08 ± 10%      -0.0        0.04 ± 59% perf-profile.children.cycles-pp.__vunmap
       0.15 ±  3%      -0.0        0.11 ±  7% perf-profile.children.cycles-pp.rcu_idle_exit
       0.12 ± 10%      -0.0        0.09 ± 14% perf-profile.children.cycles-pp.__switch_to_asm
       0.09 ± 13%      -0.0        0.07 ±  5% perf-profile.children.cycles-pp.des3_ede_encrypt
       0.06 ± 11%      +0.0        0.09 ± 13% perf-profile.children.cycles-pp.mark_page_accessed
       0.15 ±  5%      +0.0        0.19 ± 12% perf-profile.children.cycles-pp.apparmor_cred_prepare
       0.22 ±  8%      +0.0        0.27 ± 11% perf-profile.children.cycles-pp.mem_cgroup_throttle_swaprate
       0.17 ±  2%      +0.0        0.22 ± 12% perf-profile.children.cycles-pp.security_prepare_creds
       0.95 ± 17%      +0.3        1.22 ± 14% perf-profile.children.cycles-pp.filemap_map_pages
       5.92 ±  8%      +0.7        6.65 ±  7% perf-profile.children.cycles-pp.get_signal
       5.66 ±  9%      +0.8        6.44 ±  7% perf-profile.children.cycles-pp.mmput
       5.65 ±  9%      +0.8        6.43 ±  7% perf-profile.children.cycles-pp.exit_mmap
       4.40 ±  4%      +1.3        5.70 ±  7% perf-profile.children.cycles-pp.prepare_exit_to_usermode
       4.45 ±  4%      +1.3        5.75 ±  7% perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
       3.16 ± 10%      -2.4        0.77 ±155% perf-profile.self.cycles-pp.copy_page
       0.08 ± 14%      -0.1        0.03 ±100% perf-profile.self.cycles-pp.delay_tsc
       0.12 ± 10%      -0.0        0.09 ± 14% perf-profile.self.cycles-pp.__switch_to_asm
       0.08 ± 12%      -0.0        0.06 ± 17% perf-profile.self.cycles-pp.enqueue_task_fair
       0.09 ± 13%      -0.0        0.07 ±  5% perf-profile.self.cycles-pp.des3_ede_encrypt
       0.07 ± 13%      +0.0        0.08 ± 19% perf-profile.self.cycles-pp.__lru_cache_add
       0.19 ±  9%      +0.0        0.22 ± 10% perf-profile.self.cycles-pp.mem_cgroup_throttle_swaprate
       0.15 ±  5%      +0.0        0.19 ± 11% perf-profile.self.cycles-pp.apparmor_cred_prepare
       0.05 ± 58%      +0.0        0.09 ± 13% perf-profile.self.cycles-pp.mark_page_accessed
       0.58 ± 10%      +0.2        0.80 ± 20% perf-profile.self.cycles-pp.release_pages
       0.75 ±173%  +1.3e+05%       1005 ±100% interrupts.127:PCI-MSI.31981660-edge.i40e-eth0-TxRx-91
     820.75 ±111%     -99.9%       0.50 ±173% interrupts.47:PCI-MSI.31981580-edge.i40e-eth0-TxRx-11
     449.25 ± 86%    -100.0%       0.00 interrupts.53:PCI-MSI.31981586-edge.i40e-eth0-TxRx-17
      33.25 ±157%    -100.0%       0.00 interrupts.57:PCI-MSI.31981590-edge.i40e-eth0-TxRx-21
       0.75 ±110%  +63533.3%     477.25 ±162% interrupts.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
     561.50 ±160%    -100.0%       0.00 interrupts.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
      82921 ±  8%     -11.1%      73748 ±  6% interrupts.CPU11.CAL:Function_call_interrupts
      66509 ± 30%     -32.6%      44828 ±  8% interrupts.CPU14.TLB:TLB_shootdowns
      43105 ± 98%     -90.3%       4183 ± 21% interrupts.CPU17.RES:Rescheduling_interrupts
     148719 ± 70%     -69.4%      45471 ± 16% interrupts.CPU17.TLB:TLB_shootdowns
      85589 ± 42%     -52.2%      40884 ±  5% interrupts.CPU20.TLB:TLB_shootdowns
     222472 ± 41%     -98.0%       4360 ± 45% interrupts.CPU22.RES:Rescheduling_interrupts
       0.50 ±173%  +95350.0%     477.25 ±162% interrupts.CPU25.61:PCI-MSI.31981594-edge.i40e-eth0-TxRx-25
      76029 ± 10%     +14.9%      87389 ±  5% interrupts.CPU25.CAL:Function_call_interrupts
     399042 ±  6%     +13.4%     452479 ±  8% interrupts.CPU27.LOC:Local_timer_interrupts
     561.00 ±161%    -100.0%       0.00 interrupts.CPU29.65:PCI-MSI.31981598-edge.i40e-eth0-TxRx-29
       7034 ± 46%   +1083.8%      83279 ±138% interrupts.CPU29.RES:Rescheduling_interrupts
      17829 ± 99%     -71.0%       5172 ± 16% interrupts.CPU30.RES:Rescheduling_interrupts
       5569 ± 15%   +2414.7%     140059 ± 94% interrupts.CPU31.RES:Rescheduling_interrupts
      37674 ± 16%     +36.6%      51473 ± 25% interrupts.CPU31.TLB:TLB_shootdowns
      47905 ± 39%     +76.6%      84583 ± 38% interrupts.CPU34.TLB:TLB_shootdowns
     568.75 ±140%    +224.8%       1847 ± 90% interrupts.CPU36.NMI:Non-maskable_interrupts
     568.75 ±140%    +224.8%       1847 ± 90% interrupts.CPU36.PMI:Performance_monitoring_interrupts
       4236 ± 25%   +2168.5%      96092 ± 90% interrupts.CPU36.RES:Rescheduling_interrupts
      52717 ± 27%     +43.3%      75565 ± 28% interrupts.CPU37.TLB:TLB_shootdowns
      41418 ±  9%    +136.6%      98010 ± 50% interrupts.CPU39.TLB:TLB_shootdowns
       5551 ±  8%    +847.8%      52615 ± 66% interrupts.CPU40.RES:Rescheduling_interrupts
       4746 ± 25%    +865.9%      45841 ± 91% interrupts.CPU42.RES:Rescheduling_interrupts
      37556 ± 11%     +24.6%      46808 ±  6% interrupts.CPU42.TLB:TLB_shootdowns
      21846 ±124%     -84.4%       3415 ± 46% interrupts.CPU48.RES:Rescheduling_interrupts
     891.50 ± 22%     -35.2%     577.25 ± 40% interrupts.CPU49.NMI:Non-maskable_interrupts
     891.50 ± 22%     -35.2%     577.25 ± 40% interrupts.CPU49.PMI:Performance_monitoring_interrupts
      20459 ±120%     -79.2%       4263 ± 14% interrupts.CPU49.RES:Rescheduling_interrupts
      59840 ± 21%     -23.1%      46042 ± 16% interrupts.CPU5.TLB:TLB_shootdowns
      65200 ± 19%     -34.5%      42678 ±  9% interrupts.CPU51.TLB:TLB_shootdowns
      70923 ±153%     -94.0%       4270 ± 29% interrupts.CPU53.RES:Rescheduling_interrupts
      65312 ± 22%     -28.7%      46578 ± 14% interrupts.CPU56.TLB:TLB_shootdowns
      65828 ± 24%     -33.4%      43846 ±  4% interrupts.CPU59.TLB:TLB_shootdowns
      72558 ±156%     -93.2%       4906 ±  9% interrupts.CPU6.RES:Rescheduling_interrupts
      68698 ± 34%     -32.6%      46327 ± 18% interrupts.CPU61.TLB:TLB_shootdowns
     109745 ± 44%     -57.4%      46711 ± 16% interrupts.CPU62.TLB:TLB_shootdowns
      89714 ± 44%     -48.5%      46198 ±  7% interrupts.CPU63.TLB:TLB_shootdowns
      59380 ±136%     -91.5%       5066 ± 13% interrupts.CPU69.RES:Rescheduling_interrupts
      40094 ± 18%    +133.9%      93798 ± 44% interrupts.CPU78.TLB:TLB_shootdowns
     129884 ± 72%     -55.3%      58034 ±157% interrupts.CPU8.RES:Rescheduling_interrupts
      69984 ± 11%     +51.4%     105957 ± 20% interrupts.CPU80.CAL:Function_call_interrupts
      32857 ± 10%    +128.7%      75131 ± 36% interrupts.CPU80.TLB:TLB_shootdowns
      35726 ± 16%     +34.6%      48081 ± 12% interrupts.CPU82.TLB:TLB_shootdowns
      73820 ± 17%     +28.2%      94643 ±  8% interrupts.CPU84.CAL:Function_call_interrupts
      38829 ± 28%    +190.3%     112736 ± 42% interrupts.CPU84.TLB:TLB_shootdowns
      36129 ±  4%     +47.6%      53329 ± 13% interrupts.CPU85.TLB:TLB_shootdowns
       4693 ±  7%   +1323.0%      66793 ±145% interrupts.CPU86.RES:Rescheduling_interrupts
      38003 ± 11%     +94.8%      74031 ± 43% interrupts.CPU86.TLB:TLB_shootdowns
      78022 ±  3%      +7.9%      84210 ±  3% interrupts.CPU87.CAL:Function_call_interrupts
      36359 ±  6%     +54.9%      56304 ± 48% interrupts.CPU88.TLB:TLB_shootdowns
      89031 ±105%     -95.0%       4475 ± 40% interrupts.CPU9.RES:Rescheduling_interrupts
      40085 ± 11%     +60.6%      64368 ± 27% interrupts.CPU91.TLB:TLB_shootdowns
      42244 ± 10%     +44.8%      61162 ± 35% interrupts.CPU94.TLB:TLB_shootdowns
      40959 ± 15%    +109.4%      85780 ± 41% interrupts.CPU95.TLB:TLB_shootdowns


                                 stress-ng.fiemap.ops
   80000 +-------------------------------------------------------------------+
   75000 |..+.             .+..            .+..+..  .+. .+..                |
         |    +..+..+..+.+.     .+..+..  .+       +.   +. +.+..+..+..+.+..|
   70000 |-+                   + +.                                    |
   65000 |-+                                                                 |
   60000 |-+                                                                 |
   55000 |-+                                                                 |
|                                                                   |
   50000 |-+                                                                 |
   45000 |-+                                                                 |
   40000 |-+                                                                 |
   35000 |-+ O                                                |
         |  O       O                       O     O O                     |
   30000 |-+  O  O     O O     O O     O  O    O     O    O  O O  O O  O O  |
   25000 +-------------------------------------------------------------------+
                             stress-ng.fiemap.ops_per_sec
   80000 +-------------------------------------------------------------------+
   75000 |..               .+.. .+..                           |
         |  +.  .+..+..+.+.     .+..+..  .+.+. +..+.+..+..+.+..+..+..+.+..|
   70000 |-+  +.               + +.                                    |
   65000 |-+                                                                 |
   60000 |-+                                                                 |
   55000 |-+                                                                 |
|                                                                   |
   50000 |-+                                                                 |
   45000 |-+                                                                 |
   40000 |-+                                                                 |
   35000 |-+ O                                                |
         |  O       O                       O O                          |
   30000 |-+  O  O     O O     O O     O       O     O O  O  O    O O  O O  |
   25000 +-------------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad  sample



Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Rong Chen


_______________________________________________
LKP mailing list -- lkp@xxxxxxxxxxxx
To unsubscribe send an email to lkp-leave@xxxxxxxxxxxx