<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:49:20 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-12062] sanity-benchmark test iozone hangs with &#8220;too many service threads, or there were not enough hardware resources&#8221;</title>
                <link>https://jira.whamcloud.com/browse/LU-12062</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;sanity-benchmark test_iozone hangs. So far, we&#8217;re only seeing this hang for ZFS. &lt;/p&gt;

&lt;p&gt;Looking at the suite_log for the failure at &lt;a href=&quot;https://testing.whamcloud.com/test_sets/464d518e-43e5-11e9-a256-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/464d518e-43e5-11e9-a256-52540065bddc&lt;/a&gt;, we can see iozone hangs in DirectIO testing&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== sanity-benchmark test iozone: iozone ============================================================== 15:04:32 (1552143872)
min OST has 1922432kB available, using 3438744kB file size
debug=0
running as uid/gid/euid/egid 500/500/500/500, groups:
 [touch] [/mnt/lustre/d0_runas_test/f27932]
running as uid/gid/euid/egid 500/500/500/500, groups:
 [iozone] [-i] [0] [-i] [1] [-i] [2] [-e] [-+d] [-r] [512] [-s] [3438744] [-f] [/mnt/lustre/d0.iozone/iozone]
	Iozone: Performance Test of File I/O
	        Version $Revision: 3.373 $
		Compiled for 64 bit mode.
		Build: linux-AMD64 

	Contributors:William Norcott, Don Capps, Isom Crawford, Kirby Collins
	             Al Slater, Scott Rhine, Mike Wisner, Ken Goss
	             Steve Landherr, Brad Smith, Mark Kelly, Dr. Alain CYR,
	             Randy Dunlap, Mark Montague, Dan Million, Gavin Brebner,
	             Jean-Marc Zucconi, Jeff Blomberg, Benny Halevy, Dave Boone,
	             Erik Habbinga, Kris Strecker, Walter Wong, Joshua Root,
	             Fabrice Bacchella, Zhenghua Xue, Qin Li, Darren Sawyer.

	Run began: Sat Mar  9 15:04:38 2019

	Include fsync in write timing
	&amp;gt;&amp;gt;&amp;gt; I/O Diagnostic mode enabled. &amp;lt;&amp;lt;&amp;lt;
	Performance measurements are invalid in this mode.
	Record Size 512 KB
	File size set to 3438744 KB
	Command line used: iozone -i 0 -i 1 -i 2 -e -+d -r 512 -s 3438744 -f /mnt/lustre/d0.iozone/iozone
	Output is in Kbytes/sec
	Time Resolution = 0.000001 seconds.
	Processor cache size set to 1024 Kbytes.
	Processor cache line size set to 32 bytes.
	File stride size set to 17 * record size.
                                                            random  random    bkwd   record   stride                                   
              KB  reclen   write rewrite    read    reread    read   write    read  rewrite     read   fwrite frewrite   fread  freread
         3438744     512    2826    2713    32747    35474   19214    3473                                                          

iozone test complete.
debug=0x33f0484
directio on /mnt/lustre/f.iozone for 1x4194304 bytes 
PASS
debug=0
running as uid/gid/euid/egid 500/500/500/500, groups:
 [iozone] [-I] [-i] [0] [-i] [1] [-i] [2] [-e] [-+d] [-r] [512] [-s] [3438744] [-f] [/mnt/lustre/d0.iozone/iozone.odir]
	Iozone: Performance Test of File I/O
	        Version $Revision: 3.373 $
		Compiled for 64 bit mode.
		Build: linux-AMD64 

	Contributors:William Norcott, Don Capps, Isom Crawford, Kirby Collins
	             Al Slater, Scott Rhine, Mike Wisner, Ken Goss
	             Steve Landherr, Brad Smith, Mark Kelly, Dr. Alain CYR,
	             Randy Dunlap, Mark Montague, Dan Million, Gavin Brebner,
	             Jean-Marc Zucconi, Jeff Blomberg, Benny Halevy, Dave Boone,
	             Erik Habbinga, Kris Strecker, Walter Wong, Joshua Root,
	             Fabrice Bacchella, Zhenghua Xue, Qin Li, Darren Sawyer.

	Run began: Sat Mar  9 16:08:59 2019

	O_DIRECT feature enabled
	Include fsync in write timing
	&amp;gt;&amp;gt;&amp;gt; I/O Diagnostic mode enabled. &amp;lt;&amp;lt;&amp;lt;
	Performance measurements are invalid in this mode.
	Record Size 512 KB
	File size set to 3438744 KB
	Command line used: iozone -I -i 0 -i 1 -i 2 -e -+d -r 512 -s 3438744 -f /mnt/lustre/d0.iozone/iozone.odir
	Output is in Kbytes/sec
	Time Resolution = 0.000001 seconds.
	Processor cache size set to 1024 Kbytes.
	Processor cache line size set to 32 bytes.
	File stride size set to 17 * record size.
                                                            random  random    bkwd   record   stride                                   
              KB  reclen   write rewrite    read    reread    read   write    read  rewrite     read   fwrite frewrite   fread  freread
         3438744     512
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Looking at the console logs, there&#8217;s not a lot to go on. On the OSS (vm3) we do see an LNET too many service threads warning before the hang and looking through the call stacks, there are some processes in the D state&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[22345.851895] Lustre: DEBUG MARKER: == sanity-benchmark test iozone: iozone ============================================================== 15:04:32 (1552143872)
[22349.818428] Lustre: DEBUG MARKER: /usr/sbin/lctl mark min OST has 1922432kB available, using 3438744kB file size
[22350.076310] Lustre: DEBUG MARKER: min OST has 1922432kB available, using 3438744kB file size
[22402.060402] LNet: Service thread pid 32137 completed after 41.38s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
[22402.065241] LNet: Skipped 7 previous similar messages
[22410.807172] LNet: Service thread pid 1390 completed after 50.25s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
[22410.812202] LNet: Skipped 4 previous similar messages
[22431.196168] LNet: Service thread pid 10091 completed after 70.64s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
[22431.199270] LNet: Skipped 21 previous similar messages
[23561.656188] LNet: Service thread pid 19365 completed after 50.37s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
[23561.659104] LNet: Skipped 23 previous similar messages
[23566.276500] LNet: Service thread pid 1643 completed after 54.99s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
[23566.280050] LNet: Skipped 1 previous similar message
[24830.125602] LNet: Service thread pid 1385 completed after 44.33s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
[24830.129505] LNet: Skipped 4 previous similar messages
[24832.949732] LNet: Service thread pid 23024 completed after 47.15s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
[24836.845387] LNet: Service thread pid 30496 completed after 51.05s. This indicates the system was overloaded (too many service threads, or there were not enough hardware resources).
&#8230;
[36836.448221] jbd2/vda1-8     D ffff8c47b92ac100     0  1197      2 0x00000000
[36836.449552] Call Trace:
[36836.450017]  [&amp;lt;ffffffffa2d65d60&amp;gt;] ? bit_wait+0x50/0x50
[36836.450882]  [&amp;lt;ffffffffa2d67c49&amp;gt;] schedule+0x29/0x70
[36836.451724]  [&amp;lt;ffffffffa2d65721&amp;gt;] schedule_timeout+0x221/0x2d0
[36836.452739]  [&amp;lt;ffffffffa266a0ce&amp;gt;] ? kvm_clock_get_cycles+0x1e/0x20
[36836.453789]  [&amp;lt;ffffffffa2d65d60&amp;gt;] ? bit_wait+0x50/0x50
[36836.454659]  [&amp;lt;ffffffffa2d672ed&amp;gt;] io_schedule_timeout+0xad/0x130
[36836.455683]  [&amp;lt;ffffffffa2d67388&amp;gt;] io_schedule+0x18/0x20
[36836.456574]  [&amp;lt;ffffffffa2d65d71&amp;gt;] bit_wait_io+0x11/0x50
[36836.457463]  [&amp;lt;ffffffffa2d65897&amp;gt;] __wait_on_bit+0x67/0x90
[36836.458367]  [&amp;lt;ffffffffa2d65d60&amp;gt;] ? bit_wait+0x50/0x50
[36836.459228]  [&amp;lt;ffffffffa2d65a01&amp;gt;] out_of_line_wait_on_bit+0x81/0xb0
[36836.460283]  [&amp;lt;ffffffffa26c2dc0&amp;gt;] ? wake_bit_function+0x40/0x40
[36836.461282]  [&amp;lt;ffffffffa2878f1a&amp;gt;] __wait_on_buffer+0x2a/0x30
[36836.462460]  [&amp;lt;ffffffffc01e4801&amp;gt;] jbd2_journal_commit_transaction+0x1781/0x19b0 [jbd2]
[36836.463787]  [&amp;lt;ffffffffa26d0880&amp;gt;] ? finish_task_switch+0x50/0x1c0
[36836.464819]  [&amp;lt;ffffffffc01e9e89&amp;gt;] kjournald2+0xc9/0x260 [jbd2]
[36836.465805]  [&amp;lt;ffffffffa26c2d00&amp;gt;] ? wake_up_atomic_t+0x30/0x30
[36836.466785]  [&amp;lt;ffffffffc01e9dc0&amp;gt;] ? commit_timeout+0x10/0x10 [jbd2]
[36836.467836]  [&amp;lt;ffffffffa26c1c31&amp;gt;] kthread+0xd1/0xe0
[36836.468658]  [&amp;lt;ffffffffa26c1b60&amp;gt;] ? insert_kthread_work+0x40/0x40
[36836.469680]  [&amp;lt;ffffffffa2d74c37&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[36836.470778]  [&amp;lt;ffffffffa26c1b60&amp;gt;] ? insert_kthread_work+0x40/0x40
&#8230;
[36836.877778] auditd          D ffff8c47b9094100     0  3085      1 0x00000000
[36836.879028] Call Trace:
[36836.879461]  [&amp;lt;ffffffffa2d67c49&amp;gt;] schedule+0x29/0x70
[36836.880296]  [&amp;lt;ffffffffc01e97c5&amp;gt;] jbd2_log_wait_commit+0xc5/0x140 [jbd2]
[36836.881405]  [&amp;lt;ffffffffa26c2d00&amp;gt;] ? wake_up_atomic_t+0x30/0x30
[36836.882391]  [&amp;lt;ffffffffc01eae52&amp;gt;] jbd2_complete_transaction+0x52/0xa0 [jbd2]
[36836.883626]  [&amp;lt;ffffffffc0213782&amp;gt;] ext4_sync_file+0x292/0x320 [ext4]
[36836.884681]  [&amp;lt;ffffffffa2875ef7&amp;gt;] do_fsync+0x67/0xb0
[36836.885533]  [&amp;lt;ffffffffa2d74d15&amp;gt;] ? system_call_after_swapgs+0xa2/0x146
[36836.886642]  [&amp;lt;ffffffffa28761e0&amp;gt;] SyS_fsync+0x10/0x20
[36836.887505]  [&amp;lt;ffffffffa2d74ddb&amp;gt;] system_call_fastpath+0x22/0x27
[36836.888528]  [&amp;lt;ffffffffa2d74d21&amp;gt;] ? system_call_after_swapgs+0xae/0x146
&#8230;
[36837.709648] in:imjournal    D ffff8c474007e180     0  3675      1 0x00000080
[36837.711149] Call Trace:
[36837.711590]  [&amp;lt;ffffffffa2d65d60&amp;gt;] ? bit_wait+0x50/0x50
[36837.712459]  [&amp;lt;ffffffffa2d67c49&amp;gt;] schedule+0x29/0x70
[36837.713287]  [&amp;lt;ffffffffa2d65721&amp;gt;] schedule_timeout+0x221/0x2d0
[36837.714256]  [&amp;lt;ffffffffa266a0ce&amp;gt;] ? kvm_clock_get_cycles+0x1e/0x20
[36837.715294]  [&amp;lt;ffffffffa2701052&amp;gt;] ? ktime_get_ts64+0x52/0xf0
[36837.716247]  [&amp;lt;ffffffffa2d65d60&amp;gt;] ? bit_wait+0x50/0x50
[36837.717120]  [&amp;lt;ffffffffa2d672ed&amp;gt;] io_schedule_timeout+0xad/0x130
[36837.718128]  [&amp;lt;ffffffffa2d67388&amp;gt;] io_schedule+0x18/0x20
[36837.719016]  [&amp;lt;ffffffffa2d65d71&amp;gt;] bit_wait_io+0x11/0x50
[36837.719907]  [&amp;lt;ffffffffa2d65897&amp;gt;] __wait_on_bit+0x67/0x90
[36837.720823]  [&amp;lt;ffffffffa27b92be&amp;gt;] ? __find_get_pages+0x11e/0x1c0
[36837.721834]  [&amp;lt;ffffffffa27b5761&amp;gt;] wait_on_page_bit+0x81/0xa0
[36837.722787]  [&amp;lt;ffffffffa26c2dc0&amp;gt;] ? wake_bit_function+0x40/0x40
[36837.723784]  [&amp;lt;ffffffffa27c6f4b&amp;gt;] truncate_inode_pages_range+0x42b/0x700
[36837.724937]  [&amp;lt;ffffffffc024bdbc&amp;gt;] ? __ext4_journal_stop+0x3c/0xb0 [ext4]
[36837.726062]  [&amp;lt;ffffffffc0227d58&amp;gt;] ? ext4_rename+0x168/0x890 [ext4]
[36837.727101]  [&amp;lt;ffffffffa2850d51&amp;gt;] ? link_path_walk+0x81/0x8b0
[36837.728084]  [&amp;lt;ffffffffa28aee8a&amp;gt;] ? __dquot_initialize+0x3a/0x240
[36837.729106]  [&amp;lt;ffffffffa286fa5a&amp;gt;] ? __inode_wait_for_writeback+0x7a/0xf0
[36837.730234]  [&amp;lt;ffffffffa27c728f&amp;gt;] truncate_inode_pages_final+0x4f/0x60
[36837.731330]  [&amp;lt;ffffffffc021e41f&amp;gt;] ext4_evict_inode+0x10f/0x480 [ext4]
[36837.732411]  [&amp;lt;ffffffffa285eab4&amp;gt;] evict+0xb4/0x180
[36837.733242]  [&amp;lt;ffffffffa285f3bc&amp;gt;] iput+0xfc/0x190
[36837.734048]  [&amp;lt;ffffffffa2859c20&amp;gt;] __dentry_kill+0x120/0x180
[36837.734987]  [&amp;lt;ffffffffa2859d30&amp;gt;] dput+0xb0/0x160
[36837.735793]  [&amp;lt;ffffffffa2853b58&amp;gt;] SYSC_renameat2+0x518/0x5a0
[36837.736751]  [&amp;lt;ffffffffa27ef841&amp;gt;] ? __vma_rb_erase+0x121/0x220
[36837.737731]  [&amp;lt;ffffffffa2d74d21&amp;gt;] ? system_call_after_swapgs+0xae/0x146
[36837.738835]  [&amp;lt;ffffffffa2d74d15&amp;gt;] ? system_call_after_swapgs+0xa2/0x146
[36837.739945]  [&amp;lt;ffffffffa2d74d21&amp;gt;] ? system_call_after_swapgs+0xae/0x146
[36837.741050]  [&amp;lt;ffffffffa2d74d15&amp;gt;] ? system_call_after_swapgs+0xa2/0x146
[36837.742152]  [&amp;lt;ffffffffa2854a5e&amp;gt;] SyS_renameat2+0xe/0x10
[36837.743078]  [&amp;lt;ffffffffa2854a9e&amp;gt;] SyS_rename+0x1e/0x20
[36837.743951]  [&amp;lt;ffffffffa2d74ddb&amp;gt;] system_call_fastpath+0x22/0x27
[36837.744961]  [&amp;lt;ffffffffa2d74d21&amp;gt;] ? system_call_after_swapgs+0xae/0x146
&#8230;
[36842.351359] txg_sync        D ffff8c478a3d5140     0 11265      2 0x00000080
[36842.352621] Call Trace:
[36842.353049]  [&amp;lt;ffffffffa287e565&amp;gt;] ? bio_alloc_bioset+0x115/0x310
[36842.354043]  [&amp;lt;ffffffffa2d67c49&amp;gt;] schedule+0x29/0x70
[36842.354892]  [&amp;lt;ffffffffa2d65721&amp;gt;] schedule_timeout+0x221/0x2d0
[36842.355886]  [&amp;lt;ffffffffa266a0ce&amp;gt;] ? kvm_clock_get_cycles+0x1e/0x20
[36842.356935]  [&amp;lt;ffffffffa2701052&amp;gt;] ? ktime_get_ts64+0x52/0xf0
[36842.357893]  [&amp;lt;ffffffffa2d672ed&amp;gt;] io_schedule_timeout+0xad/0x130
[36842.358912]  [&amp;lt;ffffffffa26c2886&amp;gt;] ? prepare_to_wait_exclusive+0x56/0x90
[36842.360046]  [&amp;lt;ffffffffa2d67388&amp;gt;] io_schedule+0x18/0x20
[36842.360935]  [&amp;lt;ffffffffc0367242&amp;gt;] cv_wait_common+0xb2/0x150 [spl]
[36842.361962]  [&amp;lt;ffffffffa26c2d00&amp;gt;] ? wake_up_atomic_t+0x30/0x30
[36842.362944]  [&amp;lt;ffffffffc0367318&amp;gt;] __cv_wait_io+0x18/0x20 [spl]
[36842.363963]  [&amp;lt;ffffffffc053b073&amp;gt;] zio_wait+0x113/0x1c0 [zfs]
[36842.364943]  [&amp;lt;ffffffffc04efbf1&amp;gt;] vdev_config_sync+0xf1/0x180 [zfs]
[36842.366027]  [&amp;lt;ffffffffc04cfa9c&amp;gt;] spa_sync+0xa1c/0xd90 [zfs]
[36842.366985]  [&amp;lt;ffffffffa26d67c2&amp;gt;] ? default_wake_function+0x12/0x20
[36842.368064]  [&amp;lt;ffffffffc04e3c71&amp;gt;] txg_sync_thread+0x301/0x510 [zfs]
[36842.369139]  [&amp;lt;ffffffffc04e3970&amp;gt;] ? txg_fini+0x2a0/0x2a0 [zfs]
[36842.370141]  [&amp;lt;ffffffffc0362063&amp;gt;] thread_generic_wrapper+0x73/0x80 [spl]
[36842.371269]  [&amp;lt;ffffffffc0361ff0&amp;gt;] ? __thread_exit+0x20/0x20 [spl]
[36842.372345]  [&amp;lt;ffffffffa26c1c31&amp;gt;] kthread+0xd1/0xe0
[36842.373182]  [&amp;lt;ffffffffa26c1b60&amp;gt;] ? insert_kthread_work+0x40/0x40
[36842.374210]  [&amp;lt;ffffffffa2d74c37&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[36842.375312]  [&amp;lt;ffffffffa26c1b60&amp;gt;] ? insert_kthread_work+0x40/0x40
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Could this be &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8215&quot; title=&quot;sanity-benchmark/iozone should wait for space recovery&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8215&quot;&gt;&lt;del&gt;LU-8215&lt;/del&gt;&lt;/a&gt; ?&lt;/p&gt;

&lt;p&gt;Other sanity-benchmark test iozone that fail in this way&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/8d4b2322-4269-11e9-8e92-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/8d4b2322-4269-11e9-8e92-52540065bddc&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/c6afa400-4280-11e9-92fe-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/c6afa400-4280-11e9-92fe-52540065bddc&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/c681c26c-40a8-11e9-8e92-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/c681c26c-40a8-11e9-8e92-52540065bddc&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/2e66b0c6-40aa-11e9-b98a-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/2e66b0c6-40aa-11e9-b98a-52540065bddc&lt;/a&gt;&lt;/p&gt;</description>
                <environment>ZFS</environment>
        <key id="55142">LU-12062</key>
            <summary>sanity-benchmark test iozone hangs with &#8220;too many service threads, or there were not enough hardware resources&#8221;</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="jamesanunez">James Nunez</reporter>
                        <labels>
                            <label>ZFS</label>
                    </labels>
                <created>Tue, 12 Mar 2019 20:23:59 +0000</created>
                <updated>Mon, 29 Jan 2024 19:25:54 +0000</updated>
                            <resolved>Mon, 29 Jan 2024 19:25:54 +0000</resolved>
                                    <version>Lustre 2.10.7</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="243942" author="adilger" created="Thu, 14 Mar 2019 18:05:36 +0000"  >&lt;p&gt;Since this is not running on real hardware, it seems possible that the high load is just overloading the VM guest/host due to a lot of IO. This is made worse by ZFS because (AFAIK) the host only has a single virtual disk for each guest, but ZFS expects to store redundant metadata copies, so it is writing 2x copies of all metadata to different parts of the &quot;disk&quot; and executing a &lt;b&gt;lot&lt;/b&gt; of sync operations to do transaction commits. &lt;/p&gt;

&lt;p&gt;Do we have enough RAM on the host to be able to store the host virtual disks directly in /tmp using tmpfs rather than on a real HDD?  If the HDD used today for the virtual disks was made into a large swap device then the guest devices in /tmp can be spilled into swap if needed, but will mostly be served from RAM and swap is no slower than 7 OSTs and 4 MDTs all writing to a single HDD on the host. &lt;/p&gt;

&lt;p&gt;We don&apos;t really care about the persistence of the guest filesystems if the host crashes, so there shouldn&apos;t be a problem. Even for &quot;hard failover&quot; tests, the virtual devices in /tmp would be available to all guests on the same host, the same as with a local HDD. &lt;/p&gt;</comment>
                            <comment id="319707" author="JIRAUSER17102" created="Tue, 30 Nov 2021 22:20:07 +0000"  >&lt;p&gt;Might have happened here: &lt;a href=&quot;https://testing.whamcloud.com/test_sets/e89234ef-9630-40e4-abf4-feca3da9762f&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/e89234ef-9630-40e4-abf4-feca3da9762f&lt;/a&gt;&#160;&lt;/p&gt;

&lt;p&gt;2.12.8 sanity-benchmark test iozone on ZFS hangs under IO:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
...
        Run began: Sat Nov 20 05:00:05 2021

	O_DIRECT feature enabled
	Include fsync in write timing
	&amp;gt;&amp;gt;&amp;gt; I/O Diagnostic mode enabled. &amp;lt;&amp;lt;&amp;lt;
	Performance measurements are invalid in &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; mode.
	Record Size 512 kB
	File size set to 5502664 kB
	Command line used: iozone -I -i 0 -i 1 -i 2 -e -+d -r 512 -s 5502664 -f /mnt/lustre/d0.iozone/iozone.odir
	Output is in kBytes/sec
	Time Resolution = 0.000001 seconds.
	Processor cache size set to 1024 kBytes.
	Processor cache line size set to 32 bytes.
	File stride size set to 17 * record size.
                                                              random    random     bkwd    record    stride                                    
              kB  reclen    write  rewrite    read    reread    read     write     read   rewrite      read   fwrite frewrite    fread  freread
         5502664     512 &lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
[22327.751978] ll_ost_io00_063 D ffff97f5b9a1d280     0 26250      2 0x00000080
[22327.753328] Call Trace:
[22327.753790]  [&amp;lt;ffffffff89589179&amp;gt;] schedule+0x29/0x70
[22327.754705]  [&amp;lt;ffffffff89586e41&amp;gt;] schedule_timeout+0x221/0x2d0
[22327.755872]  [&amp;lt;ffffffffc066e569&amp;gt;] ? arc_read+0x499/0xab0 [zfs]
[22327.756947]  [&amp;lt;ffffffff88e6d39e&amp;gt;] ? kvm_clock_get_cycles+0x1e/0x20
[22327.758168]  [&amp;lt;ffffffff89588a2d&amp;gt;] io_schedule_timeout+0xad/0x130
[22327.759318]  [&amp;lt;ffffffff88ec6ad6&amp;gt;] ? prepare_to_wait_exclusive+0x56/0x90
[22327.760531]  [&amp;lt;ffffffff89588ac8&amp;gt;] io_schedule+0x18/0x20
[22327.761529]  [&amp;lt;ffffffffc02d7262&amp;gt;] cv_wait_common+0xb2/0x150 [spl]
[22327.762636]  [&amp;lt;ffffffff88ec6f50&amp;gt;] ? wake_up_atomic_t+0x30/0x30
[22327.763703]  [&amp;lt;ffffffffc02d7338&amp;gt;] __cv_wait_io+0x18/0x20 [spl]
[22327.764885]  [&amp;lt;ffffffffc0738a8b&amp;gt;] zio_wait+0x11b/0x1c0 [zfs]
[22327.766007]  [&amp;lt;ffffffffc0695251&amp;gt;] dmu_tx_count_write+0x191/0x1d0 [zfs]
[22327.767227]  [&amp;lt;ffffffffc069532a&amp;gt;] dmu_tx_hold_write_by_dnode+0x3a/0x50 [zfs]
[22327.768579]  [&amp;lt;ffffffffc1185449&amp;gt;] osd_declare_write_commit+0x599/0x7f0 [osd_zfs]
[22327.769964]  [&amp;lt;ffffffffc069633b&amp;gt;] ? dmu_tx_create_dd+0x9b/0xe0 [zfs]
[22327.771150]  [&amp;lt;ffffffffc12c4739&amp;gt;] ofd_commitrw_write+0x939/0x1db0 [ofd]
[22327.772363]  [&amp;lt;ffffffffc12c8e1c&amp;gt;] ofd_commitrw+0x47c/0xa50 [ofd]
[22327.773572]  [&amp;lt;ffffffffc0f9fbec&amp;gt;] obd_commitrw+0x9c/0x370 [ptlrpc]
[22327.774739]  [&amp;lt;ffffffffc0fa4092&amp;gt;] tgt_brw_write+0xf02/0x1ae0 [ptlrpc]
[22327.775959]  [&amp;lt;ffffffffc0ade393&amp;gt;] ? cfs_trace_unlock_tcd+0x33/0x90 [libcfs]
[22327.777235]  [&amp;lt;ffffffffc0ae4cc8&amp;gt;] ? libcfs_debug_vmsg2+0x6d8/0xb30 [libcfs]
[22327.778585]  [&amp;lt;ffffffffc0ef5780&amp;gt;] ? target_send_reply_msg+0x170/0x170 [ptlrpc]
[22327.779932]  [&amp;lt;ffffffffc0fa5eea&amp;gt;] tgt_request_handle+0xada/0x1570 [ptlrpc]
[22327.781187]  [&amp;lt;ffffffffc0ae5177&amp;gt;] ? libcfs_debug_msg+0x57/0x80 [libcfs]
[22327.782418]  [&amp;lt;ffffffffc0f4abcb&amp;gt;] ptlrpc_server_handle_request+0x24b/0xab0 [ptlrpc]
[22327.783913]  [&amp;lt;ffffffffc0f4e534&amp;gt;] ptlrpc_main+0xb34/0x1470 [ptlrpc]
[22327.785118]  [&amp;lt;ffffffffc0f4da00&amp;gt;] ? ptlrpc_register_service+0xf80/0xf80 [ptlrpc]
[22327.786447]  [&amp;lt;ffffffff88ec5e61&amp;gt;] kthread+0xd1/0xe0
[22327.787448]  [&amp;lt;ffffffff88ec5d90&amp;gt;] ? insert_kthread_work+0x40/0x40
[22327.788567]  [&amp;lt;ffffffff89595df7&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[22327.789751]  [&amp;lt;ffffffff88ec5d90&amp;gt;] ? insert_kthread_work+0x40/0x40 &lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i00d87:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>