<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:16:10 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-8280] Bad Lustre Read Performance with Master Build 3371</title>
                <link>https://jira.whamcloud.com/browse/LU-8280</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;*Important notes: Please do not rebuild Lustre on wolf-&lt;span class=&quot;error&quot;&gt;&amp;#91;33-45&amp;#93;&lt;/span&gt;, thanks!&lt;/p&gt;

&lt;p&gt;The read performance of Lustre master build 3371 is bad. Following are the detailed configurations for my evaluation over Lustre file system using IOR benchmark:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;Lustre Configuration&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;MDS/MDT: wolf-37&lt;br/&gt;
OST/OSS: wolf-&lt;span class=&quot;error&quot;&gt;&amp;#91;33-36&amp;#93;&lt;/span&gt; (Each OST: MD RAID 0 striped &lt;span class=&quot;error&quot;&gt;&amp;#91;Chunk = 1M&amp;#93;&lt;/span&gt; over 5 SATA Drives)&lt;br/&gt;
Clients: wolf-&lt;span class=&quot;error&quot;&gt;&amp;#91;38-45&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;Stripe Size: 4194304&lt;br/&gt;
Stripe Count: 4&lt;/p&gt;

&lt;p&gt;Lustre Build Version: 3371&lt;br/&gt;
Lustre Build Command:&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;loadjenkinsbuild -b 3371 -a x86_64 -j lustre-master --jenkinsuri &lt;a href=&quot;https://build.hpdd.intel.com&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://build.hpdd.intel.com&lt;/a&gt; -p test-el6-x86_64 -t server -d el6.7 -n wolf-45 -r -v&lt;/li&gt;
&lt;/ol&gt;



&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;IOR Benchmark&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;IOR Command:&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;mpirun -np 4 -iface ib0 -f /home/congxu/host-ib /home/congxu/Software/ior-master/src/ior -a POSIX -N 4 -d 5 -i 1 -s 32768 -b 4MiB -t 4MiB -w -r -o /mnt/lustre/cong/testfile&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;IOR Results:&lt;/p&gt;

&lt;p&gt;Share File case&lt;br/&gt;
IOR-3.0.1: MPI Coordinated Test of Parallel I/O &lt;/p&gt;

&lt;p&gt;Began: Tue Jun 14 20:16:18 2016&lt;br/&gt;
Command line used: /home/congxu/Software/ior-master/src/ior -a POSIX -N 4 -d 5 -i 1 -s 32768 -b 4MiB -t 4MiB -w -r -o /mnt/lustre/cong/testfile&lt;br/&gt;
Machine: Linux wolf-38.wolf.hpdd.intel.com&lt;/p&gt;

&lt;p&gt;Test 0 started: Tue Jun 14 20:16:18 2016&lt;br/&gt;
Summary:&lt;br/&gt;
    api                = POSIX&lt;br/&gt;
    test filename      = /mnt/lustre/cong/testfile&lt;br/&gt;
    access             = single-shared-file&lt;br/&gt;
    ordering in a file = sequential offsets&lt;br/&gt;
    ordering inter file= no tasks offsets&lt;br/&gt;
    clients            = 4 (1 per node)&lt;br/&gt;
    repetitions        = 1&lt;br/&gt;
    xfersize           = 4 MiB &lt;br/&gt;
    blocksize          = 4 MiB &lt;br/&gt;
    aggregate filesize = 512 GiB &lt;/p&gt;

&lt;p&gt;access    bw(MiB/s)  block(KiB) xfer(KiB)  open(s)    wr/rd(s)   close(s)   total(s)   iter&lt;br/&gt;
------    ---------  ---------- ---------  --------   --------   --------   --------   ----&lt;br/&gt;
write     2546.35    4096       4096       0.001613   205.90     10.62      205.90     0   &lt;br/&gt;
read      1076.82    4096       4096       0.000565   486.89     74.16      486.89     0   &lt;br/&gt;
remove    -          -          -          -          -          -          0.001506   0   &lt;/p&gt;

&lt;p&gt;Max Write: 2546.35 MiB/sec (2670.05 MB/sec)&lt;br/&gt;
Max Read:  1076.82 MiB/sec (1129.12 MB/sec)&lt;/p&gt;

&lt;p&gt;Summary of all tests:&lt;br/&gt;
Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum&lt;br/&gt;
write        2546.35    2546.35    2546.35       0.00  205.89756 0 4 1 1 0 0 1 0 0 32768 4194304 4194304 549755813888 POSIX 0&lt;br/&gt;
read         1076.82    1076.82    1076.82       0.00  486.88718 0 4 1 1 0 0 1 0 0 32768 4194304 4194304 549755813888 POSIX 0&lt;/p&gt;

&lt;p&gt;Finished: Tue Jun 14 20:28:01 2016&lt;/p&gt;



&lt;p&gt;*************************************&lt;b&gt;On the contrary, &quot;lustre-b2_8 -b 12&quot; performs well&lt;/b&gt;**********************************&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;Lustre Configuration&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;MDS/MDT: wolf-37&lt;br/&gt;
OST/OSS: wolf-&lt;span class=&quot;error&quot;&gt;&amp;#91;33-36&amp;#93;&lt;/span&gt; (Each OST: MD RAID 0 striped &lt;span class=&quot;error&quot;&gt;&amp;#91;Chunk = 1M&amp;#93;&lt;/span&gt; over 5 SATA Drives)&lt;br/&gt;
Clients: wolf-&lt;span class=&quot;error&quot;&gt;&amp;#91;38-45&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;Stripe Size: 4194304&lt;br/&gt;
Stripe Count: 4&lt;/p&gt;

&lt;p&gt;Lustre Build Version: 12&lt;br/&gt;
Lustre Build Command:&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;loadjenkinsbuild -b 12 -a x86_64 -j lustre-b2_8 --jenkinsuri &lt;a href=&quot;https://build.hpdd.intel.com&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://build.hpdd.intel.com&lt;/a&gt; -p test-el6-x86_64 -t server -d el6.7 -n wolf-45 -r -v&lt;/li&gt;
&lt;/ol&gt;



&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;IOR Benchmark&amp;#93;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;IOR Command:&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;mpirun -np 4 -iface ib0 -f /home/congxu/host-ib /home/congxu/Software/ior-master/src/ior -a POSIX -N 4 -d 5 -i 1 -s 32768 -b 4MiB -t 4MiB -w -r -o /mnt/lustre/cong/testfile&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;IOR Results:&lt;/p&gt;

&lt;p&gt;Share File case&lt;br/&gt;
IOR-3.0.1: MPI Coordinated Test of Parallel I/O&lt;/p&gt;

&lt;p&gt;Began: Wed Jun 15 01:17:36 2016&lt;br/&gt;
Command line used: /home/congxu/Software/ior-master/src/ior -a POSIX -N 4 -d 5 -i 1 -s 32768 -b 4MiB -t 4MiB -w -r -o /mnt/lustre/cong/testfile&lt;br/&gt;
Machine: Linux wolf-38.wolf.hpdd.intel.com&lt;/p&gt;

&lt;p&gt;Test 0 started: Wed Jun 15 01:17:36 2016&lt;br/&gt;
Summary:&lt;br/&gt;
	api                = POSIX&lt;br/&gt;
	test filename      = /mnt/lustre/cong/testfile&lt;br/&gt;
	access             = single-shared-file&lt;br/&gt;
	ordering in a file = sequential offsets&lt;br/&gt;
	ordering inter file= no tasks offsets&lt;br/&gt;
	clients            = 4 (1 per node)&lt;br/&gt;
	repetitions        = 1&lt;br/&gt;
	xfersize           = 4 MiB&lt;br/&gt;
	blocksize          = 4 MiB&lt;br/&gt;
	aggregate filesize = 512 GiB&lt;/p&gt;

&lt;p&gt;access    bw(MiB/s)  block(KiB) xfer(KiB)  open(s)    wr/rd(s)   close(s)   total(s)   iter&lt;br/&gt;
------    ---------  ---------- ---------  --------   --------   --------   --------   ----&lt;br/&gt;
write     2559.64    4096       4096       0.001187   204.83     5.84       204.83     0   &lt;br/&gt;
read      2840.79    4096       4096       0.000808   184.56     24.11      184.56     0   &lt;br/&gt;
remove    -          -          -          -          -          -          0.001111   0   &lt;/p&gt;

&lt;p&gt;Max Write: 2559.64 MiB/sec (2683.98 MB/sec)&lt;br/&gt;
Max Read:  2840.79 MiB/sec (2978.78 MB/sec)&lt;/p&gt;

&lt;p&gt;Summary of all tests:&lt;br/&gt;
Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum&lt;br/&gt;
write        2559.64    2559.64    2559.64       0.00  204.82859 0 4 1 1 0 0 1 0 0 32768 4194304 4194304 549755813888 POSIX 0&lt;br/&gt;
read         2840.79    2840.79    2840.79       0.00  184.55737 0 4 1 1 0 0 1 0 0 32768 4194304 4194304 549755813888 POSIX 0&lt;/p&gt;

&lt;p&gt;Finished: Wed Jun 15 01:24:15 2016&lt;/p&gt;</description>
                <environment></environment>
        <key id="37593">LU-8280</key>
            <summary>Bad Lustre Read Performance with Master Build 3371</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="6">Not a Bug</resolution>
                                        <assignee username="cliffw">Cliff White</assignee>
                                    <reporter username="czx0003">Cong Xu</reporter>
                        <labels>
                    </labels>
                <created>Wed, 15 Jun 2016 01:10:17 +0000</created>
                <updated>Sat, 23 Jul 2016 05:13:00 +0000</updated>
                            <resolved>Sat, 23 Jul 2016 05:13:00 +0000</resolved>
                                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="155726" author="green" created="Wed, 15 Jun 2016 03:36:52 +0000"  >&lt;p&gt;I moved this into Lustre project sine it seems that newer Lustre versions are slower and 2.8 is still good, so it&apos;s unlikely to do anything with the infrastructure.&lt;/p&gt;</comment>
                            <comment id="156992" author="cliffw" created="Mon, 27 Jun 2016 15:18:17 +0000"  >&lt;p&gt;Note: we normally run IOR with -i 5 at least - a single iteration does not produce a consistent result. Also, our normal performance tests use -b 4G, not -b 4m. &lt;/p&gt;</comment>
                            <comment id="157059" author="czx0003" created="Mon, 27 Jun 2016 20:55:29 +0000"  >&lt;p&gt;In this evaluation, we configure Lustre stripe size, IOR Transfer size and Block size to be 4MB, and the number of Clients equals to number of OSTs. Thus, everything is perfect matched and we expect to deliver the maximum bandwidth of Lustre file system.&lt;/p&gt;</comment>
                            <comment id="157146" author="cliffw" created="Tue, 28 Jun 2016 17:05:38 +0000"  >&lt;p&gt;I ran a comparison of 2.8.0 and 2.8.53.38 on Spirit. &lt;br/&gt;
Backend storage was a DDN 12k with 8 LUNs. Each LUN was 7.0 TB, Total filesystem size was 142TB.&lt;br/&gt;
The OSS uses dual port FDR Infiniband, the clients have a single port FDR IB connection. &lt;br/&gt;
I ran a matrix, first with 8 IOR threads per client, second with 16 IOR threads per client. &lt;br/&gt;
Test were run with 1,4, 8 and 16 clients. &lt;br/&gt;
IOR File-per-process and IOR single-shared-file tests were run. &lt;br/&gt;
For the case of a single client with 8 or 16 threads, IOR file-per-process read performance for 2.8.53.38 is well below 2.8.0.&lt;br/&gt;
For all other cases, performance on 2.8.53.38 is equal or above 2.8.0.  &lt;br/&gt;
The single-client case is interesting and may be worth further examination.&lt;br/&gt;
All test runs, vmstat and collectl data is attached to the bug. &lt;/p&gt;

</comment>
                            <comment id="157158" author="cliffw" created="Tue, 28 Jun 2016 17:43:42 +0000"  >&lt;p&gt;In my experience a single client cannot generate enough IO request to saturate a single OSS, &amp;gt;1 client per OST is normally required. For this reason we run performance across a range of client sizes, and use multiple threads per client. &lt;/p&gt;</comment>
                            <comment id="157184" author="cliffw" created="Tue, 28 Jun 2016 20:55:14 +0000"  >&lt;p&gt;This spreadsheet contains the relevant runs from Hyperion. 2.8.55 is below 2.8.0 performance @ 16 clients for both reads and writes, writes are also below standard at larger client counts. &lt;/p&gt;</comment>
                            <comment id="159662" author="pjones" created="Sat, 23 Jul 2016 05:13:00 +0000"  >&lt;p&gt;IIRC Cliff had reported that this was an issue with the way the tests were being run&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="22056" name="2.8.0 vs 2.8.54 Jun 28 2016.xlsx" size="47787" author="cliffw" created="Tue, 28 Jun 2016 20:55:14 +0000"/>
                            <attachment id="22048" name="Spirit LU-8280 Jun 27 2016.xlsx" size="54206" author="cliffw" created="Tue, 28 Jun 2016 17:40:50 +0000"/>
                            <attachment id="22044" name="lu-8280-2.8.0-alldata.tar.gz" size="1459420" author="cliffw" created="Tue, 28 Jun 2016 17:06:23 +0000"/>
                            <attachment id="22045" name="lu-8280-2.8.53-alldata.tar.gz" size="3323508" author="cliffw" created="Tue, 28 Jun 2016 17:06:23 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzyeof:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>