<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:10:58 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7677] sudden severe slowdown for some OSTs in the file system</title>
                <link>https://jira.whamcloud.com/browse/LU-7677</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Over the last day we have had reports from our  the lustre file system is unusable due to very slow write speeds. &lt;/p&gt;

&lt;p&gt;While investigating the issue, we discovered that a number of OSTs are showing write throughput of &amp;lt;1MB/s while the rest is showing the expected &amp;gt;400MB/s throughput. Initially only a very small number of OSTs were affected but the issue now seems to affect an increasing number of OSTs.&lt;/p&gt;

&lt;p&gt;We have recovered the bulk of the performance for our users by disabling any affected OST on the MDT, allowing our users to continue operating for now. This is clearly not a situation we want to continue operating for longer than necessary, but it allowed us to keep the system up while we are investigating the issue.&lt;/p&gt;

&lt;p&gt;There does not seem to be any pattern which OSTs are affected, AFAICT they are distributed over all OSS nodes while all OSS also have perfectly working OSTs.&lt;/p&gt;

&lt;p&gt;There is nothing obvious in the logs on clients, OSS, MDS  and no obvious fault on the SFA10K.&lt;/p&gt;

&lt;p&gt;Currently our users are fairly ok with the situation but starting tomorrow (UK time) we will need to investigate this with high priority.&lt;/p&gt;

&lt;p&gt;The only Lustre related log entries we have been able to find are on all OSS nodes and are similar to the one below, repeating roughly every 10 minutes.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Jan 17 15:10:48 cs04r-sc-oss03-03 kernel: Lustre: 27537:0:(client.c:1939:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1453043393/real 1453043393]  req@ffff8805c377f9c0 x1521165154263304/t0(0) o250-&amp;gt;MGC10.144.144.1@o2ib@10.144.144.1@o2ib:26/25 lens 400/544 e 0 to 1 dl 1453043448 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
Jan 17 15:10:48 cs04r-sc-oss03-03 kernel: Lustre: 27537:0:(client.c:1939:ptlrpc_expire_one_request()) Skipped 8 previous similar messages
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;10.144.144.1@o2ib is the MDS which was active until we initiated a failover to the second MDS after upgrading that to 2.7.1+patches, we have so far not been able to identify the cause of this error.&lt;/p&gt;

&lt;p&gt;Any suggestions how to debug this (at least initially preferably without further impact on the production file system) would be much appreciated.&lt;/p&gt;</description>
                <environment>Rhel 6 clients and servers, clients on Lustre 2.7.1, OSSes on 2.7.0, active MDS on 2.7.1 + patches for changelog overflow, about 2 weeks ago, the active MDS was failed over to the standby which had been upgraded previously.&lt;br/&gt;
patches applied on MDS over 2.7.1 are:&lt;br/&gt;
&lt;br/&gt;
0e98253 &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6556&quot; title=&quot;changelog catalog corruption if all possible records is define &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6556&quot;&gt;&lt;strike&gt;LU-6556&lt;/strike&gt;&lt;/a&gt; obdclass: re-allow catalog to wrap around&lt;br/&gt;
7f86741 &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6634&quot; title=&quot;(osd_handler.c:901:osd_trans_start()) ASSERTION( get_current()-&amp;gt;journal_info == ((void *)0) ) failed: when reaching Catalog full condition&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6634&quot;&gt;&lt;strike&gt;LU-6634&lt;/strike&gt;&lt;/a&gt; llog: destroy plain llog if init fails&lt;br/&gt;
4616323 &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6634&quot; title=&quot;(osd_handler.c:901:osd_trans_start()) ASSERTION( get_current()-&amp;gt;journal_info == ((void *)0) ) failed: when reaching Catalog full condition&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6634&quot;&gt;&lt;strike&gt;LU-6634&lt;/strike&gt;&lt;/a&gt; llog: add llog_trans_destroy()&lt;br/&gt;
&lt;br/&gt;
OSTs on DDN SFA10K</environment>
        <key id="34145">LU-7677</key>
            <summary>sudden severe slowdown for some OSTs in the file system</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="4">Incomplete</resolution>
                                        <assignee username="green">Oleg Drokin</assignee>
                                    <reporter username="ferner">Frederik Ferner</reporter>
                        <labels>
                    </labels>
                <created>Sun, 17 Jan 2016 16:22:06 +0000</created>
                <updated>Tue, 14 Dec 2021 22:20:43 +0000</updated>
                            <resolved>Tue, 14 Dec 2021 22:20:43 +0000</resolved>
                                    <version>Lustre 2.7.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="139178" author="ferner" created="Mon, 18 Jan 2016 14:46:47 +0000"  >&lt;p&gt;FWIW, we have now confirmed that all affected OSTs are using the same controller on the SFA10K and all OSTs on that controller are affected. DDN are looking into this.&lt;/p&gt;</comment>
                            <comment id="139191" author="pjones" created="Mon, 18 Jan 2016 18:19:25 +0000"  >&lt;p&gt;ok Frederik. Let us know if there is anything else you need from us&lt;/p&gt;</comment>
                            <comment id="140652" author="ferner" created="Mon, 1 Feb 2016 14:19:47 +0000"  >&lt;p&gt;Peter, All,&lt;/p&gt;

&lt;p&gt;could we re-open this ticket please? Looks like we need to investigate this further.&lt;/p&gt;

&lt;p&gt;We have replaced the controller, the  problem went away but came back with the same symptoms after about a week.&lt;/p&gt;

&lt;p&gt;As I see it, we are now looking at investigating individual OSTs, preferably on a low level, eliminating as much of the Lustre layers as possible. Ideally without affecting the data. Any suggestions how we can debug this would be appreciated. (We&apos;ve got a maintenance window tomorrow where we&apos;ll try to do as much as we can.)&lt;/p&gt;

&lt;p&gt;Frederik&lt;/p&gt;</comment>
                            <comment id="140659" author="jfc" created="Mon, 1 Feb 2016 16:28:53 +0000"  >&lt;p&gt;By request of Frederick.&lt;br/&gt;
~ jfc.&lt;/p&gt;</comment>
                            <comment id="140673" author="green" created="Mon, 1 Feb 2016 18:18:38 +0000"  >&lt;p&gt;So if you want to remove the lustre out of the picture completely, you can use the raw block device testing with ike dd and stuff. it should be nondestructive for reads at least.&lt;/p&gt;

&lt;p&gt;You can mount the underlying lustre fs as ldiskfs and then read/write some files with some performance-measuring tool too and that sohuld be nondestructive too, just remember to delete the files afterwards.&lt;/p&gt;

&lt;p&gt;And you can use obdsurvey tools to read write via lustre layers but no network and I believe that &apos;s also nondestructive as it does not really reformat anything. Just need to remember to nuke the objects it used for IO.&lt;/p&gt;</comment>
                            <comment id="141134" author="ferner" created="Thu, 4 Feb 2016 09:11:56 +0000"  >&lt;p&gt;Oleg,&lt;/p&gt;

&lt;p&gt;thanks for the suggestions.&lt;/p&gt;

&lt;p&gt;We have been able to see that if an OST is slow, this is also observed when reading from the raw block device with dd. (I&apos;ve not tried any of the others...)&lt;/p&gt;

&lt;p&gt;We have now also managed to trigger slow throughput on an OST by running a script randomly writing 4k blocks in a file. If we run this script in 4 instances on a single client, each writing its own file, each file ~400GB, all files on the same OST. When doing this, with pretty much nothing else ongoing on the file system, the throughput drops from 600MB/s or more to &amp;lt;10MB/s when testing with a simple dd to another file on that OST from a different node. Is that expected?&lt;/p&gt;

&lt;p&gt;Frederik&lt;/p&gt;</comment>
                            <comment id="143112" author="green" created="Sat, 20 Feb 2016 20:52:11 +0000"  >&lt;p&gt;Well, typically if you write a lot of small random IO to disks, they become really slow due to all the seeking involved, so no real surprise there.&lt;/p&gt;

&lt;p&gt;The surprising part is you are doing this over a DDN device and I read they were smart about that and had a guaranteed floor rate of the IO even in case of a stream of random small IO.&lt;/p&gt;

&lt;p&gt;If you can demonstrate this with no Lustre in the picture by just mounting an OST volume as ldiskfs (or formatting a spare LUN as ext4 if you prefer), running 4 (or 4*8 = 32 to account for multiple RPCs in flight in case of Lustre) of those &quot;random stream of 4k writes per file&quot; &lt;em&gt;in DIRECTIO mode&lt;/em&gt; (the directio is important to force data to go to disk right away to match what Lustre does). and then do a streaming write (also in directio mode in 1M chunks), if that is slow, I think you&apos;d need to talk to DDN and they should be able to help you see if there&apos;s some sort of a configuration error or some other problem going on.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10490" key="com.atlassian.jira.plugin.system.customfieldtypes:datepicker">
                        <customfieldname>End date</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>Fri, 19 Feb 2016 16:22:06 +0000</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                            <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxye7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                        <customfield id="customfield_10493" key="com.atlassian.jira.plugin.system.customfieldtypes:datepicker">
                        <customfieldname>Start date</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>Mon, 18 Jan 2016 16:22:06 +0000</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                    </customfields>
    </item>
</channel>
</rss>