<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:15:12 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-1277] Initial IO is slow after mount on the client</title>
                <link>https://jira.whamcloud.com/browse/LU-1277</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Hi, we have an customer who have existing lustre filesystem (usage is about 60-70%). When they mount the lustre on the clients and start the IO the client, the initial IO performance is always slow.&lt;br/&gt;
When it writes the files and remove them... repeating many times, the system is going to be normal and IO performance is back. &lt;br/&gt;
This is odd behavior, but it always happens after the lustre starts.&lt;/p&gt;</description>
                <environment></environment>
        <key id="13835">LU-1277</key>
            <summary>Initial IO is slow after mount on the client</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="johann">Johann Lombardi</assignee>
                                    <reporter username="ihara">Shuichi Ihara</reporter>
                        <labels>
                    </labels>
                <created>Mon, 2 Apr 2012 06:23:37 +0000</created>
                <updated>Mon, 20 Aug 2012 09:51:13 +0000</updated>
                            <resolved>Mon, 20 Aug 2012 09:51:09 +0000</resolved>
                                    <version>Lustre 1.8.7</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="33260" author="ihara" created="Mon, 2 Apr 2012 06:27:15 +0000"  >&lt;p&gt;attached is syslog (we ran sysraq+t) on one of OSSs, we saw these call-trace on all of OSS, when this problem happens.&lt;/p&gt;</comment>
                            <comment id="33283" author="johann" created="Mon, 2 Apr 2012 08:52:33 +0000"  >&lt;p&gt;Hi Ihara,&lt;/p&gt;

&lt;p&gt;Once performance are back to &quot;normal&quot; from a given client, if you then mount a new client, are performance bad again on this new client until you repeat the test several times?&lt;br/&gt;
I wonder if this bug is instead related to OST restart, like the problem we have with bitmap loading (as mentioned in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15&quot; title=&quot;strange slow IO messages and bad performance &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15&quot;&gt;&lt;del&gt;LU-15&lt;/del&gt;&lt;/a&gt;).&lt;/p&gt;

&lt;p&gt;If this is really related to the lustre client itself, then it might be due to grant. It will take several bulk writes for the client to own 32MB of grant space. To isolate the problem, could you please:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;collect rpc_stats (on client) &amp;amp; brw_stats (on OSTs) for a good and bad run?&lt;/li&gt;
	&lt;li&gt;tell us if this problem is only observed with writes or also with reads?&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="33296" author="ihara" created="Mon, 2 Apr 2012 10:02:02 +0000"  >&lt;p&gt;Johann,&lt;/p&gt;

&lt;p&gt;we just tried to mount on a new client after back to the normal performance, but the performance was no problem on that client.&lt;/p&gt;

&lt;p&gt;btw, we saw similar behavior that was reported on &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15&quot; title=&quot;strange slow IO messages and bad performance &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15&quot;&gt;&lt;del&gt;LU-15&lt;/del&gt;&lt;/a&gt; - when it writes the files after all OSTs are mounted, there are many small read IOs although there are noshing read from user space.&lt;br/&gt;
This filesystem was formatted on 1.8.4, but already upgraded 1.8.7 few month ago.&lt;/p&gt;</comment>
                            <comment id="33298" author="johann" created="Mon, 2 Apr 2012 10:15:25 +0000"  >&lt;p&gt;If the problem only happens after OST restart, then i&apos;m afraid that the only &quot;solution&quot; would be to write some tool/patch to pre-load the bitmaps in memory.&lt;/p&gt;

&lt;p&gt;Now, if the problem can also happen while the OST wasn&apos;t restarted recently, it could mean that some bitmaps got evicted from memory and it might make sense to try to disable the OSS read cache feature (there are actually 2 parameters to disable) if not done already.&lt;/p&gt;</comment>
                            <comment id="33301" author="ihara" created="Mon, 2 Apr 2012 11:24:41 +0000"  >&lt;p&gt;So far, the problem only happens after mount the OSTs..&lt;br/&gt;
I thought &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-15&quot; title=&quot;strange slow IO messages and bad performance &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-15&quot;&gt;&lt;del&gt;LU-15&lt;/del&gt;&lt;/a&gt; was fixed in 1.8.6, but we still have a potential to get this problem even on 1.8.7?&lt;/p&gt;</comment>
                            <comment id="33304" author="green" created="Mon, 2 Apr 2012 12:29:38 +0000"  >&lt;p&gt;I think this is a longstanding issue related to bitmap loading and calculating of extent tables on ext4 mount.&lt;br/&gt;
I think people used debugfs right after start to preload the bitmaps and it helped.&lt;/p&gt;</comment>
                            <comment id="42718" author="johann" created="Mon, 6 Aug 2012 04:43:37 +0000"  >&lt;p&gt;Hi Ihara, are you ok to close this bug?&lt;/p&gt;</comment>
                            <comment id="43466" author="ihara" created="Sat, 18 Aug 2012 09:18:56 +0000"  >&lt;p&gt;yes, please. we will try dumpe2fs before start OSTs to load bitmap and extent table calculation. Thanks for advice!&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="11038" name="messages.t2s007057" size="1036858" author="ihara" created="Mon, 2 Apr 2012 06:27:15 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvh4n:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>6424</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>