<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:11:51 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7780] MDS crashed with oom-killer</title>
                <link>https://jira.whamcloud.com/browse/LU-7780</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Error happened during soak testing of build &apos;20160215&apos; (see &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20150215&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20150215&lt;/a&gt;). DNE is enabled.&lt;br/&gt;
MDT had been formatted using ldiskfs, OSTs using zfs. MDS nodes are configured in active-active HA failover configuration.&lt;/p&gt;

&lt;p&gt;Please note that build 20150215 is a vanilla build of the master brunch.&lt;br/&gt;
This issue might be addressed by the changes included in build &apos;20160210&apos; as we didn&apos;t observe this issue in a two day test session.&lt;/p&gt;

&lt;ul&gt;
	&lt;li&gt;2016-02-15 15:37:51,169:fsmgmt.fsmgmt:INFO     triggering fault mds_failover (for &lt;tt&gt;lola-11&lt;/tt&gt;)&lt;/li&gt;
	&lt;li&gt;2016-02-15 15:44:57,839:fsmgmt.fsmgmt:INFO     mds_failover just completed (for &lt;tt&gt;lola-11&lt;/tt&gt;)&lt;/li&gt;
	&lt;li&gt;After that the slabs memory consumption of slabs continuously increased till all resources are exhausted at 2016-02-15 22:38.&lt;/li&gt;
	&lt;li&gt;Most pages are allocated by &lt;em&gt;size-1048576&lt;/em&gt; slabs. High score list reads as
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;#Date Time SlabName ObjInUse ObjInUseB ObjAll ObjAllB SlabInUse SlabInUseB SlabAll SlabAllB SlabChg SlabPct
20160215 22:46:20 size-1048576 29147 30562844672 29147 30562844672 29147 30562844672 29147 30562844672 0 0
20160215 22:46:20 size-262144 1793 470024192 1793 470024192 1793 470024192 1793 470024192 0 0
20160215 22:46:20 ptlrpc_cache 399364 306711552 399380 306723840 79873 327159808 79876 327172096 180224 0
20160215 22:46:20 size-1024 229179 234679296 229188 234688512 57295 234680320 57297 234688512 -24576 0
20160215 22:46:20 size-512 256540 131348480 258232 132214784 32278 132210688 32279 132214784 86016 0
20160215 22:46:20 size-192 460848 88482816 460880 88488960 23043 94384128 23044 94388224 28672 0
20160215 22:46:20 size-8192 5776 47316992 5776 47316992 5776 47316992 5776 47316992 -8192 0
20160215 22:46:20 size-128 265120 33935360 266250 34080000 8875 36352000 8875 36352000 0 0
20160215 22:46:20 size-65536 361 23658496 361 23658496 361 23658496 361 23658496 0 0
20160215 22:46:20 kmem_cache 289 9506944 289 9506944 289 18939904 289 18939904 0 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;(see attached file slab-usage-by-allocation-descending.dat)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Attached files messages, console file of &lt;tt&gt;lola-11&lt;/tt&gt;. Sorted slab usage as oom-killer was started.&lt;/p&gt;</description>
                <environment>lola&lt;br/&gt;
&amp;nbsp;build: 2.8.50-6-gf9ca359 ; commit f9ca359284357d145819beb08b316e932f7a3060 </environment>
        <key id="34687">LU-7780</key>
            <summary>MDS crashed with oom-killer</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="heckes">Frank Heckes</assignee>
                                    <reporter username="heckes">Frank Heckes</reporter>
                        <labels>
                            <label>soak</label>
                    </labels>
                <created>Tue, 16 Feb 2016 11:46:43 +0000</created>
                <updated>Mon, 6 Jun 2016 15:52:25 +0000</updated>
                            <resolved>Wed, 16 Mar 2016 19:27:29 +0000</resolved>
                                                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="143151" author="heckes" created="Mon, 22 Feb 2016 10:07:00 +0000"  >&lt;p&gt;Three (different) MDSes crashed with oom-killer started over the weekend (2016-02-20 - 2016-02-21).&lt;br/&gt;
Performance counters and event sequence can be provided if necessary&lt;/p&gt;</comment>
                            <comment id="143232" author="green" created="Mon, 22 Feb 2016 18:43:41 +0000"  >&lt;p&gt;Can you please do the tip of b2_8 testign to see if it&apos;s also affected?&lt;/p&gt;

&lt;p&gt;Also on master - if you can enable memory allocation tracing, reproduce, and then extract the debug log out of a crashdump so that we can see where are all of those allocations come from.&lt;/p&gt;</comment>
                            <comment id="144741" author="heckes" created="Mon, 7 Mar 2016 14:11:19 +0000"  >&lt;p&gt;currently &apos;+malloc&apos; has been added to thed debug on server nodes&lt;/p&gt;</comment>
                            <comment id="145476" author="adilger" created="Mon, 14 Mar 2016 19:33:28 +0000"  >&lt;p&gt;Frank or Cliff,&lt;br/&gt;
it might be useful to increase the debug buffer size and turn on trace logging (&lt;tt&gt;lctl set_param debug_mb=128 debug=+trace&lt;/tt&gt;) and then let it continue to run for a minute or two and dump the debug log without having to wait for the crash to be hit.  The &lt;tt&gt;size-1024&lt;/tt&gt; items are being allocated 10/sec so the cause will appear in any logs.  Similarly, the &lt;tt&gt;size-1048576&lt;/tt&gt; allocations are done every minute so the cause should also be in any logs dumped.   The &lt;tt&gt;size-262144&lt;/tt&gt; is only being allocated in chunks every 15 minutes, so it might be related to recovery if that is the frequency of failovers, so that might need to be timed to coordinate with a failover or similar, or the size of the log increased to 512 or 1024 MB.&lt;/p&gt;

&lt;p&gt;Di, the memory is all used in ptlrpc_cache and large (1MB and 512KB) slab allocations.  My first guess is that these large allocations relate to striped directory RPCs (i.e. OUT) and recovery, and are probably not being freed except at shutdown.  There are 30k 1MB allocations consuming a whopping 30GB of memory, 400k ptlrpc_cache entries consuming 300MB of memory, and 230k 1KB allocations consuming 235MB.  All of those slabs are growing continuously from startup, yet there aren&apos;t any objects (ldlm_lock, ldlm_resource, ldiskfs_inode, buffer_head, dentry, etc) that might normally grow as the node is caching a lot of data.&lt;/p&gt;

&lt;p&gt;It might be that we are just keeping too many requests in &lt;tt&gt;ptlrpc_cache&lt;/tt&gt;?  In that case, we might need to add a slab shrinker for ptlrpc_cache to keep the size of this slab in check, but I&apos;m not sure this is the root problem because it seems all of the &lt;tt&gt;ptlrpc_cache&lt;/tt&gt; items are in use, so it may be that there is a stray reference on the request that isn&apos;t being put somewhere? It is also a bit sad that the cache name is &lt;tt&gt;ptlrpc_cache&lt;/tt&gt; but the internal usage is &lt;tt&gt;request_cache&lt;/tt&gt;.  The slab cache name should be changed to &lt;tt&gt;ptlrpc_request_cache&lt;/tt&gt; to match the code so that grep can find all of the relevant code at once.&lt;/p&gt;

&lt;p&gt;Is it possible that ptlrpc_history is enabled on these nodes?  That would also keep a lot of requests pinned in memory in &lt;tt&gt;ptlrpc_server_drop_request()&lt;/tt&gt;.  Normally this wouldn&apos;t be a problem, but if there are so many large buffers attached to the RPCs this could be causing the OOM, when smaller (e.g. 32KB) buffers wouldn&apos;t be a problem.  It also isn&apos;t clear that there is a benefit to keeping the request buffers in the request history, since they are never used, so it might also be possible to release the buffers before inserting the request into the history, but I&apos;m not sure if that is required here or not?&lt;/p&gt;</comment>
                            <comment id="145585" author="heckes" created="Tue, 15 Mar 2016 15:36:42 +0000"  >&lt;p&gt;Debug filter has been changed to : &lt;tt&gt;debug_mb=128 debug=+malloc +trace&lt;/tt&gt;&lt;/p&gt;</comment>
                            <comment id="145839" author="pjones" created="Wed, 16 Mar 2016 18:58:03 +0000"  >&lt;p&gt;Moving to 2.9 because it seems that this issue only occurs with multiple MDTs per MDS and does not happen with the more common configuration of a single MDT per MDS. Is this a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7836&quot; title=&quot;MDSes crashed with oom-killer&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7836&quot;&gt;&lt;del&gt;LU-7836&lt;/del&gt;&lt;/a&gt;?&lt;/p&gt;</comment>
                            <comment id="145850" author="di.wang" created="Wed, 16 Mar 2016 19:24:19 +0000"  >&lt;p&gt;Yes, this is duplicate with lu-7836&lt;/p&gt;</comment>
                            <comment id="145851" author="pjones" created="Wed, 16 Mar 2016 19:27:29 +0000"  >&lt;p&gt;ok thanks Di&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="35106">LU-7836</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="20390" name="console-lola-11.log.bz2" size="89490" author="heckes" created="Tue, 16 Feb 2016 12:00:40 +0000"/>
                            <attachment id="20391" name="lola-11-memory-counter-20160215.dat.bz2" size="38231" author="heckes" created="Tue, 16 Feb 2016 12:00:40 +0000"/>
                            <attachment id="20392" name="messages-lola-11.log.bz2" size="73568" author="heckes" created="Tue, 16 Feb 2016 12:00:40 +0000"/>
                            <attachment id="20393" name="slab-details.tar.bz2" size="851051" author="heckes" created="Tue, 16 Feb 2016 12:00:40 +0000"/>
                            <attachment id="20394" name="slab-usage-by-allocation-descending.dat.bz2" size="4524" author="heckes" created="Tue, 16 Feb 2016 12:00:40 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzy1cf:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>