<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:45:07 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4705] LustreError: 89827:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2</title>
                <link>https://jira.whamcloud.com/browse/LU-4705</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Unexpected MDC LustreError&apos;s on most clients.  &lt;/p&gt;

&lt;p&gt;Client 10:&lt;br/&gt;
Mar  4 03:27:11 lustre10 kernel: LustreError: 183913:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&lt;/p&gt;

&lt;p&gt;Client 11:&lt;br/&gt;
Mar  4 00:37:25 lustre11 kernel: LustreError: 89827:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&lt;/p&gt;

&lt;p&gt;Client 12:&lt;br/&gt;
Mar  4 00:39:36 lustre12 kernel: LustreError: 11-0: cal-MDT0000-mdc-ffff8807b75c4000: Communicating with 192.168.20.1@tcp1, operation ldlm_enqueue failed with -116.&lt;br/&gt;
Mar  4 00:39:36 lustre12 kernel: LustreError: 70225:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -116&lt;br/&gt;
Mar  4 00:39:36 lustre12 kernel: LustreError: 70225:0:(vvp_io.c:1227:vvp_io_init()) cal: refresh file layout &lt;span class=&quot;error&quot;&gt;&amp;#91;0x200001c0b:0x176e:0x0&amp;#93;&lt;/span&gt; error -116.&lt;br/&gt;
Mar  4 03:09:33 lustre12 kernel: LustreError: 70225:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&lt;/p&gt;

&lt;p&gt;Client 13:&lt;br/&gt;
Mar  4 00:29:54 lustre13 kernel: LustreError: 167294:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&lt;/p&gt;

&lt;p&gt;Client 14:&lt;br/&gt;
Mar  4 01:18:04 lustre14 kernel: LustreError: 11-0: cal-MDT0000-mdc-ffff880787af8400: Communicating with 192.168.20.1@tcp1, operation ldlm_enqueue failed with -116.&lt;br/&gt;
Mar  4 01:18:04 lustre14 kernel: LustreError: 11503:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -116&lt;br/&gt;
Mar  4 01:18:04 lustre14 kernel: LustreError: 11503:0:(vvp_io.c:1227:vvp_io_init()) cal: refresh file layout &lt;span class=&quot;error&quot;&gt;&amp;#91;0x200001c12:0xbbe2:0x0&amp;#93;&lt;/span&gt; error -116.&lt;/p&gt;

&lt;p&gt;Client 16:&lt;br/&gt;
Mar  4 01:00:46 lustre16 kernel: LustreError: 141605:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&lt;/p&gt;

&lt;p&gt;Client 17:&lt;br/&gt;
Mar  4 00:13:39 lustre17 kernel: LustreError: 11-0: cal-MDT0000-mdc-ffff8808038aa000: Communicating with 192.168.20.1@tcp1, operation ldlm_enqueue failed with -116.&lt;br/&gt;
Mar  4 00:13:39 lustre17 kernel: LustreError: 126770:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -116&lt;br/&gt;
Mar  4 00:13:39 lustre17 kernel: LustreError: 126770:0:(vvp_io.c:1227:vvp_io_init()) cal: refresh file layout &lt;span class=&quot;error&quot;&gt;&amp;#91;0x200001beb:0x1aedf:0x0&amp;#93;&lt;/span&gt; error -116.&lt;br/&gt;
Mar  4 02:02:43 lustre17 kernel: LustreError: 126770:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&lt;/p&gt;

&lt;p&gt;Client 18:&lt;br/&gt;
Mar  1 05:34:03 lustre18 kernel: LustreError: 146331:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&lt;/p&gt;</description>
                <environment>Running tip of Lustre b2_5, 1 MGS, 1 MDS, 2 OSS, 12 clients. </environment>
        <key id="23453">LU-4705</key>
            <summary>LustreError: 89827:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="brett">Brett Lee</reporter>
                        <labels>
                    </labels>
                <created>Tue, 4 Mar 2014 15:28:35 +0000</created>
                <updated>Thu, 26 Oct 2017 16:49:41 +0000</updated>
                            <resolved>Tue, 24 Oct 2017 13:01:01 +0000</resolved>
                                    <version>Lustre 2.5.1</version>
                                    <fixVersion>Lustre 2.11.0</fixVersion>
                    <fixVersion>Lustre 2.10.2</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="78882" author="keith" created="Mon, 10 Mar 2014 15:52:31 +0000"  >&lt;p&gt;I see these same errors with a Lustre 2.5.0 Client.  The do not seem to impact the usability of the filesystem.   But this is listed as a Error so there could be something happening. &lt;/p&gt;</comment>
                            <comment id="78910" author="adilger" created="Mon, 10 Mar 2014 19:29:24 +0000"  >&lt;p&gt;Is the filesystem re-exported via NFS, or possibly have concurrent threads that are accessing and unlinking files?&lt;/p&gt;

&lt;p&gt;These messages mean that the client was looking up some file, but it was deleted by the time it tried to access it.&lt;/p&gt;

&lt;p&gt;-116 = -ESTALE, -2 = -ENOENT.&lt;/p&gt;

&lt;p&gt;The errors are not really fatal, and could probably be quieted from the console.&lt;/p&gt;</comment>
                            <comment id="78929" author="keith" created="Mon, 10 Mar 2014 21:17:50 +0000"  >&lt;p&gt;I have seen this error with IOR no NFS.  I am not sure if the errors were generated during one single file or file per process. &lt;/p&gt;</comment>
                            <comment id="79020" author="brett" created="Tue, 11 Mar 2014 17:20:41 +0000"  >&lt;p&gt;No, there was no re-exporting, but each Lustre client did have four (4) mounts of the file system - each mount appearing active via the stats files in /proc.&lt;/p&gt;</comment>
                            <comment id="79255" author="adilger" created="Thu, 13 Mar 2014 17:17:32 +0000"  >&lt;p&gt;Brett, what was the workload being run here?  Something that is creating and deleting files concurrently (e.g. racer), or possibly multiple threads doing &quot;rm -r&quot; on the same tree?  Either this is &quot;normal&quot; and maybe we should quiet the error messages, or it might imply some sort of bug on the MDS with inode lookup or files unexpectedly being deleted.  Are there application-visible errors that are unexpected (&quot;No such file or directory&quot;)?&lt;/p&gt;</comment>
                            <comment id="80401" author="brett" created="Thu, 27 Mar 2014 18:41:12 +0000"  >&lt;p&gt;Andreas, the workload was a mix of real jobs with varying IO patterns - most prominent of which was many small reads from large files.  There was no artificial creating/deleting of files.  As for the application, am now noticing that a setting disabled printing of &quot;some&quot; error an warning messages during this run, however, each job completed successfully.  No unexpected application-visible errors were seen.&lt;/p&gt;</comment>
                            <comment id="140616" author="mjo" created="Sat, 30 Jan 2016 19:32:17 +0000"  >&lt;p&gt;This is being seen at Gulfstream.  In their environment, there doesn&apos;t appear to be any operational consequence to it.  But, it scared them. It&apos;d be nice if we could mute these errors, as discussed in &lt;a href=&quot;https://jira.hpdd.intel.com/browse/LU-4705?focusedCommentId=79255&amp;amp;page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-79255&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://jira.hpdd.intel.com/browse/LU-4705?focusedCommentId=79255&amp;amp;page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-79255&lt;/a&gt;&lt;/p&gt;
</comment>
                            <comment id="142872" author="kjstrosahl" created="Thu, 18 Feb 2016 20:43:57 +0000"  >&lt;p&gt;I just saw an instance of this error in the Lustre file system at TJNAF.  It is the only instance I can recall of it being seen here, we are running lustre 2.5.3 pristine &lt;/p&gt;

&lt;p&gt;To expand a bit more... I have a test environment that I&apos;m using to benchmark oss systems.  Presently I have three osts on a single server running lustre 2.5.3.  I&apos;ve mounted it on a single client and am running IOR tests with the following parameters:&lt;/p&gt;

&lt;p&gt;mpirun -np 12 -bynode -machinefile ./nodelist ./ior -F -e -m -g -i 10 -t 1024k -b 42G -o /testL/benchmark/test&lt;/p&gt;

&lt;p&gt;where nodelist contains a single node.&lt;/p&gt;</comment>
                            <comment id="208270" author="gerrit" created="Wed, 13 Sep 2017 16:53:45 +0000"  >&lt;p&gt;Andreas Dilger (andreas.dilger@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/28978&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/28978&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4705&quot; title=&quot;LustreError: 89827:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4705&quot;&gt;&lt;del&gt;LU-4705&lt;/del&gt;&lt;/a&gt; mdc: improve mdc_enqueue() error message&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 9d8f53da6ac5482262c188ba1e0ca3fb395aedfd&lt;/p&gt;</comment>
                            <comment id="211765" author="gerrit" created="Tue, 24 Oct 2017 07:18:46 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/28978/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/28978/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4705&quot; title=&quot;LustreError: 89827:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4705&quot;&gt;&lt;del&gt;LU-4705&lt;/del&gt;&lt;/a&gt; mdc: improve mdc_enqueue() error message&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 471c5303eb29d5ea1ba5a683173bda63095dae78&lt;/p&gt;</comment>
                            <comment id="211788" author="pjones" created="Tue, 24 Oct 2017 13:01:01 +0000"  >&lt;p&gt;Landed for 2.11&lt;/p&gt;</comment>
                            <comment id="211818" author="gerrit" created="Tue, 24 Oct 2017 15:11:06 +0000"  >&lt;p&gt;Minh Diep (minh.diep@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/29736&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/29736&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4705&quot; title=&quot;LustreError: 89827:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4705&quot;&gt;&lt;del&gt;LU-4705&lt;/del&gt;&lt;/a&gt; mdc: improve mdc_enqueue() error message&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_10&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: b51accd1a652406afbe41ad764d116d0f361a0fb&lt;/p&gt;</comment>
                            <comment id="212090" author="gerrit" created="Thu, 26 Oct 2017 16:09:20 +0000"  >&lt;p&gt;John L. Hammond (john.hammond@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/29736/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/29736/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4705&quot; title=&quot;LustreError: 89827:0:(mdc_locks.c:916:mdc_enqueue()) ldlm_cli_enqueue: -2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4705&quot;&gt;&lt;del&gt;LU-4705&lt;/del&gt;&lt;/a&gt; mdc: improve mdc_enqueue() error message&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_10&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: c27470755cf40ee33056011883a0d0600ce00340&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="24482">LU-4973</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="22829">LU-4522</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwgpb:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>12942</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>