<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:50:07 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-5282] (mdc_request.c:791:mdc_free_open()) ASSERTION( mod-&gt;mod_open_req-&gt;rq_replay == 0 ) failed</title>
                <link>https://jira.whamcloud.com/browse/LU-5282</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;In ll_close_inode_openhandle() if mdc_close() is not called or if it fails to clear mod-&amp;gt;mod_open_req-&amp;gt;rq_replay then we have the following LBUG:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[ 2759.567725] LustreError: 23179:0:(mdc_request.c:791:mdc_free_open()) ASSERTION( mod-&amp;gt;m\
od_open_req-&amp;gt;rq_replay == 0 ) failed:
[ 2759.568559] LustreError: 23203:0:(mdc_request.c:791:mdc_free_open()) ASSERTION( mod-&amp;gt;m\
od_open_req-&amp;gt;rq_replay == 0 ) failed:
[ 2759.568563] LustreError: 23203:0:(mdc_request.c:791:mdc_free_open()) LBUG
[ 2759.568564] Pid: 23203, comm: ls
[ 2759.568566]
[ 2759.568566] Call Trace:
[ 2759.568584]  [&amp;lt;ffffffffa0ef18c5&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
[ 2759.568594]  [&amp;lt;ffffffffa0ef1ec7&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
[ 2759.568601]  [&amp;lt;ffffffffa055ce60&amp;gt;] mdc_clear_open_replay_data+0x0/0x240 [mdc]
[ 2759.568607]  [&amp;lt;ffffffffa055ceb8&amp;gt;] mdc_clear_open_replay_data+0x58/0x240 [mdc]
[ 2759.568630]  [&amp;lt;ffffffffa0523fce&amp;gt;] lmv_clear_open_replay_data+0x20e/0x6f0 [lmv]
[ 2759.568655]  [&amp;lt;ffffffffa0b13375&amp;gt;] ll_close_inode_openhandle+0x485/0xfe0 [lustre]
[ 2759.568670]  [&amp;lt;ffffffffa0b1463f&amp;gt;] ll_md_real_close+0x1af/0x220 [lustre]
[ 2759.568685]  [&amp;lt;ffffffffa0b14d15&amp;gt;] ll_file_release+0x665/0xb90 [lustre]
[ 2759.568698]  [&amp;lt;ffffffffa0afbaab&amp;gt;] ll_dir_release+0xdb/0xf0 [lustre]
[ 2759.568703]  [&amp;lt;ffffffff811a4318&amp;gt;] __fput+0x108/0x230
[ 2759.568706]  [&amp;lt;ffffffff811a4465&amp;gt;] fput+0x25/0x30
[ 2759.568708]  [&amp;lt;ffffffff8119f64d&amp;gt;] filp_close+0x5d/0x90
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;This easiest way to see it is to inject a failure to allocate op_data in ll_close_inode_openhandle(). But early returns from lmv_close() or mdc_close() will likely trigger the same assertion.&lt;/p&gt;</description>
                <environment></environment>
        <key id="25402">LU-5282</key>
            <summary>(mdc_request.c:791:mdc_free_open()) ASSERTION( mod-&gt;mod_open_req-&gt;rq_replay == 0 ) failed</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="jhammond">John Hammond</reporter>
                        <labels>
                            <label>llite</label>
                            <label>malloc</label>
                            <label>mdc</label>
                            <label>patch</label>
                    </labels>
                <created>Tue, 1 Jul 2014 22:19:08 +0000</created>
                <updated>Fri, 4 Aug 2017 19:35:08 +0000</updated>
                            <resolved>Fri, 8 Jul 2016 20:27:45 +0000</resolved>
                                    <version>Lustre 2.6.0</version>
                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>10</watches>
                                                                            <comments>
                            <comment id="87993" author="adilger" created="Wed, 2 Jul 2014 18:06:17 +0000"  >&lt;p&gt;John, how did you hit this originally, and how easily is it reproduced?  Just trying to decide if this is a regression that needs to be addressed for 2.6.0?&lt;/p&gt;</comment>
                            <comment id="87995" author="jhammond" created="Wed, 2 Jul 2014 18:13:11 +0000"  >&lt;p&gt;After &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5274&quot; title=&quot;ll_statahead_thread() may leak parent, uses parent after dput&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5274&quot;&gt;&lt;del&gt;LU-5274&lt;/del&gt;&lt;/a&gt; I dug up my OBD_ALLOC() fault injection patch and ran racer with it applied.&lt;/p&gt;</comment>
                            <comment id="133510" author="amk" created="Fri, 13 Nov 2015 20:53:02 +0000"  >&lt;p&gt;A couple of Cray customers have hit this assertion using Lustre 2.5. One case suggests that the bug may be reproduced by running du against a large file system. The du causes Lustre to run the node out of memory (millions of unused ldlm_locks being canceled). An ldlm_bl thread calls into mdc_close but cannot allocate memory for the rpc. The mdc_close does not reset the rq_replay flag, and the subsequent call to mdc_free_open asserts. The ldlm_locks were held primarily by 2 processes, one of which had finished by the time of the LBUG. So there may have been at least 2 processes running du that triggered the LBUG.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&amp;gt;c0-0c1s1n2 LustreError: 28818:0:(file.c:172:ll_close_inode_openhandle()) inode 4035227916085244137 mdc close failed: rc = -12
&amp;gt;c0-0c1s1n2 LustreError: 28818:0:(mdc_request.c:816:mdc_free_open()) ASSERTION( mod-&amp;gt;mod_open_req-&amp;gt;rq_replay == 0 ) failed:
&amp;gt; c0-0c1s1n2 LustreError: 28818:0:(mdc_request.c:816:mdc_free_open()) LBUG
&amp;gt; c0-0c1s1n2 Pid: 28818, comm: ldlm_bl_51
&amp;gt; c0-0c1s1n2 Call Trace:
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffff81005f31&amp;gt;] try_stack_unwind+0x191/0x1a0
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffff810047eb&amp;gt;] dump_trace+0x8b/0x350
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa059e813&amp;gt;] libcfs_debug_dumpstack+0x53/0x80 [libcfs]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa059eda5&amp;gt;] lbug_with_loc+0x45/0xc0 [libcfs]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa0b5912a&amp;gt;] mdc_free_open+0xda/0x110 [mdc]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa0b5b9e8&amp;gt;] mdc_clear_open_replay_data+0x58/0x280 [mdc]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa0269c3d&amp;gt;] lmv_clear_open_replay_data+0x17d/0x690 [lmv]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa0bb3a57&amp;gt;] ll_close_inode_openhandle+0x437/0xf20 [lustre]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa0bb5d0a&amp;gt;] ll_md_real_close+0x11a/0x210 [lustre]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa0be5635&amp;gt;] ll_md_blocking_ast+0x665/0x780 [lustre]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa09355cd&amp;gt;] ldlm_cancel_callback+0x6d/0x150 [ptlrpc]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa0940ca0&amp;gt;] ldlm_cli_cancel_local+0xa0/0x420 [ptlrpc]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa094575a&amp;gt;] ldlm_cli_cancel_list_local+0xea/0x280 [ptlrpc]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffffa094abe7&amp;gt;] ldlm_bl_thread_main+0xd7/0x3f0 [ptlrpc]
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffff81069a70&amp;gt;] kthread+0xc0/0xd0
&amp;gt; c0-0c1s1n2 [&amp;lt;ffffffff815a0a88&amp;gt;] ret_from_fork+0x58/0x90

&amp;gt; PID: 11210  TASK: ffff880c5835f200  CPU: 4   COMMAND: &quot;du&quot;
&amp;gt;  #0 [ffff880cfba2bbc8] schedule at ffffffff81595e45
&amp;gt;  #1 [ffff880cfba2bc48] cl_lock_state_wait at ffffffffa0825025 [obdclass]
&amp;gt;  #2 [ffff880cfba2bcc0] cl_enqueue_locked at ffffffffa0829003 [obdclass]
&amp;gt;  #3 [ffff880cfba2bcf8] cl_lock_request at ffffffffa0829d9b [obdclass]
&amp;gt;  #4 [ffff880cfba2bd50] cl_glimpse_lock at ffffffffa0c03785 [lustre]
&amp;gt;  #5 [ffff880cfba2bda0] cl_glimpse_size0 at ffffffffa0c03c25 [lustre]
&amp;gt;  #6 [ffff880cfba2bde0] ll_inode_revalidate_it at ffffffffa0bbf416 [lustre]
&amp;gt;  #7 [ffff880cfba2be10] ll_getattr_it at ffffffffa0bbf496 [lustre]
&amp;gt;  #8 [ffff880cfba2be40] ll_getattr at ffffffffa0bbf5d4 [lustre]
&amp;gt;  #9 [ffff880cfba2be98] vfs_getattr_nosec at ffffffff8117f4a9
&amp;gt; #10 [ffff880cfba2bea8] vfs_fstatat at ffffffff8117f622
&amp;gt; #11 [ffff880cfba2bef0] SYSC_newfstatat at ffffffff8117fb15
&amp;gt; #12 [ffff880cfba2bf70] sys_newfstatat at ffffffff8117fcfe
&amp;gt; #13 [ffff880cfba2bf80] system_call_fastpath at ffffffff815a0b32
&amp;gt;     RIP: 00007f1ab40b383a  RSP: 00007ffc0061e630  RFLAGS: 00000246
&amp;gt;     RAX: 0000000000000106  RBX: ffffffff815a0b32  RCX: 0000000000000000
&amp;gt;     RDX: 0000000000e695e8  RSI: 0000000000e69678  RDI: 0000000000000005
&amp;gt;     RBP: 0000000000e695e8   R8: 0000000000000100   R9: 0000000000000002
&amp;gt;     R10: 0000000000000100  R11: 0000000000000246  R12: ffffffff8117fcfe
&amp;gt;     R13: ffff880cfba2bf78  R14: 0000000000e69570  R15: 0000000000000000
&amp;gt;     ORIG_RAX: 0000000000000106  CS: 0033  SS: 002b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="135067" author="gerrit" created="Thu, 3 Dec 2015 06:59:59 +0000"  >&lt;p&gt;Alexander Boyko (alexander.boyko@seagate.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/17442&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17442&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5282&quot; title=&quot;(mdc_request.c:791:mdc_free_open()) ASSERTION( mod-&amp;gt;mod_open_req-&amp;gt;rq_replay == 0 ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5282&quot;&gt;&lt;del&gt;LU-5282&lt;/del&gt;&lt;/a&gt; test: check error during mdc_close&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 5f71f93bac5472e2835ee1544770c243de1f95a0&lt;/p&gt;</comment>
                            <comment id="135383" author="pjones" created="Mon, 7 Dec 2015 13:48:06 +0000"  >&lt;p&gt;Fix pushed to gerrit - &lt;a href=&quot;http://review.whamcloud.com/#/c/17495/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/17495/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="136837" author="aboyko" created="Fri, 18 Dec 2015 10:49:20 +0000"  >&lt;p&gt;The patch change the assertion to warning. It is not a fatal error if client doesn`t send a close request to MDS. In worst case MDS evict client and client doesn`t send cookies to OST. With reproducer I see that mdc export was not fully released.&lt;br/&gt;
18 ST mdc lustre-MDT0000-mdc-ffff88005be2d800 18902755-d63e-aa43-5b3f-5e81f3be1b56 2&lt;/p&gt;</comment>
                            <comment id="157726" author="gerrit" created="Tue, 5 Jul 2016 23:48:16 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/17495/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/17495/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5282&quot; title=&quot;(mdc_request.c:791:mdc_free_open()) ASSERTION( mod-&amp;gt;mod_open_req-&amp;gt;rq_replay == 0 ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5282&quot;&gt;&lt;del&gt;LU-5282&lt;/del&gt;&lt;/a&gt; mdc: fix panic at mdc_free_open()&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 62713a8530349a75b7202c9bfc6be121409a0203&lt;/p&gt;</comment>
                            <comment id="158170" author="jgmitter" created="Fri, 8 Jul 2016 20:27:45 +0000"  >&lt;p&gt;Landed to master for 2.9.0&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="15283" name="alloc-fail.patch" size="4459" author="jhammond" created="Wed, 2 Jul 2014 18:13:26 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwqen:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>14743</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>