<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:47:26 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4971] sanity-scrub test_2: ldlm_lock2desc()) ASSERTION( lock-&gt;l_policy_data.l_inodebits.bits == (MDS_INODELOCK_LOOKUP | MDS_INODELOCK_UPDATE | MDS_INODELOCK_LAYOUT)failed</title>
                <link>https://jira.whamcloud.com/browse/LU-4971</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Nathaniel Clark &amp;lt;nathaniel.l.clark@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run:&lt;br/&gt;
&lt;a href=&quot;http://maloo.whamcloud.com/test_sets/598caeaa-cd2c-11e3-b548-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://maloo.whamcloud.com/test_sets/598caeaa-cd2c-11e3-b548-52540035b04c&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/fc94ece0-a552-11e3-9fee-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/fc94ece0-a552-11e3-9fee-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The sub-test test_2 failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;test failed to respond and timed out&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Info required for matching: sanity-scrub 2&lt;/p&gt;</description>
                <environment></environment>
        <key id="24438">LU-4971</key>
            <summary>sanity-scrub test_2: ldlm_lock2desc()) ASSERTION( lock-&gt;l_policy_data.l_inodebits.bits == (MDS_INODELOCK_LOOKUP | MDS_INODELOCK_UPDATE | MDS_INODELOCK_LAYOUT)failed</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="yong.fan">nasf</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Mon, 28 Apr 2014 18:52:45 +0000</created>
                <updated>Wed, 8 Feb 2017 16:54:53 +0000</updated>
                            <resolved>Thu, 10 Jul 2014 23:52:41 +0000</resolved>
                                    <version>Lustre 2.6.0</version>
                    <version>Lustre 2.5.4</version>
                                    <fixVersion>Lustre 2.6.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="82710" author="yong.fan" created="Tue, 29 Apr 2014 01:15:27 +0000"  >&lt;p&gt;Some log on the client side:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;23:56:52:LustreError: 23554:0:(ldlm_lock.c:671:ldlm_lock2desc()) ASSERTION( lock-&amp;gt;l_policy_data.l_inodebits.bits == (MDS_INODELOCK_LOOKUP | MDS_INODELOCK_UPDATE | MDS_INODELOCK_LAYOUT) ) failed: Inappropriate inode lock bits during conversion 3
23:56:52:LustreError: 23554:0:(ldlm_lock.c:671:ldlm_lock2desc()) LBUG
23:56:52:Pid: 23554, comm: diff
23:56:52:
23:56:52:Call Trace:
23:56:52: [&amp;lt;ffffffffa03d0895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
23:56:52: [&amp;lt;ffffffffa03d0e97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
23:56:52: [&amp;lt;ffffffffa08dc239&amp;gt;] ldlm_lock2desc+0x179/0x180 [ptlrpc]
23:56:52: [&amp;lt;ffffffffa08eed90&amp;gt;] ldlm_cli_enqueue+0x1f0/0x790 [ptlrpc]
23:56:52: [&amp;lt;ffffffffa0914c0a&amp;gt;] ? ptlrpc_request_set_replen+0x3a/0x60 [ptlrpc]
23:56:52: [&amp;lt;ffffffffa08f3bd0&amp;gt;] ? ldlm_completion_ast+0x0/0x930 [ptlrpc]
23:56:52: [&amp;lt;ffffffffa13434d0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7f0 [lustre]
23:56:52: [&amp;lt;ffffffffa0a3d88e&amp;gt;] mdc_enqueue+0x2be/0x1b00 [mdc]
23:56:52: [&amp;lt;ffffffffa08e25e5&amp;gt;] ? ldlm_lock_match+0x215/0x8b0 [ptlrpc]
23:56:52: [&amp;lt;ffffffffa0a3f2ce&amp;gt;] mdc_intent_lock+0x1fe/0x63f [mdc]
23:56:52: [&amp;lt;ffffffffa13434d0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7f0 [lustre]
23:56:52: [&amp;lt;ffffffffa08f3bd0&amp;gt;] ? ldlm_completion_ast+0x0/0x930 [ptlrpc]
23:56:52: [&amp;lt;ffffffffa053ba28&amp;gt;] ? lprocfs_counter_add+0x1a8/0x1c0 [obdclass]
23:56:52: [&amp;lt;ffffffffa082eae1&amp;gt;] lmv_intent_remote+0x481/0xa90 [lmv]
23:56:52: [&amp;lt;ffffffffa13434d0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7f0 [lustre]
23:56:52: [&amp;lt;ffffffffa082fd49&amp;gt;] lmv_intent_lookup+0xc59/0xd00 [lmv]
23:56:52: [&amp;lt;ffffffffa13434d0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7f0 [lustre]
23:56:52: [&amp;lt;ffffffffa03d027b&amp;gt;] ? cfs_set_ptldebug_header+0x2b/0xc0 [libcfs]
23:56:52: [&amp;lt;ffffffffa0830ada&amp;gt;] lmv_intent_lock+0x32a/0x380 [lmv]
23:56:52: [&amp;lt;ffffffffa13434d0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7f0 [lustre]
23:56:52: [&amp;lt;ffffffffa1325ed8&amp;gt;] ? ll_prep_md_op_data+0x1a8/0x490 [lustre]
23:56:52: [&amp;lt;ffffffffa1344d79&amp;gt;] ll_lookup_it+0x299/0xb50 [lustre]
23:56:52: [&amp;lt;ffffffffa13434d0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7f0 [lustre]
23:56:52: [&amp;lt;ffffffffa134589f&amp;gt;] ll_lookup_nd+0x26f/0x3e0 [lustre]
23:56:52: [&amp;lt;ffffffff811a394e&amp;gt;] ? d_alloc+0x13e/0x1b0
23:56:52: [&amp;lt;ffffffff81198a25&amp;gt;] do_lookup+0x1a5/0x230
23:56:52: [&amp;lt;ffffffff81198db0&amp;gt;] __link_path_walk+0x200/0xff0
23:56:52: [&amp;lt;ffffffff8114a647&amp;gt;] ? handle_pte_fault+0xf7/0xb00
23:56:52: [&amp;lt;ffffffff81199e5a&amp;gt;] path_walk+0x6a/0xe0
23:56:52: [&amp;lt;ffffffff8119a06b&amp;gt;] filename_lookup+0x6b/0xc0
23:56:52: [&amp;lt;ffffffff81196996&amp;gt;] ? final_putname+0x26/0x50
23:56:52: [&amp;lt;ffffffff8119b197&amp;gt;] user_path_at+0x57/0xa0
23:56:52: [&amp;lt;ffffffff8104a98c&amp;gt;] ? __do_page_fault+0x1ec/0x480
23:56:52: [&amp;lt;ffffffff812826f5&amp;gt;] ? _atomic_dec_and_lock+0x55/0x80
23:56:52: [&amp;lt;ffffffff811aaa10&amp;gt;] ? mntput_no_expire+0x30/0x110
23:56:52: [&amp;lt;ffffffff8118e7a4&amp;gt;] ? cp_new_stat+0xe4/0x100
23:56:52: [&amp;lt;ffffffff8118e9e0&amp;gt;] vfs_fstatat+0x50/0xa0
23:56:52: [&amp;lt;ffffffff8118eb5b&amp;gt;] vfs_stat+0x1b/0x20
23:56:52: [&amp;lt;ffffffff8118eb84&amp;gt;] sys_newstat+0x24/0x50
23:56:52: [&amp;lt;ffffffff810e2057&amp;gt;] ? audit_syscall_entry+0x1d7/0x200
23:56:52: [&amp;lt;ffffffff810e1e4e&amp;gt;] ? __audit_syscall_exit+0x25e/0x290
23:56:52: [&amp;lt;ffffffff8100b072&amp;gt;] system_call_fastpath+0x16/0x1b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="82765" author="green" created="Tue, 29 Apr 2014 17:34:54 +0000"  >&lt;p&gt;So it appears tha client side exp_connect_flags is somehow corrupted because we had no servers not supporting inode bits for many years now. Not since 1.4.x I suspect.&lt;/p&gt;</comment>
                            <comment id="82909" author="adilger" created="Wed, 30 Apr 2014 19:58:33 +0000"  >&lt;p&gt;It is strange that the LASSERT() is on the client, but LFSCK is running on the server (only some parts of the test scripts are setting up the environment on the client).&lt;/p&gt;

&lt;p&gt;Oleg notes that this problem indicates that the client is trying to use an IBITS lock (as it should) but it thinks the server does not support this (due to missing connect feature flags).  That might indicate some problem or corruption with the connection state.&lt;/p&gt;</comment>
                            <comment id="82937" author="yong.fan" created="Wed, 30 Apr 2014 22:06:48 +0000"  >&lt;p&gt;Yes, I also suspected that it should not related with OI scrub, instead, might because of some connection race cases. I remembered that we have met similar connection flags issues about client_is_remote() before at LLNL site. But I cannot remember the detail...&lt;/p&gt;</comment>
                            <comment id="87654" author="hongchao.zhang" created="Fri, 27 Jun 2014 04:24:01 +0000"  >&lt;p&gt;&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/099fd572-fd95-11e3-b84d-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/099fd572-fd95-11e3-b84d-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="87831" author="jamesanunez" created="Mon, 30 Jun 2014 20:16:18 +0000"  >&lt;p&gt;sanity-scrub test 4 is displaying this error and crashing the client node.&lt;/p&gt;

&lt;p&gt;I&apos;m not sure why so few logs were produced, but here are the logs from the client node: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/77df2c0e-0092-11e4-a42b-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/77df2c0e-0092-11e4-a42b-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="88028" author="jamesanunez" created="Wed, 2 Jul 2014 20:35:28 +0000"  >&lt;p&gt;I&apos;ve run sanity-scrub four times on the OpenSFS cluster and sanity-scrub fails with this error each time in test 2, 4 or 5. &lt;/p&gt;

&lt;p&gt;I&apos;ve run with one, two and four clients with two MDSs each with two MDTs and four OSSs with two OSTs each with Lustre 2.5.60 build # 2535.&lt;/p&gt;</comment>
                            <comment id="88070" author="yong.fan" created="Thu, 3 Jul 2014 04:38:21 +0000"  >&lt;p&gt;A debug patch to check the export status when assertion:&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/#/c/10958/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/10958/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="88133" author="jamesanunez" created="Thu, 3 Jul 2014 18:39:36 +0000"  >&lt;p&gt;I&apos;ve run sanity-scrub twice with the debug patch and only hit this error one of the runs. Here is the output from the debug message for a single client run that failed in sanity-scrub test 2:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;...
restore data
restore EA
/usr/lib64/lustre/tests
remove recovery logs
removed `/lustre/brpt/CATALOGS&apos;
starting MDTs without disabling OI scrub
Starting client: c17: -o user_xattr,flock mds01@o2ib:/scratch /lustre/scratch

Message from syslogd@c17 at Jul  3 11:36:24 ...
 kernel:LustreError: 22841:0:(ldlm_lock.c:672:ldlm_lock2desc()) ASSERTION( lock-&amp;gt;l_policy_data.l_inodebits.bits == (MDS_INODELOCK_LOOKUP | MDS_INODELOCK_UPDATE | MDS_INODELOCK_LAYOUT) ) failed: Inappropriate inode lock bits during conversion, bits 0x3, exp ffff880827d6dc00, flags 0x0

Message from syslogd@c17 at Jul  3 11:36:24 ...
 kernel:LustreError: 22841:0:(ldlm_lock.c:672:ldlm_lock2desc()) LBUG
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="88179" author="adilger" created="Fri, 4 Jul 2014 05:46:35 +0000"  >&lt;p&gt;Hmm, it has bits 0x3 set, so it is missing the LAYOUT flag. That would be possible if the file is being migrated, or maybe if it is a remote MDT object?  I suspect that the LASSERT is wrong here, but I&apos;m not positive yet as I haven&apos;t looked closely at the surrounding code. &lt;/p&gt;</comment>
                            <comment id="88180" author="yong.fan" created="Fri, 4 Jul 2014 05:57:02 +0000"  >&lt;p&gt;The bits may be right, because the former code cannot find OBD_CONNECT_IBITS in the connection_flags, so the layout ibits will be dropped automatically.&lt;/p&gt;

&lt;p&gt;Here we get the connection_flags as &quot;0x0&quot;, that is impossible for any connection. So there should be data corruption.&lt;/p&gt;</comment>
                            <comment id="88207" author="adilger" created="Fri, 4 Jul 2014 18:03:45 +0000"  >&lt;p&gt;Since this only happens during sanity-scrub, I expect there is some kind of memory corruption or use-after-free problem in the code?  It might be useful to run sanity-scrub on a kernel with CONFIG_DEBUG_SLAB and other memory debugging features (CONFIG_DEBUG_SPINLOCK, CONFIG_DEBUG_OBJECTS, CONFIG_DEBUG_KMEMLEAK, CONFIG_DEBUG_MUTEXES, CONFIG_DEBUG_KOBJECT) enabled to see if this triggers a problem.&lt;/p&gt;</comment>
                            <comment id="88223" author="yong.fan" created="Sat, 5 Jul 2014 03:01:50 +0000"  >&lt;p&gt;I have run related tests locally for hundreds of times, but cannot reproduce the failure by myself. So I have to go though all related code. Finally, I found that: it may be not related with data corruption, but more seems race condition between the asynchronous connection and synchronous lock_enqueue operation. Consider the following scenario:&lt;/p&gt;

&lt;p&gt;1) There are two MDTs in the system, when the client mount, it will trigger MDS_CONNECT request to each of the MDT. But the connect request will be handled asynchronously. The mount processing will go ahead when the connection between the client and MDT0 is established, but will not wait for the connection between the client and the MDT1 to be established.&lt;/p&gt;

&lt;p&gt;2) After the client mount returned, the application tries to &quot;stat $MOUNT/a/b&quot;, the directory &quot;a&quot; resides on the MDT0, the &quot;b&quot; resides on the MDT1. So the client will send lock_enqueue RPC to the MDT0 firstly.&lt;/p&gt;

&lt;p&gt;3) The RPC handler on the MDT0 finds that the target object &quot;b&quot; resides on the remote MDT1, then it replies the client to try to talk with the MDT1.&lt;/p&gt;

&lt;p&gt;4) The client gets the MDT0&apos;s reply, then calls lmv_intent_remote() to forward the request to the MDT1 via MDC1, but at that time, the connection between the client and the MDT1 has not been established yet.&lt;/p&gt;

&lt;p&gt;5) For most of other RPCs, above case is not a problem, because the client-side RPC sender (in ptlrpc) will handle kinds of connection issues. But for the ldlm_enqueue, it is some different, because before giving the ldlm_enqueue RPC to the sender (or to the ptlrpc), the sponsor will check the export connect flags that will not be set until the connection between the client and the MDT has been established. So at such time point, the sponsor of the ldlm_enqueue request will get &quot;0&quot; flags from the export connection flags, then trigger the ASSERTION as described in this ticket.&lt;/p&gt;


&lt;p&gt;So it is not a special bug for OI scrub, but more general issue for the whole Lustre framework. Because the sanity-scrub will access some cross-MDTs objects immediately after the client mount (without any wait), then it is more easy to reproduce the failure.&lt;/p&gt;

&lt;p&gt;About the fixing, I do not want to change the general connection/ldlm framework, because I am afraid of introducing some potential bugs. Instead, the patch will make the ldlm_request request to wait the connection to be established for such race case.&lt;/p&gt;</comment>
                            <comment id="88224" author="yong.fan" created="Sat, 5 Jul 2014 03:03:14 +0000"  >&lt;p&gt;James, would you please to verify the new patch? Thanks!&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/#/c/10958/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/10958/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="88226" author="di.wang" created="Sat, 5 Jul 2014 06:58:47 +0000"  >&lt;p&gt;Hmm, I think this is the problem for single MDT FS too, i.e. ldlm_cli_enqueue should not use exp connection flag before it make sure the connection is established?&lt;br/&gt;
This inodebit connect flag is being added even before 1.6 I guess?  At least this inodebit interop code is added in 2007&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;commit 113303973ec9f8484eb2355a1a6ef3c4c7fd6a56
Author: nathan &amp;lt;nathan&amp;gt;
Date:   Sat Feb 10 06:33:41 2007 +0000

    land b1_5 onto HEAD
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Which I thought it use exp connection flag in a wrong way.  So we probably just remove these code to save some trouble here. Since I do not think we plan to let 2.6 lustre client to connect some non-inodebit support server (1.4 ?) &lt;/p&gt;</comment>
                            <comment id="88228" author="yong.fan" created="Sat, 5 Jul 2014 12:56:09 +0000"  >&lt;p&gt;No, for single MDT case, there is no such race condition. Because the mount process will call md_getstatus() on the ROOT object firstly, it is NOT ldlm_enqueue RPC. Then when the mount return to user-space, it can guarantee that the connection between the client and the MDT0 has been established successfully.&lt;/p&gt;</comment>
                            <comment id="88377" author="di.wang" created="Mon, 7 Jul 2014 22:21:52 +0000"  >&lt;p&gt;Ah, I see. Hmm, I still think remove those incompatible inodebit handling is a better choice, or can we do sth like this&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;root@testnode lustre-release]# git diff lustre/ldlm/
diff --git a/lustre/ldlm/ldlm_request.c b/lustre/ldlm/ldlm_request.c
index 76f329d..2c0d05d 100644
--- a/lustre/ldlm/ldlm_request.c
+++ b/lustre/ldlm/ldlm_request.c
@@ -909,7 +909,7 @@ int ldlm_cli_enqueue(struct obd_export *exp, struct ptlrpc_request **reqp,
                              OBD_CONNECT_IBITS))
                                 lock-&amp;gt;l_policy_data.l_inodebits.bits =
                                         MDS_INODELOCK_LOOKUP |
-                                        MDS_INODELOCK_UPDATE;
+                                        MDS_INODELOCK_UPDATE | MDS_INODELOCK_LAYOUT;
                         else
                                 lock-&amp;gt;l_policy_data = *policy;
                 }
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
</comment>
                            <comment id="88379" author="yong.fan" created="Mon, 7 Jul 2014 22:37:41 +0000"  >&lt;p&gt;Wangdi, I have verified the race conditions locally. But I am not sure whether removing the incompatible bits is a suitable solution or not. Because the key issue is that we are checking un-iniitialized export flags.&lt;/p&gt;</comment>
                            <comment id="88380" author="di.wang" created="Mon, 7 Jul 2014 22:43:21 +0000"  >&lt;blockquote&gt;
&lt;p&gt;Because the key issue is that we are checking un-iniitialized export flags.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;Yes, but that is for checking if the server support inodebit lock. But since 2.6 client will be only connected to inodebit enable server, so why do we need this check? TBH, I really do not like add &quot;connection wait logic&quot; above ptlrpc, which might bring us troubles or even worse.  &lt;/p&gt;</comment>
                            <comment id="88381" author="yong.fan" created="Mon, 7 Jul 2014 22:48:00 +0000"  >&lt;p&gt;Even though we drop the incompatible ibits, we still needs to prevent the user to use the non-initialised export, otherwise, there may be other potential bugs.&lt;/p&gt;</comment>
                            <comment id="88383" author="jamesanunez" created="Mon, 7 Jul 2014 23:09:35 +0000"  >&lt;p&gt;I&apos;ve tested the patch at &lt;a href=&quot;http://review.whamcloud.com/#/c/10958/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/10958/&lt;/a&gt; (version 2) and, after running sanity-scrub three times, I have not seen this assertion.&lt;/p&gt;

&lt;p&gt;The patch fixes the crash/error for the configuration I&apos;ve been testing with.&lt;/p&gt;</comment>
                            <comment id="88384" author="yong.fan" created="Mon, 7 Jul 2014 23:13:35 +0000"  >&lt;p&gt;Thanks James!&lt;/p&gt;</comment>
                            <comment id="88386" author="di.wang" created="Mon, 7 Jul 2014 23:41:07 +0000"  >&lt;p&gt;&quot;Even though we drop the incompatible ibits, we still needs to prevent the user to use the non-initialised export, otherwise, there may be other potential bugs.&quot;&lt;/p&gt;

&lt;p&gt;That is something should never happen, IMHO.&lt;/p&gt;</comment>
                            <comment id="88394" author="yong.fan" created="Tue, 8 Jul 2014 01:54:56 +0000"  >&lt;p&gt;Here is the patch to drop the redundant ibits lock interoperability check:&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/11004&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/11004&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="88786" author="yong.fan" created="Thu, 10 Jul 2014 23:52:41 +0000"  >&lt;p&gt;The patch has been landed to master.&lt;/p&gt;</comment>
                            <comment id="100753" author="yujian" created="Thu, 4 Dec 2014 20:39:12 +0000"  >&lt;p&gt;While verifying patch &lt;a href=&quot;http://review.whamcloud.com/12606&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/12606&lt;/a&gt; on Lustre b2_5 branch, the same failure occurred:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LustreError: 19407:0:(ldlm_lock.c:669:ldlm_lock2desc()) ASSERTION( lock-&amp;gt;l_policy_data.l_inodebits.bits == (MDS_INODELOCK_LOOKUP | MDS_INODELOCK_UPDATE | MDS_INODELOCK_LAYOUT) ) failed: Inappropriate inode lock bits during conversion 3
LustreError: 19407:0:(ldlm_lock.c:669:ldlm_lock2desc()) LBUG
Pid: 19407, comm: stat

Call Trace:
 [&amp;lt;ffffffffa03d0895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs] 
 [&amp;lt;ffffffffa03d0e97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs] 
 [&amp;lt;ffffffffa067a239&amp;gt;] ldlm_lock2desc+0x179/0x180 [ptlrpc] 
 [&amp;lt;ffffffffa068cb90&amp;gt;] ldlm_cli_enqueue+0x1f0/0x790 [ptlrpc] 
 [&amp;lt;ffffffffa06b29ea&amp;gt;] ? ptlrpc_request_set_replen+0x3a/0x60 [ptlrpc] 
 [&amp;lt;ffffffffa06919d0&amp;gt;] ? ldlm_completion_ast+0x0/0x920 [ptlrpc] 
 [&amp;lt;ffffffffa0a825c0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7d0 [lustre] 
 [&amp;lt;ffffffffa08ded8e&amp;gt;] mdc_enqueue+0x2be/0x1a10 [mdc]
 [&amp;lt;ffffffffa01bc294&amp;gt;] ? fld_client_rpc+0x864/0xed0 [fld]
 [&amp;lt;ffffffffa08e06dd&amp;gt;] mdc_intent_lock+0x1fd/0x64a [mdc]
 [&amp;lt;ffffffffa0a825c0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7d0 [lustre] 
 [&amp;lt;ffffffffa06919d0&amp;gt;] ? ldlm_completion_ast+0x0/0x920 [ptlrpc] 
 [&amp;lt;ffffffffa00e58b8&amp;gt;] ? lprocfs_counter_add+0x1a8/0x1d6 [lvfs]
 [&amp;lt;ffffffffa08a846e&amp;gt;] lmv_intent_remote+0x47e/0xa80 [lmv]
 [&amp;lt;ffffffffa0a825c0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7d0 [lustre] 
 [&amp;lt;ffffffffa08a9137&amp;gt;] lmv_intent_lookup+0x6c7/0x700 [lmv]
 [&amp;lt;ffffffffa0a825c0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7d0 [lustre] 
 [&amp;lt;ffffffffa08a9d6a&amp;gt;] lmv_intent_lock+0x32a/0x380 [lmv]
 [&amp;lt;ffffffffa0a825c0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7d0 [lustre] 
 [&amp;lt;ffffffffa0a81d0e&amp;gt;] ? ll_i2gids+0x2e/0xd0 [lustre] 
 [&amp;lt;ffffffffa0a68c0d&amp;gt;] ? ll_prep_md_op_data+0x10d/0x3b0 [lustre] 
 [&amp;lt;ffffffffa0a8491f&amp;gt;] ll_lookup_it+0x33f/0xb00 [lustre] 
 [&amp;lt;ffffffffa0a825c0&amp;gt;] ? ll_md_blocking_ast+0x0/0x7d0 [lustre] 
 [&amp;lt;ffffffffa0a50c51&amp;gt;] ? __ll_inode_revalidate_it+0x1e1/0xc30 [lustre] 
 [&amp;lt;ffffffffa0a8535f&amp;gt;] ll_lookup_nd+0x27f/0x3f0 [lustre] 
 [&amp;lt;ffffffff811a42fe&amp;gt;] ? d_alloc+0x13e/0x1b0
 [&amp;lt;ffffffff81198a35&amp;gt;] do_lookup+0x1a5/0x230
 [&amp;lt;ffffffff81199100&amp;gt;] __link_path_walk+0x200/0x1000
 [&amp;lt;ffffffff8114a3d7&amp;gt;] ? handle_pte_fault+0xf7/0xb00
 [&amp;lt;ffffffff8119a1ba&amp;gt;] path_walk+0x6a/0xe0
 [&amp;lt;ffffffff8119a3cb&amp;gt;] filename_lookup+0x6b/0xc0
 [&amp;lt;ffffffff8119b4f7&amp;gt;] user_path_at+0x57/0xa0
 [&amp;lt;ffffffff8104a98c&amp;gt;] ? __do_page_fault+0x1ec/0x480
 [&amp;lt;ffffffff8118e990&amp;gt;] vfs_fstatat+0x50/0xa0
 [&amp;lt;ffffffff811515a5&amp;gt;] ? do_mmap_pgoff+0x335/0x380
 [&amp;lt;ffffffff8118ea4e&amp;gt;] vfs_lstat+0x1e/0x20
 [&amp;lt;ffffffff8118ea74&amp;gt;] sys_newlstat+0x24/0x50
 [&amp;lt;ffffffff810e1e07&amp;gt;] ? audit_syscall_entry+0x1d7/0x200
 [&amp;lt;ffffffff8100b072&amp;gt;] system_call_fastpath+0x16/0x1b

Kernel panic - not syncing: LBUG
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Maloo report: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/f4332d04-7af0-11e4-956d-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/f4332d04-7af0-11e4-956d-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hi Nasf, could you please take a look whether this is a regression introduced by the patch &lt;a href=&quot;http://review.whamcloud.com/12606&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/12606&lt;/a&gt; or not? If not, then this is an issue on Lustre b2_5 branch and I&apos;ll back-port your patch to fix it. Thank you!&lt;/p&gt;</comment>
                            <comment id="100779" author="yong.fan" created="Thu, 4 Dec 2014 23:25:15 +0000"  >&lt;p&gt;Yujian,&lt;/p&gt;

&lt;p&gt;I do not think it is &lt;a href=&quot;http://review.whamcloud.com/#/c/12606/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/12606/&lt;/a&gt; caused the trouble. You need to back-port patch(es) to fix it.&lt;/p&gt;</comment>
                            <comment id="100854" author="yujian" created="Fri, 5 Dec 2014 18:32:42 +0000"  >&lt;p&gt;Thank you, Nasf. Will do.&lt;/p&gt;</comment>
                            <comment id="100863" author="yujian" created="Fri, 5 Dec 2014 19:29:20 +0000"  >&lt;p&gt;One more instance on Lustre b2_5 branch:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/c800bb44-5e7f-11e4-9843-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/c800bb44-5e7f-11e4-9843-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="100865" author="gerrit" created="Fri, 5 Dec 2014 19:44:04 +0000"  >&lt;p&gt;Jian Yu (jian.yu@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/12960&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/12960&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4971&quot; title=&quot;sanity-scrub test_2: ldlm_lock2desc()) ASSERTION( lock-&amp;gt;l_policy_data.l_inodebits.bits == (MDS_INODELOCK_LOOKUP | MDS_INODELOCK_UPDATE | MDS_INODELOCK_LAYOUT)failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4971&quot;&gt;&lt;del&gt;LU-4971&lt;/del&gt;&lt;/a&gt; ldlm: drop redundant ibits lock interoperability check&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_5&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 2b44f94c2bfbefd9b0f85a52218979b851d7af58&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="25367">LU-5273</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="25366">LU-5272</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="35779">LU-7975</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwla7:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>13763</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>