<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:36:24 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-3727] LBUG (llite_nfs.c:281:ll_get_parent()) ASSERTION(body-&gt;valid &amp; OBD_MD_FLID) failed</title>
                <link>https://jira.whamcloud.com/browse/LU-3727</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;At GE Global Research, we ran into an LBUG with a 1.8.9 client that is re-exporting 2.1.5 Lustre:&lt;/p&gt;

&lt;p&gt;Jul 31 10:26:46 scinfra3 kernel: Installing knfsd (copyright (C) 1996 okir@monad.swb.de).&lt;br/&gt;
Jul 31 10:26:46 scinfra3 kernel: NFSD: Using /var/lib/nfs/v4recovery as the NFSv4 state recovery directory&lt;br/&gt;
Jul 31 10:26:46 scinfra3 kernel: NFSD: starting 90-second grace period&lt;br/&gt;
Jul 31 10:26:53 scinfra3 ntpd&lt;span class=&quot;error&quot;&gt;&amp;#91;8318&amp;#93;&lt;/span&gt;: synchronized to 3.40.208.30, stratum 2&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel: LustreError: 27396:0:(llite_nfs.c:281:ll_get_parent()) ASSERTION(body-&amp;gt;valid &amp;amp; OBD_MD_FLID) failed&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel: LustreError: 27396:0:(llite_nfs.c:281:ll_get_parent()) LBUG&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel: Pid: 27396, comm: nfsd&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel: &lt;br/&gt;
Jul 31 10g:29:46 scinfra3 kernel: Call Trace:&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] libcfs_debug_dumpstack+0x51/0x60 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] lbug_with_loc+0x7a/0xd0 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] tracefile_init+0x0/0x110 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] ll_get_parent+0x1e3/0x2b0 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] ll_get_dentry+0x6b/0xe0 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] mutex_lock+0xd/0x1d&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] find_exported_dentry+0x241/0x486 &lt;span class=&quot;error&quot;&gt;&amp;#91;exportfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd_acceptable+0x0/0xdc &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] autoremove_wake_function+0x0/0x2e&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] sunrpc_cache_lookup+0x4b/0x128 &lt;span class=&quot;error&quot;&gt;&amp;#91;sunrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] exp_get_by_name+0x5b/0x71 &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] exp_find_key+0x89/0x9c &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd_acceptable+0x0/0xdc &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] ll_decode_fh+0x197/0x240 &lt;span class=&quot;error&quot;&gt;&amp;#91;lustre&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] set_current_groups+0x116/0x164&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] fh_verify+0x29c/0x4cf &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd3_proc_getattr+0x8a/0xbe &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd_dispatch+0xd8/0x1d6 &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] svc_process+0x3f8/0x6bf &lt;span class=&quot;error&quot;&gt;&amp;#91;sunrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] __down_read+0x12/0x92&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd+0x0/0x2cb &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd+0x1a5/0x2cb &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] child_rip+0xa/0x11&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd+0x0/0x2cb &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] nfsd+0x0/0x2cb &lt;span class=&quot;error&quot;&gt;&amp;#91;nfsd&amp;#93;&lt;/span&gt;&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel:  [ ] child_rip+0x0/0x11&lt;br/&gt;
Jul 31 10:29:46 scinfra3 kernel: &lt;/p&gt;

&lt;p&gt;It appears to be easily reproducible, we are going to try to get a core dump, but I was wondering if there was anything obvious from this trace or any other jira tickets I might have missed. Also is there any other information that might be useful?&lt;/p&gt;

&lt;p&gt;Thanks.&lt;/p&gt;</description>
                <environment></environment>
        <key id="20245">LU-3727</key>
            <summary>LBUG (llite_nfs.c:281:ll_get_parent()) ASSERTION(body-&gt;valid &amp; OBD_MD_FLID) failed</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="emoly.liu">Emoly Liu</assignee>
                                    <reporter username="orentas">Oz Rentas</reporter>
                        <labels>
                            <label>patch</label>
                    </labels>
                <created>Thu, 8 Aug 2013 17:53:18 +0000</created>
                <updated>Thu, 14 Jun 2018 21:41:36 +0000</updated>
                            <resolved>Tue, 10 Feb 2015 12:39:17 +0000</resolved>
                                    <version>Lustre 2.1.5</version>
                    <version>Lustre 1.8.9</version>
                    <version>Lustre 2.4.1</version>
                                    <fixVersion>Lustre 2.7.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>14</watches>
                                                                            <comments>
                            <comment id="63911" author="pjones" created="Thu, 8 Aug 2013 19:24:17 +0000"  >&lt;p&gt;Thanks for the report Kit!&lt;/p&gt;</comment>
                            <comment id="63988" author="pjones" created="Fri, 9 Aug 2013 17:25:22 +0000"  >&lt;p&gt;Emoly&lt;/p&gt;

&lt;p&gt;What do you suggest here?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="64056" author="emoly.liu" created="Mon, 12 Aug 2013 01:41:26 +0000"  >&lt;p&gt;Kit, could you please show how to reproduce this LBUG in detail? And a core dump file will be helpful. Thanks!&lt;/p&gt;</comment>
                            <comment id="64084" author="kitwestneat" created="Mon, 12 Aug 2013 14:43:05 +0000"  >&lt;p&gt;Hi Emoly,&lt;/p&gt;

&lt;p&gt;The customer is currently unable to reproduce after setting up kdump, so we are in a holding pattern. I will ask what they were doing to reproduce before.&lt;/p&gt;

&lt;p&gt;Thanks. &lt;/p&gt;</comment>
                            <comment id="64205" author="kitwestneat" created="Tue, 13 Aug 2013 21:36:03 +0000"  >&lt;p&gt;Hi Emoly,&lt;/p&gt;

&lt;p&gt;We were able to capture a crash dump from a different client node. Strangely, this client gets a null pointer dereference when trying to print the LBUG stack trace. &lt;/p&gt;

&lt;p&gt;Here is the vmcore:&lt;br/&gt;
&lt;a href=&quot;http://ddntsr.com/ftp/2013-08-13-SR23991-vmcore&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://ddntsr.com/ftp/2013-08-13-SR23991-vmcore&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is the kernel debuginfo for it:&lt;br/&gt;
&lt;a href=&quot;http://debuginfo.centos.org/6/x86_64/kernel-debuginfo-2.6.32-279.2.1.el6.x86_64.rpm&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://debuginfo.centos.org/6/x86_64/kernel-debuginfo-2.6.32-279.2.1.el6.x86_64.rpm&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://debuginfo.centos.org/6/x86_64/kernel-debuginfo-common-x86_64-2.6.32-279.2.1.el6.x86_64.rpm&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://debuginfo.centos.org/6/x86_64/kernel-debuginfo-common-x86_64-2.6.32-279.2.1.el6.x86_64.rpm&lt;/a&gt;&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;crash&amp;gt; bt
PID: 23553  TASK: ffff881019a32080  CPU: 10  COMMAND: &lt;span class=&quot;code-quote&quot;&gt;&quot;nfsd&quot;&lt;/span&gt;
 #0 [ffff88100a241520] machine_kexec at ffffffff8103281b
 #1 [ffff88100a241580] crash_kexec at ffffffff810ba792
 #2 [ffff88100a241650] text_poke at ffffffff815016f0
 #3 [ffff88100a241680] no_context at ffffffff81043bab
 #4 [ffff88100a2416d0] __bad_area_nosemaphore at ffffffff81043e35
 #5 [ffff88100a241720] bad_area_nosemaphore at ffffffff81043f03
 #6 [ffff88100a241730] __do_page_fault at ffffffff81044661
 #7 [ffff88100a241850] debugfs_kprobe_init at ffffffff815036ce
 #8 [ffff88100a241880] do_debug at ffffffff81500a85
 #9 [ffff88100a2419d8] libcfs_debug_dumpstack at ffffffffa01d78f5 [libcfs]
#10 [ffff88100a2419f8] lbug_with_loc at ffffffffa01d7f25 [libcfs]
#11 [ffff88100a241a48] libcfs_assertion_failed at ffffffffa01e0696 [libcfs]
#12 [ffff88100a241a98] ll_get_parent at ffffffffa08836f8 [lustre]
#13 [ffff88100a241b38] reconnect_path at ffffffffa021e3b0 [exportfs]
#14 [ffff88100a241ba8] exportfs_decode_fh at ffffffffa021e7aa [exportfs]
#15 [ffff88100a241d18] fh_verify at ffffffffa064abea [nfsd]
#16 [ffff88100a241da8] nfsd3_proc_getattr at ffffffffa0655b6c [nfsd]
#17 [ffff88100a241dd8] nfsd_dispatch at ffffffffa064743e [nfsd]
#18 [ffff88100a241e18] svc_process_common at ffffffffa05fb5d4 [sunrpc]
#19 [ffff88100a241e98] svc_process at ffffffffa05fbc10 [sunrpc]
#20 [ffff88100a241eb8] nfsd at ffffffffa0647b62 [nfsd]
#21 [ffff88100a241ee8] kthread at ffffffff81091d66
#22 [ffff88100a241f48] kernel_thread at ffffffff8100c14a
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The customer uses NFS for staging files onto the clusters. So it should just be a lot of copies, removes, that sort of thing. There shouldn&apos;t be any jobs, but it&apos;s possible there are some desktop applications that run against it. It happens pretty frequently, though they don&apos;t have a reproducer. &lt;/p&gt;

&lt;p&gt;Let me know if there is any other data I can get.&lt;/p&gt;

&lt;p&gt;Thanks.&lt;/p&gt;</comment>
                            <comment id="64226" author="lixi" created="Wed, 14 Aug 2013 08:56:44 +0000"  >&lt;p&gt;We hit the same problem. Here is how we reproduce it:&lt;br/&gt;
1. Create a directory under the Lustre directory which is exported, e.g. create /mnt/lustre/export/dir.&lt;br/&gt;
2. Change the directory&apos;s mode to 700 so that no one other than root has the right to access it.&lt;br/&gt;
3. Export the Lustre directry, i.e. /mnt/lustre/export. Please make sure &apos;no_root_squash&apos; option is not enabled.&lt;br/&gt;
4. Mount the NFS client on some node.&lt;br/&gt;
5. On the NFS client, cd into the Lustre directory, i.e. directory of /mnt/lustre/export.&lt;br/&gt;
6. On the NFS server (and the Lustre client), unexport the NFS server, unmount Lustre client,&lt;br/&gt;
7. On the NFS client, run &apos;ls -l&apos;. The operation will stuck.&lt;br/&gt;
8. On the NFS server, remount Lustre client, reexport  NFS server (i.e. simulate a reboot).&lt;br/&gt;
9. Wait for a while, we will hit the LBUG (on master branch) or get EACCES (on Lustre-2.1.4)&lt;/p&gt;

&lt;p&gt;After some inverstigation, we found that the cause is that the nfs daemon user does not have the permission to access &quot;..&quot; directory.&lt;/p&gt;

&lt;p&gt;We will upload the fix patch soon.&lt;/p&gt;</comment>
                            <comment id="64227" author="lixi" created="Wed, 14 Aug 2013 09:32:57 +0000"  >&lt;p&gt;Here is the patch which tries to fix the problem.&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/7327&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7327&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="64228" author="emoly.liu" created="Wed, 14 Aug 2013 09:51:09 +0000"  >&lt;p&gt;Thanks for your patch!&lt;/p&gt;</comment>
                            <comment id="64378" author="lixi" created="Fri, 16 Aug 2013 02:40:14 +0000"  >&lt;p&gt;Here is a patch which tries to fix the problem in another way:&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/#/c/7349/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7349/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The last patch (&lt;a href=&quot;http://review.whamcloud.com/7327&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7327&lt;/a&gt;) tries to fix the problem by skipping the permission check on MDT. This patch avoids permission denied by pretending the client is sending a RPC of root user. It is useful for us because this is a client-side-only patch, and it is difficult for us to get downtime for the customer.&lt;/p&gt;

&lt;p&gt;I am not sure which patch is better. Maybe we should think more about their security problem?&lt;/p&gt;</comment>
                            <comment id="64438" author="yong.fan" created="Sun, 18 Aug 2013 13:59:18 +0000"  >&lt;p&gt;I am not sure whether I understand your case correctly or not. But consider the following case:&lt;/p&gt;

&lt;p&gt;1) root user &quot;mkdir /tmp/test&quot;&lt;br/&gt;
2) root user &quot;chmod 0700 /tmp/test&quot;&lt;br/&gt;
3) root user &quot;cd /tmp/test&quot;&lt;br/&gt;
4) root user &quot;su USER1&quot;&lt;br/&gt;
5) USER1 &quot;ls -l&quot;&lt;/p&gt;

&lt;p&gt;The result is &quot;ls: cannot open directory .: Permission denied&quot;.&lt;/p&gt;

&lt;p&gt;Back to the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3727&quot; title=&quot;LBUG (llite_nfs.c:281:ll_get_parent()) ASSERTION(body-&amp;gt;valid &amp;amp; OBD_MD_FLID) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3727&quot;&gt;&lt;del&gt;LU-3727&lt;/del&gt;&lt;/a&gt;, if the &quot;/mnt/lustre/export&quot; is only accessible for root user, then &quot;ls -l&quot; under it as non-root user should get &quot;Permission denied&quot; instead of bypassing permission check as the patch will do, right? If yes, we need to remove invalid &quot;ASSERT()&quot; and report &quot;Permission denied&quot; as expected.&lt;/p&gt;</comment>
                            <comment id="64439" author="lixi" created="Sun, 18 Aug 2013 14:29:18 +0000"  >&lt;p&gt;Sorry, maybe my former explaination was not accurate. Here is how to reproduce the problem.&lt;br/&gt;
1) NFS server &quot;mkdir /mnt/lustre/export&quot;&lt;br/&gt;
2) NFS server exports /mnt/lustre/export&lt;br/&gt;
3) NFS server &quot;mkdir /mnt/lustre/export/dir&quot;&lt;br/&gt;
4) NFS server &quot;chmod 0700 /mnt/lustre/export/dir&quot;&lt;br/&gt;
5) NFS client &quot;cd /mnt/lustre/export/&quot;&lt;br/&gt;
6) NFS server &quot;service nfs stop&quot;&lt;br/&gt;
7) NFS client &quot;ls -l&quot;&lt;br/&gt;
8) NFS server &quot;umount /mnt/lustre&quot;&lt;br/&gt;
9) NFS server mount /mnt/lustre again&lt;br/&gt;
10) NFS server &quot;service nfs start&quot;&lt;br/&gt;
11) NFS client hits LBUG after a few seconds&lt;/p&gt;

&lt;p&gt;Please notice that at step 7), we are under &quot;/mnt/lustre/export&quot;, and we have the permission for this operation.&lt;br/&gt;
If &quot;no_root_squash&quot; is disabled, the problem will show up no matter the user is root or not.&lt;/p&gt;</comment>
                            <comment id="65041" author="ihara" created="Mon, 26 Aug 2013 02:56:42 +0000"  >&lt;p&gt;Hi FanYong &amp;amp; Emoly&lt;/p&gt;

&lt;p&gt;Would you please review Li&apos;s patch quickly? &lt;br/&gt;
And please give us advices/comments on this, please?&lt;/p&gt;

&lt;p&gt;Thanks!&lt;/p&gt;
</comment>
                            <comment id="65046" author="yong.fan" created="Mon, 26 Aug 2013 06:02:36 +0000"  >&lt;p&gt;I think the patch (&lt;a href=&quot;http://review.whamcloud.com/#/c/7349/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7349/&lt;/a&gt;) is some hack, any will introduce security hole. There may be UID/GID mapping on MDT side, so even if you specify the special case as root user, it still possible be mapped to another user in the future. It is MDT to decide how to handle such case.&lt;/p&gt;</comment>
                            <comment id="65048" author="lixi" created="Mon, 26 Aug 2013 06:31:27 +0000"  >&lt;p&gt;Thank you very much, Fan Yong!&lt;/p&gt;

&lt;p&gt;I guess the first patch (&lt;a href=&quot;http://review.whamcloud.com/7327&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7327&lt;/a&gt;) is more acceptable, right?&lt;/p&gt;</comment>
                            <comment id="65050" author="yong.fan" created="Mon, 26 Aug 2013 07:08:16 +0000"  >&lt;p&gt;I think 7327 is better than 7349, although the former one still can be improved.&lt;/p&gt;</comment>
                            <comment id="68032" author="shadow" created="Tue, 1 Oct 2013 11:41:30 +0000"  >&lt;p&gt;Li Xi, &lt;/p&gt;

&lt;p&gt;I have some questions about you script to reproduce a bug.&lt;br/&gt;
1) if nfs client will able to do &apos;cd $dir&apos; they should be lookup a permissions in last directory as minimum.&lt;br/&gt;
so you should be have EACCESS in that step.&lt;br/&gt;
2) in case nfs client don&apos;t have any verify permissions in that step - have a EACCESS in return for ls -l command, looks reasonable, but you tried avoid that permission check and take output.&lt;/p&gt;

&lt;p&gt;if (2) is true - i think question - why &apos;md_getattr_name&apos; isn&apos;t return an error as it&apos;s expected.&lt;/p&gt;

&lt;p&gt;may you attach a full lustre debug from lustre in that crash? i think you use a single node configuration for a nfs server so we will be see all operations in logs.&lt;/p&gt;</comment>
                            <comment id="68036" author="lixi" created="Tue, 1 Oct 2013 12:26:20 +0000"  >&lt;p&gt;The trace output when this BUG happens. Please note that this log is trace on Lustre-2.1, so no LBUG happens. But basically, it is the same problem.&lt;/p&gt;</comment>
                            <comment id="68037" author="lixi" created="Tue, 1 Oct 2013 12:27:29 +0000"  >&lt;p&gt;Hi Alexey,&lt;/p&gt;

&lt;p&gt;Yeah, NFS client has the permission to access &apos;/mnt/lustre/export/&apos; but it has no permission to access &apos;/mnt/lustre/export/dir&apos;. However, when NFS daemon restart (which is not normal and that is when ll_get_parent() is called), NFS client should get the attribute of &apos;/mnt/lustre/export/dir/../&apos;, which will make Lustre client (and NFS server) hit the LBUG. Normally, NFS client could access &apos;/mnt/lustre/export/&apos; withouth any problem though it has no permission to access &apos;/mnt/lustre/export/dir&apos;, but ll_get_parent() is different. Since there is no reason to require that a user doing &apos;ls -l /mnt/lustre/export/&apos; has the permission to access &apos;/mnt/lustre/export/dir&apos;, I think the best way to fix this is to avoid the permission check of &apos;/mnt/lustre/export/dir&apos; in this case.&lt;/p&gt;

&lt;p&gt;I&apos;ve post the lustre debug file as &apos;log.txt&apos;. Sorry, I should have post it earlier.&lt;/p&gt;
</comment>
                            <comment id="69664" author="paf" created="Wed, 23 Oct 2013 19:05:59 +0000"  >&lt;p&gt;We encountered this assertion while running a set of tests from the Linux Test Project over NFS.  We did not do any unmount/remounting of NFS or Lustre as part of this, but are consistently able to hit the bug with this test.  We hit it with both a SLES11SP1 Lustre client and a CentOS 6.4 Lustre client, in both cases to CentOS 6.4 servers running 2.4.1.&lt;/p&gt;

&lt;p&gt;I&apos;ll attached the source file for this particular test (the LTP is a GPL&apos;ed project).&lt;/p&gt;

&lt;p&gt;The test set is called unlink08, and is a series of tests of the unlink system call.&lt;/p&gt;

&lt;p&gt;We&apos;re planning to test the patch in &lt;a href=&quot;http://review.whamcloud.com/#/c/7327/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7327/&lt;/a&gt; today, I&apos;ll report back with results.&lt;/p&gt;</comment>
                            <comment id="69678" author="paf" created="Wed, 23 Oct 2013 20:49:32 +0000"  >&lt;p&gt;With 3727 applied, the &apos;unlink8&apos; test no longer hits this bug.&lt;/p&gt;</comment>
                            <comment id="69681" author="shadow" created="Wed, 23 Oct 2013 21:14:45 +0000"  >&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@rhel6-64 WC-review&amp;#93;&lt;/span&gt;# gcc unlink08.c &lt;br/&gt;
unlink08.c:119:18: error: test.h: No such file or directory&lt;br/&gt;
unlink08.c:120:21: error: usctest.h: No such file or directory&lt;/p&gt;</comment>
                            <comment id="69684" author="paf" created="Wed, 23 Oct 2013 21:30:57 +0000"  >&lt;p&gt;Sorry, Alexey - I hadn&apos;t intended that to be buildable/useable by itself, I just posted it for reference.  &lt;/p&gt;

&lt;p&gt;It&apos;s part of LTP, which can be downloaded from here:&lt;br/&gt;
&lt;a href=&quot;http://sourceforge.net/projects/ltp/files/latest/download&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://sourceforge.net/projects/ltp/files/latest/download&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You&apos;ll find it in testcases/kernel/syscalls/unlink/ in the untarred package, but you&apos;ll have to figure out how to build and run that specific test.&lt;/p&gt;</comment>
                            <comment id="69685" author="shadow" created="Wed, 23 Oct 2013 21:38:51 +0000"  >&lt;p&gt;I have checked on ~2 days ago with some older ltp code, and don&apos;t able to replicate that assertion.&lt;br/&gt;
but my test env hit an hung while accessing an exported dir.&lt;/p&gt;

&lt;p&gt;after hung finished &lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@rhel6-64 ltp]# export TMPDIR=/mnt/lustre2/; /Users/shadow/work/lustre/work/ltp/testcases/kernel/syscalls/unlink/unlink08 

unlink08    1  TPASS  :  unlink(&amp;lt;unwritable directory&amp;gt;) returned 0
unlink08    2  TPASS  :  unlink(&amp;lt;unsearchable directory&amp;gt;) returned 0
unlink08    3  TPASS  :  unlink(&amp;lt;directory&amp;gt;) Failed, errno=21
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;where &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;cat /etc/exports&lt;br/&gt;
/mnt/lustre/export       192.168.0.0/16(rw,no_root_squash,fsid=100)&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;/without fsid i have issues with exporting directory while root lustre dir exported fine/&lt;/p&gt;

&lt;p&gt;ps. same for last LTP from git.&lt;/p&gt;</comment>
                            <comment id="69687" author="paf" created="Wed, 23 Oct 2013 21:52:52 +0000"  >&lt;p&gt;Interesting.  I confirmed and we were able to cause that assertion by running only unlink8, as well as running unlink8 as part of the larger test suite.&lt;br/&gt;
No idea why we&apos;re seeing it and you aren&apos;t.  Regardless, as mentioned above, the patch fixes that issue for us.&lt;/p&gt;

&lt;p&gt;By the way...&lt;br/&gt;
You may already know the details on the FSID issue, but if not, they&apos;re here:&lt;br/&gt;
&lt;a href=&quot;https://jira.hpdd.intel.com/browse/LU-3550&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://jira.hpdd.intel.com/browse/LU-3550&lt;/a&gt;&lt;br/&gt;
Very briefly, Lustre 2.4 uses 64 bit inodes.  NFS does not support 64 bit root inodes, because it explicitly casts inodes to 32 bit in FSID generation, and when those generated FSIDs are returned by the client and unparsed by the server, they cannot be matched back to the 64 bit root inode on Lustre.  (In contrast, when you provide an FSID with -o fsid, it is used directly, so this issue is avoided.  Why a generated FSID is parsed to an inode and looked up by the server - rather than used as an arbitrary ID for the export, like when an FSID is provided - is something of a mystery.)&lt;/p&gt;</comment>
                            <comment id="69689" author="shadow" created="Wed, 23 Oct 2013 22:09:55 +0000"  >&lt;p&gt;question about fsid - is long story. lustre put native fid with parent fid in NFS file handle, FSID may be any number and it&apos;s now same as block device id (for compatibility with older servers in pair). but main question - why i able to export /mnt/lustre without fsid set, but /mnt/lustre/export need fsid.&lt;/p&gt;

&lt;p&gt;as about assert - did you able to take a crashdump and extract lustre debug log from core file?&lt;br/&gt;
and provide a git hash of lustre sources.&lt;/p&gt;</comment>
                            <comment id="69691" author="paf" created="Wed, 23 Oct 2013 22:18:27 +0000"  >&lt;p&gt;re FSID:&lt;br/&gt;
I&apos;m guessing, but I believe because /mnt/lustre has an inode # in your normal Linux file system, whereas /mnt/lustre/export exists only in Lustre.  &lt;br/&gt;
It may matter if you mount Lustre before or after exporting /mnt/lustre - Possibly if you mount lustre at /mnt/lustre/ first and then do the exportfs, you would get the Lustre inode # and see the problem.  &lt;br/&gt;
I recall having problems exporting the root of Lustre mounts as well as sub-directories of those mounts, but I was definitely mounting Lustre before doing the export.&lt;/p&gt;

&lt;p&gt;re: The assertion.&lt;br/&gt;
We were running Cray&apos;s version of Lustre 2.4.1, so no Git hash.  However, it&apos;s very close to the Intel 2.4.1 tag.&lt;/p&gt;

&lt;p&gt;We could probably extract logs from the dump we have, but the logs only had the default debugging options, IE:&lt;br/&gt;
debug=ioctl neterror warning error emerg ha config console&lt;/p&gt;

&lt;p&gt;If you&apos;re still interested, I could probably get those for you tomorrow.&lt;/p&gt;</comment>
                            <comment id="69692" author="shadow" created="Wed, 23 Oct 2013 22:22:29 +0000"  >&lt;p&gt;it&apos;s don&apos;t enough &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/sad.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;... i need full debug logs for both - mdt and client nodes.. primary for a client, but mdt also interested. I will try checkout to 2.4.1 and see a difference.&lt;/p&gt;

&lt;p&gt;tried to replicate with v2_4_1 tag, but without success.&lt;/p&gt;</comment>
                            <comment id="69696" author="shadow" created="Wed, 23 Oct 2013 22:54:08 +0000"  >&lt;p&gt;Li,&lt;/p&gt;

&lt;p&gt;attached log don&apos;t have an information about rpc and isn&apos;t full - it&apos;s have just last step when we hit an error..&lt;/p&gt;</comment>
                            <comment id="69716" author="lixi" created="Thu, 24 Oct 2013 04:11:37 +0000"  >&lt;p&gt;I reproduced the problem with following steps.&lt;br/&gt;
1. Run all the servers/client on the same machine.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# df&lt;br/&gt;
Filesystem           1K-blocks      Used Available Use% Mounted on&lt;br/&gt;
/dev/sda3            100798068  92258100   3419652  97% /&lt;br/&gt;
none                  16424096         0  16424096   0% /dev/shm&lt;br/&gt;
/dev/sda1               521780    216388    278888  44% /boot&lt;br/&gt;
/dev/sda5              1043548     43396    947140   5% /mnt/mgs&lt;br/&gt;
/dev/sda6              7865980    456800   6885408   7% /mnt/mdt0&lt;br/&gt;
/dev/sda7              7865980    456396   6885812   7% /mnt/mdt1&lt;br/&gt;
/dev/sda8             33442176    562620  31201168   2% /mnt/ost0&lt;br/&gt;
/dev/sda9             80025512    464196  75545068   1% /mnt/ost1&lt;br/&gt;
192.168.3.100@tcp:/server1&lt;br/&gt;
                     113467688   1026816 106746236   1% /mnt/lustre&lt;br/&gt;
2. NFS server is on the same machine. Please note no_root_squash should NOt be enabled to reproduce this problem using root&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# cat /etc/exports &lt;br/&gt;
/mnt/lustre/nfs *(fsid=0,rw)&lt;br/&gt;
3. The directory should have no execute permission for normal users.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# ll /mnt/lustre/nfs/&lt;br/&gt;
total 4&lt;br/&gt;
drwx------ 2 root root 4096 Oct 24 11:39 dir&lt;br/&gt;
4. Trace the path&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# echo trace &amp;gt; /proc/sys/lnet/debug&lt;br/&gt;
5. Clearup the messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# lctl debug_kernel /tmp/lustre.log&lt;br/&gt;
6. Start the NFS&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# service nfs restart&lt;br/&gt;
7. Start the NFS client&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# mount 192.168.3.100:/mnt/lustre/nfs /mnt/nfs&lt;br/&gt;
8. Go down to the NFS client directory&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# cd /mnt/nfs&lt;br/&gt;
9. Get the target dentry so that ll_get_parent will call for it.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 nfs&amp;#93;&lt;/span&gt;# ls -l&lt;br/&gt;
10. Get the target dentry so that ll_get_parent will call for it.&lt;br/&gt;
11. Stop the NFS server to simulate a reboot&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 nfs&amp;#93;&lt;/span&gt;# service nfs stop&lt;br/&gt;
Shutting down NFS daemon: [  OK  ]&lt;br/&gt;
Shutting down NFS mountd: [  OK  ]&lt;br/&gt;
Shutting down NFS quotas: [  OK  ]&lt;br/&gt;
Shutting down NFS services:  [  OK  ]&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 nfs&amp;#93;&lt;/span&gt;# umount /mnt/lustre&lt;br/&gt;
12. Access the directory again, this command will stuck&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 nfs&amp;#93;&lt;/span&gt;# ls -l&lt;br/&gt;
13. Start the NFS server again on another terminal&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# mount -t lustre 192.168.3.100@tcp:/server1 /mnt/lustre&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 ~&amp;#93;&lt;/span&gt;# service nfs start&lt;br/&gt;
WARNING: Deprecated config file /etc/modprobe.conf, all config files belong into /etc/modprobe.d/.&lt;br/&gt;
Starting NFS services:  [  OK  ]&lt;br/&gt;
Starting NFS quotas: [  OK  ]&lt;br/&gt;
Starting NFS mountd: [  OK  ]&lt;br/&gt;
Stopping RPC idmapd: [  OK  ]&lt;br/&gt;
Starting RPC idmapd: [  OK  ]&lt;br/&gt;
Starting NFS daemon: [  OK  ]&lt;br/&gt;
14. After a few seconds the stucked command hit the LBUG.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@server1 nfs&amp;#93;&lt;/span&gt;# ls -l&lt;/p&gt;

&lt;p&gt;Message from syslogd@server1 at Oct 24 12:08:53 ...&lt;br/&gt;
 kernel:LustreError: 3348:0:(llite_nfs.c:311:ll_get_parent()) ASSERTION( body-&amp;gt;valid &amp;amp; (0x00000001ULL) ) failed: &lt;/p&gt;

&lt;p&gt;Message from syslogd@server1 at Oct 24 12:08:53 ...&lt;br/&gt;
 kernel:LustreError: 3348:0:(llite_nfs.c:311:ll_get_parent()) LBUG&lt;br/&gt;
15. Get the Lustre log&lt;br/&gt;
lctl debug_file /tmp/lustre-log.1382641733.3348 /tmp/lustre.log &lt;/p&gt;</comment>
                            <comment id="69717" author="lixi" created="Thu, 24 Oct 2013 04:14:14 +0000"  >&lt;p&gt;Alexey,&lt;/p&gt;

&lt;p&gt;&apos;lustre.log&apos; is the trace log which I got when reproducing the problem. Hope it helps.&lt;/p&gt;</comment>
                            <comment id="69719" author="shadow" created="Thu, 24 Oct 2013 05:49:16 +0000"  >&lt;p&gt;Li,&lt;/p&gt;

&lt;p&gt;thanks!&lt;/p&gt;</comment>
                            <comment id="69816" author="shadow" created="Thu, 24 Oct 2013 17:26:56 +0000"  >&lt;p&gt;Li,&lt;/p&gt;

&lt;p&gt;thanks again. devil in details.. we need additional directory created in exported dir.&lt;br/&gt;
without it ll isn&apos;t trigger a bug.&lt;/p&gt;</comment>
                            <comment id="69827" author="shadow" created="Thu, 24 Oct 2013 18:32:20 +0000"  >&lt;p&gt;as i talk before - mdt generate an error as before&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000004:00000001:1.0:1382635559.670116:0:15672:0:(mdd_permission.c:309:__mdd_permission_internal()) Process leaving (rc=18446744073709551603 : -13 : fffffffffffffff3)
00000004:00000001:1.0:1382635559.670117:0:15672:0:(mdd_dir.c:90:__mdd_lookup()) Process leaving (rc=18446744073709551603 : -13 : fffffffffffffff3)
00000004:00000001:1.0:1382635559.670117:0:15672:0:(mdd_dir.c:115:mdd_lookup()) Process leaving (rc=18446744073709551603 : -13 : fffffffffffffff3)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;but that error isn&apos;t returned to the caller&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000004:00000001:1.0:1382635559.670119:0:15672:0:(mdt_handler.c:1273:mdt_getattr_name_lock()) Process leaving (rc=0 : 0 : 0)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;i that case client correctly trigger a panic as we have none errors in processing, but reply isn&apos;t filled correctly.&lt;br/&gt;
that bug should be affect isn&apos;t NFS only.&lt;/p&gt;</comment>
                            <comment id="69828" author="shadow" created="Thu, 24 Oct 2013 18:37:04 +0000"  >&lt;p&gt;Li,&lt;/p&gt;

&lt;p&gt;may you look into MDT code to verify - why that error isn&apos;t returned correctly to the client?&lt;br/&gt;
from my point view it&apos;s should be addressed to the&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;#&lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; 0
        &lt;span class=&quot;code-comment&quot;&gt;/* XXX is raw_lookup possible as intent operation? */&lt;/span&gt;
        &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (rc != 0) {
                &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (rc == -ENOENT)
                        mdt_set_disposition(info, ldlm_rep, DISP_LOOKUP_NEG);
                RETURN(rc);
        } &lt;span class=&quot;code-keyword&quot;&gt;else&lt;/span&gt;
                mdt_set_disposition(info, ldlm_rep, DISP_LOOKUP_POS);

        repbody = req_capsule_server_get(info-&amp;gt;mti_pill, &amp;amp;RMF_MDT_BODY);
#endif
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;or we need to replace an &apos;RETURN(1);&apos; to &quot;return(rc)&apos; at end of mdt_raw_lookup() function.&lt;/p&gt;</comment>
                            <comment id="69852" author="paf" created="Thu, 24 Oct 2013 22:21:52 +0000"  >&lt;p&gt;At Alexey&apos;s request, we reproduced this.&lt;/p&gt;

&lt;p&gt;Here&apos;s the procedure from our test engineer:&lt;br/&gt;
&amp;#8212;&lt;br/&gt;
1)      Mount lustre on NFS server&lt;/p&gt;

&lt;p&gt;2)      Start nfsserver daemon on NFS server&lt;/p&gt;

&lt;p&gt;3)      Export nfs (sudo /usr/sbin/exportfs -i -o rw,insecure,no_root_squash,no_subtree_check,fsid=538 &amp;#42;:/extlus )&lt;/p&gt;

&lt;p&gt;4)      Mount NFS on client (sudo mount perses-esl3:/extlus /tmp/lus)&lt;/p&gt;

&lt;p&gt;5)      Run test using /tmp/lus&lt;br/&gt;
&amp;#8212;&lt;br/&gt;
Other than fsid=, the other options are just what we usually use when testing NFS internally.&lt;/p&gt;

&lt;p&gt;Attaching logs shortly.&lt;/p&gt;</comment>
                            <comment id="69853" author="paf" created="Thu, 24 Oct 2013 22:22:49 +0000"  >&lt;p&gt;MDS log during the test.  Client LBUGged doing unlink8 test from LTP as described earlier.&lt;/p&gt;</comment>
                            <comment id="69861" author="lixi" created="Fri, 25 Oct 2013 00:16:28 +0000"  >&lt;p&gt;Hi Alexey,&lt;/p&gt;

&lt;p&gt;I agree on that mdt_raw_lookup() should not return 1 all the time. And follwoing patch tries to fix that too.&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/#/c/7327&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7327&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="69862" author="shadow" created="Fri, 25 Oct 2013 00:28:14 +0000"  >&lt;p&gt;Hi Li,&lt;/p&gt;

&lt;p&gt;main question for it - did we need a set intent disposition in reply. may you check - how it send from client? via mdc_intent_lock or other way ?&lt;/p&gt;</comment>
                            <comment id="70040" author="paf" created="Mon, 28 Oct 2013 17:37:04 +0000"  >&lt;p&gt;It might be worth noting that we hit this on 2.4.1.  The ticket only lists 1.8.9/2.1.5.&lt;/p&gt;</comment>
                            <comment id="71518" author="shadow" created="Thu, 14 Nov 2013 12:31:00 +0000"  >&lt;p&gt;any ability to answer ?&lt;/p&gt;</comment>
                            <comment id="71524" author="lixi" created="Thu, 14 Nov 2013 13:53:16 +0000"  >&lt;p&gt;Hi Alexey,&lt;/p&gt;

&lt;p&gt;I am sorry, maybe because the lack of background knowledge, I don&apos;t understand the question well. Would you please explain a little bit about it? And do you have any specific problems about the patch?&lt;/p&gt;</comment>
                            <comment id="77067" author="ihara" created="Fri, 14 Feb 2014 09:09:06 +0000"  >&lt;p&gt;Alexey, can you please describe your quesiton in detial here?&lt;/p&gt;</comment>
                            <comment id="92131" author="ferner" created="Thu, 21 Aug 2014 13:28:25 +0000"  >&lt;p&gt;Looks like we&apos;ve just hit this as well on a NFS server/lustre client which is still running 1.8.9 after upgrading one file system to 2.5.2. We intend to upgrade the client to 2.5.2 as well ASAP but need to upgrade all file system first. &lt;/p&gt;

&lt;p&gt;Is there any indication that this might be fixed in 2.5.2?&lt;/p&gt;</comment>
                            <comment id="92225" author="paf" created="Fri, 22 Aug 2014 15:17:53 +0000"  >&lt;p&gt;Frederik - There&apos;s no movement towards a fix at the moment.  If you&apos;re building your own Lustre, there&apos;s an option: Alexey and Oleg dislike &lt;a href=&quot;http://review.whamcloud.com/#/c/7327/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7327/&lt;/a&gt;, but it does avoid the bug &amp;amp; we&apos;ve been running it at Cray for a bit.&lt;/p&gt;</comment>
                            <comment id="100297" author="lixi" created="Mon, 1 Dec 2014 13:34:44 +0000"  >&lt;p&gt;Patch of &apos;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3952&quot; title=&quot;llite_nfs.c:349:ll_get_parent()) ASSERTION( body-&amp;gt;valid &amp;amp; (0x00000001ULL) ) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3952&quot;&gt;&lt;del&gt;LU-3952&lt;/del&gt;&lt;/a&gt; nfs: don&apos;t panic NFS server if MDS fails to find FID&apos; helps to walk around the problem for master branch. But that patch does not help earlier versions such as b2_1. And I don&apos;t think the root cause has been fixed by that patch. Refreshed &lt;a href=&quot;http://review.whamcloud.com/#/c/7327/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/7327/&lt;/a&gt; again.&lt;/p&gt;</comment>
                            <comment id="102320" author="gerrit" created="Fri, 26 Dec 2014 16:57:20 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/7327/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/7327/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3727&quot; title=&quot;LBUG (llite_nfs.c:281:ll_get_parent()) ASSERTION(body-&amp;gt;valid &amp;amp; OBD_MD_FLID) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3727&quot;&gt;&lt;del&gt;LU-3727&lt;/del&gt;&lt;/a&gt; nfs: Fix ll_get_parent() LBUG caused by permission&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: a0b959c53d10bf3f0fd6b22de46397d0c7e5f667&lt;/p&gt;</comment>
                            <comment id="102743" author="gerrit" created="Wed, 7 Jan 2015 15:41:31 +0000"  >&lt;p&gt;Lai Siyao (lai.siyao@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/13270&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/13270&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3727&quot; title=&quot;LBUG (llite_nfs.c:281:ll_get_parent()) ASSERTION(body-&amp;gt;valid &amp;amp; OBD_MD_FLID) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3727&quot;&gt;&lt;del&gt;LU-3727&lt;/del&gt;&lt;/a&gt; nfs: Fix ll_get_parent() LBUG caused by permission&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_5&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: cffa66fed624973d71bfc2c3382f1ef0a19397d4&lt;/p&gt;</comment>
                            <comment id="106428" author="pjones" created="Tue, 10 Feb 2015 12:39:17 +0000"  >&lt;p&gt;Landed for 2.7&lt;/p&gt;</comment>
                            <comment id="112363" author="gerrit" created="Mon, 20 Apr 2015 07:20:45 +0000"  >&lt;p&gt;Lai Siyao (lai.siyao@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/14498&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/14498&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3727&quot; title=&quot;LBUG (llite_nfs.c:281:ll_get_parent()) ASSERTION(body-&amp;gt;valid &amp;amp; OBD_MD_FLID) failed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3727&quot;&gt;&lt;del&gt;LU-3727&lt;/del&gt;&lt;/a&gt; nfs: Fix ll_get_parent() LBUG caused by permission&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_3&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: bc90ead229cdaa940cca1819fe6fbfb983506014&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="26977">LU-5730</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="13569" name="log.txt" size="44993" author="lixi" created="Tue, 1 Oct 2013 12:26:20 +0000"/>
                            <attachment id="13691" name="log.unlink08.lctl.dk.out.gz" size="3691833" author="paf" created="Thu, 24 Oct 2013 22:22:49 +0000"/>
                            <attachment id="13678" name="lustre.log" size="3778749" author="lixi" created="Thu, 24 Oct 2013 04:14:14 +0000"/>
                            <attachment id="13673" name="unlink08.c" size="10550" author="paf" created="Wed, 23 Oct 2013 19:07:46 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvxgf:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9597</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>