<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:45:06 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-11578] ldiskfs_map_blocks: comm mdt00_100: lblock 0 mapped to illegal pblock</title>
                <link>https://jira.whamcloud.com/browse/LU-11578</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We hit the following issue today on Oak&apos;s MDT0. We just added another MDT a few days ago so it was empty but I started today to lfs migrate a test directory (from MDT0 to MDT1) when this happened on MDT0. &lt;tt&gt;lfs migrate&lt;/tt&gt;&#160;did actually work for a while (about ~40k inodes have been migrated) until MDT0 did this:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Oct 26 17:26:13 oak-md1-s2 kernel: LDISKFS-fs error (device dm-0): ldiskfs_map_blocks:594: inode #659619751: block 774843950: comm mdt00_100: lblock 0 mapped to illegal pblock (length 1)
Oct 26 17:26:13 oak-md1-s2 kernel: Aborting journal on device dm-0-8.
Oct 26 17:26:13 oak-md1-s2 kernel: LustreError: 3844:0:(osd_handler.c:1586:osd_trans_commit_cb()) transaction @0xffff881b024c6a80 commit error: 2
Oct 26 17:26:13 oak-md1-s2 kernel: LDISKFS-fs (dm-0): Remounting filesystem read-only
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;I performed a fsck with the new &lt;tt&gt;e2fsprogs-1.44.3.wc1-0.el7.x86_64&lt;/tt&gt;&lt;br/&gt;
 Not sure this fixed our issue as I don&apos;t see any reference to inode &lt;tt&gt;659619751&lt;/tt&gt; in it. I&apos;m attaching the full fsck logs.&lt;/p&gt;

&lt;p&gt;Hopefully, MDT0 then restarted without problem but I haven&apos;t touched the partially migrated directory at this time to avoid further issues on a Friday night on this production system. My feeling is that &lt;tt&gt;lfs migrate&lt;/tt&gt; hit some migrated inode that somehow got corrupted on MDT0? The system has been working fine for weeks now (we only have the issue reported in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-11205&quot; title=&quot;Failure to clear the changelog for user 1 on MDT&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-11205&quot;&gt;&lt;del&gt;LU-11205&lt;/del&gt;&lt;/a&gt; regarding changelog_clear errors), so I assume this is due to my running &lt;tt&gt;lfs migrate&lt;/tt&gt;.&#160;I can perform troubleshooting next week. Any recommendation to avoid this in the future is welcome, thanks much!&lt;/p&gt;

&lt;p&gt;Stephane&lt;/p&gt;</description>
                <environment>CentOS 7.4</environment>
        <key id="53837">LU-11578</key>
            <summary>ldiskfs_map_blocks: comm mdt00_100: lblock 0 mapped to illegal pblock</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="ys">Yang Sheng</assignee>
                                    <reporter username="sthiell">Stephane Thiell</reporter>
                        <labels>
                    </labels>
                <created>Sat, 27 Oct 2018 04:35:06 +0000</created>
                <updated>Tue, 26 Jul 2022 03:56:38 +0000</updated>
                            <resolved>Tue, 26 Jul 2022 03:56:38 +0000</resolved>
                                    <version>Lustre 2.10.4</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="235642" author="pjones" created="Sat, 27 Oct 2018 10:23:48 +0000"  >&lt;p&gt;Yang Sheng&lt;/p&gt;

&lt;p&gt;Can you please advise?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="235930" author="ys" created="Tue, 30 Oct 2018 17:01:22 +0000"  >&lt;p&gt;Hi, Stephane,&lt;/p&gt;

&lt;p&gt;Could please collect superblock info of  MDS as below:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;
# tune2fs -l {mds device}

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
YangSheng&lt;/p&gt;</comment>
                            <comment id="235934" author="sthiell" created="Tue, 30 Oct 2018 17:35:49 +0000"  >&lt;p&gt;Hi Yang,&lt;/p&gt;

&lt;p&gt;Absolutely! Thanks for taking a look at this.&lt;br/&gt;
 &#160;&lt;br/&gt;
 Original MDT0 (formatted at the time of Lustre 2.9):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@oak-md1-s2 ~]# tune2fs -l /dev/mapper/md1-rbod1-mdt0 
tune2fs 1.44.3.wc1 (23-July-2018)
device /dev/dm-0 mounted by lustre
Filesystem volume name:   oak-MDT0000
Last mounted on:          /
Filesystem UUID:          0ed1cfdd-8e25-4b6b-9cb9-7be1e89d70ad
Filesystem magic number:  0xEF53
Filesystem revision #:    1 (dynamic)
Filesystem features:      has_journal ext_attr resize_inode dir_index filetype needs_recovery mmp flex_bg dirdata sparse_super large_file huge_file uninit_bg dir_nlink quota
Filesystem flags:         signed_directory_hash 
Default mount options:    user_xattr acl
Filesystem state:         clean
Errors behavior:          Continue
Filesystem OS type:       Linux
Inode count:              873332736
Block count:              436659200
Reserved block count:     21832960
Free blocks:              243851408
Free inodes:              571125380
First block:              0
Block size:               4096
Fragment size:            4096
Reserved GDT blocks:      1024
Blocks per group:         16384
Fragments per group:      16384
Inodes per group:         32768
Inode blocks per group:   4096
Flex block group size:    16
Filesystem created:       Mon Feb 13 12:36:07 2017
Last mount time:          Fri Oct 26 18:39:01 2018
Last write time:          Fri Oct 26 18:39:01 2018
Mount count:              1
Maximum mount count:      -1
Last checked:             Fri Oct 26 17:41:18 2018
Check interval:           0 (&amp;lt;none&amp;gt;)
Lifetime writes:          52 TB
Reserved blocks uid:      0 (user root)
Reserved blocks gid:      0 (group root)
First inode:              11
Inode size:	          512
Required extra isize:     28
Desired extra isize:      28
Journal inode:            8
Default directory hash:   half_md4
Directory Hash Seed:      be3bd996-8da4-4d22-80e4-e7a4c8ce22a0
Journal backup:           inode blocks
MMP block number:         13560
MMP update interval:      5
User quota inode:         3
Group quota inode:        4
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Just in case, same command with the new MDT0001, formatted with 2.10:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@oak-md1-s1 ~]# tune2fs -l /dev/mapper/md1-rbod1-mdt1 
tune2fs 1.42.13.wc6 (05-Feb-2017)
device /dev/dm-2 mounted by lustre
Filesystem volume name:   oak-MDT0001
Last mounted on:          /
Filesystem UUID:          169de89e-6b5d-4480-b118-8f726d7af07b
Filesystem magic number:  0xEF53
Filesystem revision #:    1 (dynamic)
Filesystem features:      has_journal ext_attr resize_inode dir_index filetype needs_recovery mmp flex_bg dirdata sparse_super large_file huge_file uninit_bg dir_nlink quota
Filesystem flags:         signed_directory_hash 
Default mount options:    user_xattr acl
Filesystem state:         clean
Errors behavior:          Continue
Filesystem OS type:       Linux
Inode count:              698600160
Block count:              436659200
Reserved block count:     21832704
Free blocks:              260890660
Free inodes:              698600147
First block:              0
Block size:               4096
Fragment size:            4096
Reserved GDT blocks:      1024
Blocks per group:         20472
Fragments per group:      20472
Inodes per group:         32752
Inode blocks per group:   8188
Flex block group size:    16
Filesystem created:       Thu Oct 18 11:43:21 2018
Last mount time:          Thu Oct 18 11:45:28 2018
Last write time:          Thu Oct 18 11:45:30 2018
Mount count:              3
Maximum mount count:      -1
Last checked:             Thu Oct 18 11:43:21 2018
Check interval:           0 (&amp;lt;none&amp;gt;)
Lifetime writes:          4145 MB
Reserved blocks uid:      0 (user root)
Reserved blocks gid:      0 (group root)
First inode:              11
Inode size:	          1024
Required extra isize:     32
Desired extra isize:      32
Journal inode:            8
Default directory hash:   half_md4
Directory Hash Seed:      c48690f3-5625-496d-adb1-9c3288cc8b00
Journal backup:           inode blocks
MMP block number:         17606
MMP update interval:      5
User quota inode:         3
Group quota inode:        4
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Let me know if you need anything else.&lt;br/&gt;
 Best,&lt;br/&gt;
 Stephane&lt;/p&gt;</comment>
                            <comment id="236264" author="sthiell" created="Fri, 2 Nov 2018 22:28:00 +0000"  >&lt;p&gt;YangSheng,&lt;br/&gt;
Any idea of what could have gone wrong here?&lt;br/&gt;
I&apos;m very reluctant to use &lt;tt&gt;lfs migrate&lt;/tt&gt; again until we figure this out. Thanks.&lt;/p&gt;</comment>
                            <comment id="236278" author="ys" created="Sat, 3 Nov 2018 15:46:00 +0000"  >&lt;p&gt;Hi, Stephane,&lt;/p&gt;

&lt;p&gt;First, Looks like the block No 774843950 is really wrong. You only have  436659200 blocks in your MDS.  You still have free inodes 571125380. So the inode No 659619751 also suspected. You can check it use command debugfs as below:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;#
# debugfs {MDS device}
debugfs: ncheck  659619751 

It should report a pathname to be consistent with this inode. If so, you can check blocks occupied by this pathname.

debug: blocks {pathname}


&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Please paste output in here.&lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
YangSheng&lt;/p&gt;</comment>
                            <comment id="236345" author="sthiell" created="Mon, 5 Nov 2018 19:33:25 +0000"  >&lt;p&gt;Hi YangSheng,&lt;/p&gt;

&lt;p&gt;Thanks! Looks like the inode can&apos;t be found anymore on MDT 0:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@oak-md1-s2 ~]# debugfs /dev/mapper/md1-rbod1-mdt0
debugfs 1.44.3.wc1 (23-July-2018)
debugfs:  ncheck 659619751
Inode	Pathname
debugfs:  
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;The command took at least two hours to perform so I guess it worked. Perhaps the file was migrated to MDT 1 after all. But how to know... the inode # is probably different on MDT 1 after migration. Anyway, what do you think?&lt;/p&gt;

&lt;p&gt;Thanks again,&lt;br/&gt;
 Stephane&lt;/p&gt;</comment>
                            <comment id="236414" author="ys" created="Tue, 6 Nov 2018 11:53:05 +0000"  >&lt;p&gt;Hi, Stephane,&lt;/p&gt;

&lt;p&gt;I don&apos;t think this inode can be moved to other place, Since the failure ldiskfs_map_blocks was issued by read syscall. And then MDT0 was remounted as read-only immediately after this failure. From failure location we can see the block No(774843950) comes from extent status tree. This is a cache was managed by ldiskfs internal. It means this block No should valid before(it can be cached) and become invalid when read again. So i think this issue should relate to memory corruption rather than migration.&lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
YangSheng &lt;/p&gt;</comment>
                            <comment id="236459" author="sthiell" created="Tue, 6 Nov 2018 17:16:41 +0000"  >&lt;p&gt;Hi YangSheng,&lt;/p&gt;

&lt;p&gt;Thanks for the explanation, this is helpful and that makes sense. And this would also explain why fsck didn&apos;t find anything interesting. I&apos;ll try to check what exactly have been migrated and possibly launch &lt;tt&gt;lfs migrate&lt;/tt&gt; again today and see how it goes. I&apos;ll report back any useful info.&lt;/p&gt;

&lt;p&gt;Best,&lt;br/&gt;
Stephane&lt;/p&gt;</comment>
                            <comment id="236468" author="sthiell" created="Tue, 6 Nov 2018 18:04:27 +0000"  >&lt;p&gt;Ok this is not good. I tried to launch &lt;tt&gt;lfs migrate&lt;/tt&gt; again:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;lfs migrate --mdt-index 1 ./sthiell
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Immediately the MDT0 crashed with:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[5516611.089815] LustreError: 15595:0:(osd_handler.c:1982:osd_object_release()) LBUG
[5516611.114441] Pid: 15595, comm: mdt00_105
[5516611.127614] 
Call Trace:
[5516611.141658]  [&amp;lt;ffffffffc08587ae&amp;gt;] libcfs_call_trace+0x4e/0x60 [libcfs]
[5516611.163987]  [&amp;lt;ffffffffc085883c&amp;gt;] lbug_with_loc+0x4c/0xb0 [libcfs]
[5516611.185171]  [&amp;lt;ffffffffc1019610&amp;gt;] osd_get_ldiskfs_dirent_param+0x0/0x130 [osd_ldiskfs]
[5516611.212390]  [&amp;lt;ffffffffc099ec4d&amp;gt;] lu_object_put+0x17d/0x3d0 [obdclass]
[5516611.235012]  [&amp;lt;ffffffffc116d233&amp;gt;] mdt_reint_migrate_internal.isra.38+0x793/0x16e0 [mdt]
[5516611.261959]  [&amp;lt;ffffffffc0b44851&amp;gt;] ? ldlm_cli_enqueue_local+0x271/0x860 [ptlrpc]
[5516611.287397]  [&amp;lt;ffffffffc116e3e5&amp;gt;] mdt_reint_rename_or_migrate.isra.39+0x265/0x860 [mdt]
[5516611.315164]  [&amp;lt;ffffffffc0b47ee0&amp;gt;] ? ldlm_blocking_ast+0x0/0x170 [ptlrpc]
[5516611.337804]  [&amp;lt;ffffffffc0b42340&amp;gt;] ? ldlm_completion_ast+0x0/0x920 [ptlrpc]
[5516611.361002]  [&amp;lt;ffffffffc116e9f0&amp;gt;] mdt_reint_migrate+0x10/0x20 [mdt]
[5516611.382451]  [&amp;lt;ffffffffc1172a40&amp;gt;] mdt_reint_rec+0x80/0x210 [mdt]
[5516611.402775]  [&amp;lt;ffffffffc115431b&amp;gt;] mdt_reint_internal+0x5fb/0x9c0 [mdt]
[5516611.425379]  [&amp;lt;ffffffffc115fe47&amp;gt;] mdt_reint+0x67/0x140 [mdt]
[5516611.444583]  [&amp;lt;ffffffffc0bdad95&amp;gt;] tgt_request_handle+0x925/0x1370 [ptlrpc]
[5516611.467794]  [&amp;lt;ffffffffc0b83bf6&amp;gt;] ptlrpc_server_handle_request+0x236/0xa90 [ptlrpc]
[5516611.494998]  [&amp;lt;ffffffffc0b80228&amp;gt;] ? ptlrpc_wait_event+0x98/0x340 [ptlrpc]
[5516611.517896]  [&amp;lt;ffffffff810c4822&amp;gt;] ? default_wake_function+0x12/0x20
[5516611.539085]  [&amp;lt;ffffffff810ba588&amp;gt;] ? __wake_up_common+0x58/0x90
[5516611.558854]  [&amp;lt;ffffffffc0b87332&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
[5516611.580053]  [&amp;lt;ffffffffc0b868a0&amp;gt;] ? ptlrpc_main+0x0/0x1e40 [ptlrpc]
[5516611.602365]  [&amp;lt;ffffffff810b098f&amp;gt;] kthread+0xcf/0xe0
[5516611.618967]  [&amp;lt;ffffffff810b08c0&amp;gt;] ? kthread+0x0/0xe0
[5516611.635856]  [&amp;lt;ffffffff816b4f58&amp;gt;] ret_from_fork+0x58/0x90
[5516611.654178]  [&amp;lt;ffffffff810b08c0&amp;gt;] ? kthread+0x0/0xe0
[5516611.671066] 
[5516611.676514] Kernel panic - not syncing: LBUG
[5516611.691113] CPU: 2 PID: 15595 Comm: mdt00_105 Tainted: G           OE  ------------   3.10.0-693.2.2.el7_lustre.pl1.x86_64 #1
[5516611.728879] Hardware name: Dell Inc. PowerEdge R630/0CNCJW, BIOS 2.6.0 10/26/2017
[5516611.754060]  ffff880e12696e00 00000000e8661d26 ffff8816667b3a08 ffffffff816a3db1
[5516611.778993]  ffff8816667b3a88 ffffffff8169dc74 ffffffff00000008 ffff8816667b3a98
[5516611.803930]  ffff8816667b3a38 00000000e8661d26 00000000e8661d26 ffff88103ce4f8b8
[5516611.828860] Call Trace:
[5516611.837456]  [&amp;lt;ffffffff816a3db1&amp;gt;] dump_stack+0x19/0x1b
[5516611.854920]  [&amp;lt;ffffffff8169dc74&amp;gt;] panic+0xe8/0x20d
[5516611.871240]  [&amp;lt;ffffffffc0858854&amp;gt;] lbug_with_loc+0x64/0xb0 [libcfs]
[5516611.892136]  [&amp;lt;ffffffffc1019610&amp;gt;] osd_object_release+0x60/0x60 [osd_ldiskfs]
[5516611.915909]  [&amp;lt;ffffffffc099ec4d&amp;gt;] lu_object_put+0x17d/0x3d0 [obdclass]
[5516611.942513]  [&amp;lt;ffffffffc116d233&amp;gt;] mdt_reint_migrate_internal.isra.38+0x793/0x16e0 [mdt]
[5516611.972007]  [&amp;lt;ffffffffc0b44851&amp;gt;] ? ldlm_cli_enqueue_local+0x271/0x860 [ptlrpc]
[5516611.996623]  [&amp;lt;ffffffffc116e3e5&amp;gt;] mdt_reint_rename_or_migrate.isra.39+0x265/0x860 [mdt]
[5516612.024113]  [&amp;lt;ffffffffc0b47ee0&amp;gt;] ? ldlm_blocking_ast_nocheck+0x310/0x310 [ptlrpc]
[5516612.049600]  [&amp;lt;ffffffffc0b42340&amp;gt;] ? ldlm_expired_completion_wait+0x240/0x240 [ptlrpc]
[5516612.075932]  [&amp;lt;ffffffffc116e9f0&amp;gt;] mdt_reint_migrate+0x10/0x20 [mdt]
[5516612.097115]  [&amp;lt;ffffffffc1172a40&amp;gt;] mdt_reint_rec+0x80/0x210 [mdt]
[5516612.118587]  [&amp;lt;ffffffffc115431b&amp;gt;] mdt_reint_internal+0x5fb/0x9c0 [mdt]
[5516612.140628]  [&amp;lt;ffffffffc115fe47&amp;gt;] mdt_reint+0x67/0x140 [mdt]
[5516612.160410]  [&amp;lt;ffffffffc0bdad95&amp;gt;] tgt_request_handle+0x925/0x1370 [ptlrpc]
[5516612.183611]  [&amp;lt;ffffffffc0b83bf6&amp;gt;] ptlrpc_server_handle_request+0x236/0xa90 [ptlrpc]
[5516612.209387]  [&amp;lt;ffffffffc0b80228&amp;gt;] ? ptlrpc_wait_event+0x98/0x340 [ptlrpc]
[5516612.232549]  [&amp;lt;ffffffff810c4822&amp;gt;] ? default_wake_function+0x12/0x20
[5516612.253726]  [&amp;lt;ffffffff810ba588&amp;gt;] ? __wake_up_common+0x58/0x90
[5516612.273494]  [&amp;lt;ffffffffc0b87332&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
[5516612.294971]  [&amp;lt;ffffffffc0b868a0&amp;gt;] ? ptlrpc_register_service+0xe30/0xe30 [ptlrpc]
[5516612.322155]  [&amp;lt;ffffffff810b098f&amp;gt;] kthread+0xcf/0xe0
[5516612.338755]  [&amp;lt;ffffffff810b08c0&amp;gt;] ? insert_kthread_work+0x40/0x40
[5516612.360224]  [&amp;lt;ffffffff816b4f58&amp;gt;] ret_from_fork+0x58/0x90
[5516612.378539]  [&amp;lt;ffffffff810b08c0&amp;gt;] ? insert_kthread_work+0x40/0x40
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;After than, when trying to remount MDT0, I got:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Nov 06 09:56:39 oak-md1-s2 kernel: Lustre: Lustre: Build Version: 2.10.4
Nov 06 09:56:39 oak-md1-s2 kernel: LNet: Using FMR for registration
Nov 06 09:56:39 oak-md1-s2 kernel: LNet: Added LNI 10.0.2.52@o2ib5 [8/256/0/180]
Nov 06 09:56:41 oak-md1-s2 kernel: LDISKFS-fs warning (device dm-1): ldiskfs_multi_mount_protect:322: MMP interval 42 higher than expected, please wait.
Nov 06 09:57:33 oak-md1-s2 kernel: LDISKFS-fs (dm-1): recovery complete
Nov 06 09:57:33 oak-md1-s2 kernel: LDISKFS-fs (dm-1): mounted filesystem with ordered data mode. Opts: user_xattr,errors=remount-ro,acl,no_mbcache,nodelalloc
Nov 06 09:57:36 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450747
Nov 06 09:57:36 oak-md1-s2 kernel: Aborting journal on device dm-1-8.
Nov 06 09:57:36 oak-md1-s2 kernel: LDISKFS-fs (dm-1): Remounting filesystem read-only
Nov 06 09:57:36 oak-md1-s2 kernel: LustreError: 3564:0:(osd_scrub.c:1962:osd_ios_lookup_one_len()) Fail to find #460450747 in lost+found (11/0): rc = -5
Nov 06 09:57:36 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450866
Nov 06 09:57:36 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450930
Nov 06 09:57:37 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450942
Nov 06 09:57:37 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450954
Nov 06 09:57:37 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450963
Nov 06 09:57:37 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450965
Nov 06 09:57:37 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460450970
Nov 06 09:57:37 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460451000
Nov 06 09:57:37 oak-md1-s2 kernel: LDISKFS-fs error (device dm-1): ldiskfs_lookup:1809: inode #11: comm mount.lustre: deleted inode referenced: 460451002
Nov 06 09:57:37 oak-md1-s2 kernel: LustreError: 137-5: oak-MDT0000_UUID: not available for connect from 10.9.107.16@o2ib4 (no target). If you are running an HA pair check that the target is mounted on the other server.
Nov 06 09:57:37 oak-md1-s2 kernel: LustreError: Skipped 1 previous similar message
Nov 06 09:57:37 oak-md1-s2 kernel: LustreError: 3564:0:(obd_mount_server.c:304:server_mgc_set_fs()) can&apos;t set_fs -30
Nov 06 09:57:37 oak-md1-s2 kernel: LustreError: 3564:0:(obd_mount_server.c:1879:server_fill_super()) Unable to start targets: -30
Nov 06 09:57:37 oak-md1-s2 kernel: LustreError: 3564:0:(obd_mount_server.c:1589:server_put_super()) no obd oak-MDT0000
Nov 06 09:57:37 oak-md1-s2 kernel: LustreError: 3564:0:(obd_mount_server.c:135:server_deregister_mount()) oak-MDT0000 not registered
Nov 06 09:57:37 oak-md1-s2 kernel: Lustre: server umount oak-MDT0000 complete
Nov 06 09:57:37 oak-md1-s2 kernel: LustreError: 3564:0:(obd_mount.c:1582:lustre_fill_super()) Unable to mount  (-30)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Oak is down. I&apos;m running fsck on it at the moment.&lt;br/&gt;
 Thanks,&lt;br/&gt;
 Stephane&lt;/p&gt;</comment>
                            <comment id="236480" author="sthiell" created="Tue, 6 Nov 2018 19:03:41 +0000"  >&lt;p&gt;Attached results of fsck on MDT0000 as&#160;oak-MDT0000-fsck-2018-11-06.log&lt;/p&gt;</comment>
                            <comment id="236481" author="sthiell" created="Tue, 6 Nov 2018 19:04:51 +0000"  >&lt;p&gt;The MDT crashed again when starting:&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Nov 06 10:59:34 oak-md1-s2 kernel: Lustre: oak-MDT0000: Recovery already passed deadline 0:29. If you do not want to wait more, please abort the recovery by force.
Nov 06 10:59:46 oak-md1-s2 kernel: Lustre: oak-MDT0000: Denying connection for new client 8b854071-cc99-990d-3c95-75ea796b9ff2(at 10.8.13.14@o2ib6), waiting for 1417 known clients (1391 recovered, 9 in progress, and 0 evicted) to recover in 0:16
Nov 06 10:59:46 oak-md1-s2 kernel: Lustre: Skipped 1 previous similar message
Nov 06 11:00:03 oak-md1-s2 kernel: Lustre: oak-MDT0000: recovery is timed out, evict stale exports
Nov 06 11:00:03 oak-md1-s2 kernel: Lustre: oak-MDT0000: disconnecting 17 stale clients
Nov 06 11:00:06 oak-md1-s2 kernel: LustreError: 5052:0:(osd_handler.c:1982:osd_object_release()) LBUG
Nov 06 11:00:06 oak-md1-s2 kernel: Pid: 5052, comm: tgt_recover_0
Nov 06 11:00:06 oak-md1-s2 kernel: 
                                   Call Trace:
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc07e47ae&amp;gt;] libcfs_call_trace+0x4e/0x60 [libcfs]

Message from syslogd@oak-md1-s2 at Nov  6 11:00:06 ...
 kernel:LustreError: 5052:0:(osd_handler.c:1982:osd_object_release()) LBUG
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc07e483c&amp;gt;] lbug_with_loc+0x4c/0xb0 [libcfs]
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc1017610&amp;gt;] osd_get_ldiskfs_dirent_param+0x0/0x130 [osd_ldiskfs]
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc099cc4d&amp;gt;] lu_object_put+0x17d/0x3d0 [obdclass]
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc0bf3778&amp;gt;] distribute_txn_replay_handle+0x418/0xc00 [ptlrpc]
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc0b39987&amp;gt;] replay_request_or_update.isra.21+0x247/0x8c0 [ptlrpc]
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffff810ba588&amp;gt;] ? __wake_up_common+0x58/0x90
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc0b3a616&amp;gt;] target_recovery_thread+0x616/0x1360 [ptlrpc]
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffff816a8fad&amp;gt;] ? __schedule+0x39d/0x8b0
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffffc0b3a000&amp;gt;] ? target_recovery_thread+0x0/0x1360 [ptlrpc]
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffff810b098f&amp;gt;] kthread+0xcf/0xe0
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffff810b08c0&amp;gt;] ? kthread+0x0/0xe0
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffff816b4f58&amp;gt;] ret_from_fork+0x58/0x90
Nov 06 11:00:06 oak-md1-s2 kernel:  [&amp;lt;ffffffff810b08c0&amp;gt;] ? kthread+0x0/0xe0
Nov 06 11:00:06 oak-md1-s2 kernel: 
Nov 06 11:00:06 oak-md1-s2 kernel: Kernel panic - not syncing: LBUG
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="236483" author="sthiell" created="Tue, 6 Nov 2018 19:32:09 +0000"  >&lt;p&gt;I tried a second time after shutting down the client that ran the &lt;tt&gt;lfs migrate&lt;/tt&gt; but the same LBUG happened. I was able to put it back into production by trying a third time and doing an abort_recovery after most of the clients recovered.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[  156.180372] Lustre: oak-MDT0000: Denying connection for new client cdb722e9-f2fd-2e1e-3be0-ae3ca5c64dbb(at 10.8.15.2@o2ib6), waiting for 1417 known clients (1406 recovered, 9 in progress, and 0 evicted) to recover in 8:45
[  156.202310] Lustre: Skipped 1 previous similar message
[  159.651156] LustreError: 3530:0:(mdt_handler.c:6167:mdt_iocontrol()) oak-MDT0000: Aborting recovery for device
[  159.662323] LustreError: 3530:0:(ldlm_lib.c:2606:target_stop_recovery_thread()) oak-MDT0000: Aborting recovery
[  159.673500] Lustre: 3850:0:(ldlm_lib.c:2059:target_recovery_overseer()) recovery is aborted, evict exports in recovery
[  159.685661] Lustre: oak-MDT0000: disconnecting 11 stale clients
[  159.720906] Lustre: 3850:0:(ldlm_lib.c:1619:abort_req_replay_queue()) @@@ aborted:  req@ffff88100d339200 x1616374803034688/t0(279195624051) o36-&amp;gt;ebc6458d-1c97-7178-4cec-be2598c570b5@10.9.114.12@o2ib4:319/0 lens 624/0 e 0 to 0 dl 1541532139 ref 1 fl Complete:/4/ffffffff rc 0/-1
[  159.748315] LustreError: 3850:0:(ldlm_lib.c:1640:abort_lock_replay_queue()) @@@ aborted:  req@ffff88102435a100 x1610729933017152/t0(0) o101-&amp;gt;10a6d338-b751-f002-23d5-4f33526aa200@10.12.4.29@o2ib:537/0 lens 328/0 e 0 to 0 dl 1541532357 ref 1 fl Complete:/40/ffffffff rc 0/-1
[  159.789840] Lustre: oak-MDT0000: Recovery over after 0:55, of 1417 clients 1406 recovered and 11 were evicted.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I have two vmcore&apos;s of the above LBUG and system logs also available if needed.&lt;/p&gt;

&lt;p&gt;Best,&lt;br/&gt;
Stephane&lt;/p&gt;</comment>
                            <comment id="237021" author="sthiell" created="Thu, 15 Nov 2018 07:44:56 +0000"  >&lt;p&gt;Hi YangSheng,&lt;/p&gt;

&lt;p&gt;Any ideas of what could have happened here?&lt;/p&gt;

&lt;p&gt;Thanks!&lt;br/&gt;
Stephane&lt;/p&gt;</comment>
                            <comment id="237085" author="ys" created="Fri, 16 Nov 2018 06:51:21 +0000"  >&lt;p&gt;Hi, Stephane,&lt;/p&gt;

&lt;p&gt;Sorry for the late reply, Could you please upload vmcore on our ftp site? ftp.whamcloud.com. This crash is really stranger for me. Also debuginfo rpms if you still keep them. If not, as comment in this ticket, you should use 2.10.4 + 3.10.0-692.2.2 kernel for your MDS?&lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
YangSheng&lt;/p&gt;</comment>
                            <comment id="237109" author="sthiell" created="Fri, 16 Nov 2018 19:09:21 +0000"  >&lt;p&gt;Hi YangSheng,&lt;/p&gt;

&lt;p&gt;Sure, thanks for taking a look. I just uploaded the vmcore and matching debuginfo rpms:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;&lt;tt&gt;lustre-debuginfo-2.10.4-1.el7.centos.x86_64.rpm&lt;/tt&gt;&lt;/li&gt;
	&lt;li&gt;&lt;tt&gt;kernel-debuginfo-3.10.0-693.2.2.el7_lustre.pl1.x86_64.rpm&lt;/tt&gt;&lt;/li&gt;
	&lt;li&gt;&lt;tt&gt;kernel-debuginfo-common-x86_64-3.10.0-693.2.2.el7_lustre.pl1.x86_64.rpm&lt;/tt&gt;&lt;/li&gt;
	&lt;li&gt;and one of the vmcores as &lt;tt&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-11578&quot; title=&quot;ldiskfs_map_blocks: comm mdt00_100: lblock 0 mapped to illegal pblock&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-11578&quot;&gt;&lt;del&gt;LU-11578&lt;/del&gt;&lt;/a&gt;-vmcore-oak-md1-s2-2018-11-06-11-00-23&lt;/tt&gt;&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Thanks!&lt;br/&gt;
Stephane&lt;/p&gt;</comment>
                            <comment id="237553" author="sthiell" created="Tue, 27 Nov 2018 20:03:03 +0000"  >&lt;p&gt;Hey YangSheng,&lt;/p&gt;

&lt;p&gt;Have you been able to take a look at this? Let me know if you need anything else from me.&lt;/p&gt;

&lt;p&gt;Thanks!&lt;br/&gt;
Stephane&lt;/p&gt;</comment>
                            <comment id="237600" author="ys" created="Wed, 28 Nov 2018 16:31:41 +0000"  >&lt;p&gt;Hi, Stephane,&lt;/p&gt;

&lt;p&gt;Sorry for the late. I have downloaded the packages. Will start to analyze later. Thanks for the patience. &lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
YangSheng&lt;/p&gt;</comment>
                            <comment id="238461" author="ys" created="Wed, 12 Dec 2018 18:21:44 +0000"  >&lt;p&gt;Hi, Stephane,&lt;/p&gt;

&lt;p&gt;I have got some data from vmcore as below:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;crash&amp;gt; struct osd_object ffff88202b660400 -x
struct osd_object {
  oo_dt = {
    do_lu = {
      lo_header = 0xffff881c939377b0,
      lo_dev = 0xffff882027e22000,
      lo_ops = 0xffffffffc1058060 &amp;lt;osd_lu_obj_ops&amp;gt;,
      lo_linkage = {
        next = 0xffff881c939377f0,
        prev = 0xffff881c87ad52c8
      },
      lo_dev_ref = {&amp;lt;No data fields&amp;gt;}
    },
    do_ops = 0xffffffffc1057f60 &amp;lt;osd_obj_ops&amp;gt;,
    do_body_ops = 0xffffffffc1059b20 &amp;lt;osd_body_ops&amp;gt;,
    do_index_ops = 0x0
  },
  oo_inode = 0xffff88125f9d9a38,
  oo_hl_head = 0xffff881c87497200,
  oo_ext_idx_sem = {
    {
      count = {
        counter = 0x0
      },
      __UNIQUE_ID_rh_kabi_hide3 = {
        count = 0x0
      },
      {&amp;lt;No data fields&amp;gt;}
    },
    wait_lock = {
      raw_lock = {
        val = {
          counter = 0x0
        }
      }
    },
    osq = {
      tail = {
        counter = 0x0
      }
    },
    wait_list = {
      next = 0xffff88202b660460
    },
    owner = 0x0
  },
  oo_sem = {
    {
      count = {
        counter = 0x0
      },
      __UNIQUE_ID_rh_kabi_hide3 = {
        count = 0x0
      },
      {&amp;lt;No data fields&amp;gt;}
    },
    wait_lock = {
      raw_lock = {
        val = {
          counter = 0x0
        }
      }
    },
    osq = {
      tail = {
        counter = 0x0
      }
    },
    wait_list = {
      next = 0xffff88202b660480
    },
    owner = 0x0
  },
  oo_dir = 0x0,
  oo_guard = {
    {
      rlock = {
        raw_lock = {
          val = {
            counter = 0x0
          }
        }
      }
    }
  },
  oo_destroyed = 0x0,
  oo_lma_flags = 0x0,
  oo_compat_dot_created = 0x1,
  oo_compat_dotdot_created = 0x1,
  oo_owner = 0x0,
  oo_xattr_list = {
    next = 0xffff88202b6604b8,
    prev = 0xffff88202b6604b8
  }
}
.....
crash&amp;gt; struct inode 0xffff88125f9d9a38
struct inode {
  i_mode = 33184,
  i_opflags = 0,
  i_uid = {
    val = 282232
  },
  i_gid = {
    val = 3199
  },
  i_flags = 128,
  i_acl = 0xffffffffffffffff,
  i_default_acl = 0xffffffffffffffff,
  i_op = 0xffffffffc0f5afc0 &amp;lt;ldiskfs_file_inode_operations&amp;gt;,
  i_sb = 0xffff8802d8e7b000,
  i_mapping = 0xffff88125f9d9b88,
  i_security = 0x0,
  i_ino = 564154962,
  {
    i_nlink = 0,
    __i_nlink = 0
  },
  i_rdev = 0,
  i_size = 0,
  i_atime = {
    tv_sec = 1501119951,
    tv_nsec = 0
  },
...........
crash&amp;gt; struct lu_object_header 0xffff881c939377b0 -x
struct lu_object_header {
  loh_fid = {
    f_seq = 0x200002b64,
    f_oid = 0xce11,
    f_ver = 0x0
  },
  loh_flags = 0x0,
  loh_ref = {
    counter = 0x0
  },
  loh_attr = 0x8001,
  loh_hash = {
    next = 0x0,
    pprev = 0xffffc90016edd928
  },
  loh_lru = {
    next = 0xffff881c939377e0,
    prev = 0xffff881c939377e0
  },
  loh_layers = {
    next = 0xffff881c93937818,
    prev = 0xffff88202b660418
  },
  loh_reference = {&amp;lt;No data fields&amp;gt;}
}
..........
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;So please help collect info as below:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;on MDS:
# debugfs [your mds device]
debugfs: ncheck 564154962
Inode	Pathname
564154962   [pathname]
debugfs: stat [pathname]
paste output in here
......
On client:
# lfs fid2path &amp;lt;your mount point&amp;gt; [0x200002b64:0xce11:0x0]

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
YangSheng&lt;/p&gt;</comment>
                            <comment id="238486" author="sthiell" created="Wed, 12 Dec 2018 20:46:08 +0000"  >&lt;p&gt;Hi YangSheng!&lt;/p&gt;

&lt;p&gt;Results:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@oak-md1-s2 ~]# debugfs /dev/mapper/md1-rbod1-mdt0
debugfs 1.44.3.wc1 (23-July-2018)
debugfs:  ncheck 564154962
Inode	Pathname
564154962	/REMOTE_PARENT_DIR/0x200002b64:0x65a2:0x0/openssl-1.0.2l-0/ssl/cacert.pem
debugfs:  stat /REMOTE_PARENT_DIR/0x200002b64:0x65a2:0x0/openssl-1.0.2l-0/ssl/cacert.pem
Inode: 564154962   Type: regular    Mode:  0640   Flags: 0x0
Generation: 529132813    Version: 0x00000005:44289388
User: 282232   Group:  3199   Size: 0
File ACL: 0
Links: 1   Blockcount: 0
Fragment:  Address: 0    Number: 0    Size: 0
 ctime: 0x59790e8c:00000000 -- Wed Jul 26 14:50:04 2017
 atime: 0x597945cf:00000000 -- Wed Jul 26 18:45:51 2017
 mtime: 0x59270fc5:00000000 -- Thu May 25 10:09:25 2017
crtime: 0x59790e35:e633c9d8 -- Wed Jul 26 14:48:37 2017
Size of extra inode fields: 28
Extended attributes:
  system.posix_acl_access (28) = 01 00 00 00 01 00 06 00 04 00 07 00 08 00 07 00 7f 0c 00 00 10 00 04 00 20 00 00 00 
  trusted.lma (24) = 00 00 00 00 00 00 00 00 64 2b 00 00 02 00 00 00 11 ce 00 00 00 00 00 00 
  lma: fid=[0x200002b64:0xce11:0x0] compat=0 incompat=0
  trusted.lov (56)
  trusted.link (80)
BLOCKS:

debugfs:
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;On a client:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@oak-gw02 ~]# lfs fid2path /oak [0x200002b64:0xce11:0x0]
/oak/stanford/groups/ruthm/sthiell/anaconda2.off/pkgs/openssl-1.0.2l-0/ssl/cacert.pem
/oak/stanford/groups/ruthm/sthiell/anaconda2.off/ssl/cacert.pem
[root@oak-gw02 ~]# 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="238518" author="ys" created="Thu, 13 Dec 2018 05:03:37 +0000"  >&lt;p&gt;Hi, Stephane,&lt;/p&gt;

&lt;p&gt;Looks like it is similar  with &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10581&quot; title=&quot;osd_handler.c:1978:osd_object_release()) LBUG&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10581&quot;&gt;LU-10581&lt;/a&gt;, Some code dec nlink and then fail to exit but leave inode as is. So the cache is  inconsistent  with on-disk data. I&apos;ll try to dig out where it occurred. Thanks for your help.&lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
YangSheng&lt;/p&gt;</comment>
                            <comment id="341528" author="ys" created="Tue, 26 Jul 2022 03:56:38 +0000"  >&lt;p&gt;Close as duplicated with &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-10581&quot; title=&quot;osd_handler.c:1978:osd_object_release()) LBUG&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-10581&quot;&gt;LU-10581&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="56220">LU-12485</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="50469">LU-10581</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="57839">LU-13157</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="63289">LU-14511</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="31430" name="oak-MDT0000-fsck-2018-11-06.log" size="20979" author="sthiell" created="Tue, 6 Nov 2018 19:03:15 +0000"/>
                            <attachment id="31348" name="oak-MDT0000_fsck.log" size="305753" author="sthiell" created="Sat, 27 Oct 2018 04:32:14 +0000"/>
                            <attachment id="31349" name="oak-md1-s2_kernel.log" size="2600584" author="sthiell" created="Sat, 27 Oct 2018 04:30:41 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i0058f:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>