<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:45:04 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-11574] LustreError: 157-3: Trying to start OBD nbp13-OST000b_UUID using the wrong disk</title>
                <link>https://jira.whamcloud.com/browse/LU-11574</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Filesystem is down.&lt;/p&gt;

&lt;p&gt;We had raid backend issues and oss crashed. We ran run fsck after crash. when trying to remount ost we get this error. I have tried to run fsck several times.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;

 [ 4993.782431] LustreError: 157-3: Trying to start OBD nbp13-OST000b_UUID using the wrong disk . Were the /dev/ assignments rearranged?
[ 4993.825963] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) header@ffff882dcd3a2f00[0x0, 1, [0x1:0x0:0x0] hash exist]{
[ 4993.825963] 
[ 4993.864146] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....local_storage@ffff882dcd3a2f50
[ 4993.864146] 
[ 4993.896057] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....osd-ldiskfs@ffff882df011ba00osd-ldiskfs-object@ffff882df011ba00(i:ffff882b7af88958:78/2138703796)[plain]
[ 4993.896057] 
[ 4993.947314] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) } header@ffff882dcd3a2f00
[ 4993.947314] 
[ 4993.978176] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) header@ffff882dcd3a3140[0x0, 1, [0x200000003:0x0:0x0] hash exist]{
[ 4993.978176] 
[ 4994.018450] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....local_storage@ffff882dcd3a3190
[ 4994.018450] 
[ 4994.050361] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....osd-ldiskfs@ffff882eee0e8d00osd-ldiskfs-object@ffff882eee0e8d00(i:ffff882b7af6f2d0:77/2138703762)[plain]
[ 4994.050361] 
[ 4994.101617] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) } header@ffff882dcd3a3140
[ 4994.101617] 
[ 4994.137076] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) header@ffff882dcd3a2c00[0x0, 1, [0xa:0x0:0x0] hash exist]{
[ 4994.137076] 
[ 4994.175261] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....local_storage@ffff882dcd3a2c50
[ 4994.175261] 
[ 4994.207171] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....osd-ldiskfs@ffff882eee0e9100osd-ldiskfs-object@ffff882eee0e9100(i:ffff882b7af90d90:79/2138703830)[plain]
[ 4994.207171] 
[ 4994.258427] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) } header@ffff882dcd3a2c00
[ 4994.258427] 
[ 4994.299979] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) header@ffff882dcd3a2e40[0x0, 1, [0x200000001:0x1017:0x0] hash exist]{
[ 4994.299979] 
[ 4994.341038] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....local_storage@ffff882dcd3a2e90
[ 4994.341038] 
[ 4994.372949] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....osd-ldiskfs@ffff882df011a600osd-ldiskfs-object@ffff882df011a600(i:ffff882c098426e0:1090561/1004891530)[plain]
[ 4994.372949] 
[ 4994.425511] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) } header@ffff882dcd3a2e40
[ 4994.425511] 
[ 4994.464017] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) header@ffff882bfaae7380[0x0, 1, [0xa:0x18:0x0] hash exist]{
[ 4994.464017] 
[ 4994.502461] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....local_storage@ffff882bfaae73d0
[ 4994.502461] 
[ 4994.534372] LustreError: 17122:0:(ofd_dev.c:251:ofd_stack_fini()) ....osd-ldiskfs@ffff882edbe28200osd-ldiskfs-object@ffff882edbe28200(i:ffff882b7af91600:80/2939569139)[plain]
[ 4994.534372] 
[ 4994.571667] Lustre: nbp13-OST000b: Not available &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; connect from 10.151.25.231@o2ib (not set up)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="53829">LU-11574</key>
            <summary>LustreError: 157-3: Trying to start OBD nbp13-OST000b_UUID using the wrong disk</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="mhanafi">Mahmoud Hanafi</reporter>
                        <labels>
                    </labels>
                <created>Fri, 26 Oct 2018 08:21:37 +0000</created>
                <updated>Fri, 4 Sep 2020 17:44:12 +0000</updated>
                            <resolved>Tue, 30 Oct 2018 12:57:12 +0000</resolved>
                                    <version>Lustre 2.10.5</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="235576" author="jgmitter" created="Fri, 26 Oct 2018 10:56:20 +0000"  >&lt;p&gt;We have engineers looking into the issue and will provide an update ASAP.&lt;/p&gt;</comment>
                            <comment id="235578" author="bzzz" created="Fri, 26 Oct 2018 11:33:46 +0000"  >&lt;p&gt;last_rcvd seem to be corrupted. please wait few minutes, I&apos;ll try to reproduce locally and figure out a solution.&lt;/p&gt;</comment>
                            <comment id="235580" author="bobijam" created="Fri, 26 Oct 2018 11:36:27 +0000"  >&lt;p&gt;Looks like the last_rcvd file of the OST target got corrupted, can you backup the target as a fail-safe, and mount the target as ldiskfs, then manually delete the last_rcvd file, umount it, and try to remount the ost normally again?&lt;/p&gt;</comment>
                            <comment id="235581" author="bzzz" created="Fri, 26 Oct 2018 11:40:59 +0000"  >&lt;p&gt;I&apos;d suggest to make a copy of last_rcvd ..&lt;br/&gt;
at least locally that worked fine - mount as ldiskfs, save last_rcvd, rm the original one, umount &lt;br/&gt;
 and mount as Lustre.&lt;/p&gt;</comment>
                            <comment id="235584" author="mhanafi" created="Fri, 26 Oct 2018 13:07:09 +0000"  >&lt;p&gt;what do you mean&lt;/p&gt;

&lt;p&gt;&quot;at least locally that worked fine&quot;?&#160;&lt;/p&gt;

&lt;p&gt;just to verify.&lt;/p&gt;

&lt;p&gt;1. Mount the OSTs that are having the issue as ldiskfs.&lt;/p&gt;

&lt;p&gt;2. copy last_rcvd&lt;/p&gt;

&lt;p&gt;3. rm last_rcvd from each ost&lt;/p&gt;

&lt;p&gt;4. umount ost&lt;/p&gt;

&lt;p&gt;5. and remount as lustre.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="235595" author="mhanafi" created="Fri, 26 Oct 2018 15:05:48 +0000"  >&lt;p&gt;A releated issue is some of the OST have wrong free space.&lt;/p&gt;

&lt;p&gt;Here we have nbp15_1-OST13&#160; says has 11TB used but this was a unused OST.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;nbp15-srv2 /mnt/lustre/nbp15_1-OST13 # du -sk &lt;br/&gt;
 4311660 .&lt;/p&gt;

&lt;p&gt;nbp15-srv2 /mnt/lustre/nbp15_1-OST13 # df -h&lt;br/&gt;
 /dev/mapper/nbp15_1-OST13 72T 11T 62T 15% /mnt/lustre/nbp15_1-OST13&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
 nbp15-srv2 ~ # e2fsck -vf /dev/mapper/nbp15_1-OST13
e2fsck 1.42.13.wc6 (05-Feb-2017)
Pass 1: Checking inodes, blocks, and sizes
Pass 2: Checking directory structure
Pass 3: Checking directory connectivity
Pass 4: Checking reference counts
Pass 5: Checking group summary information       80857 inodes used (0.87%, out of 9337344)
           8 non-contiguous files (0.0%)
           0 non-contiguous directories (0.0%)
             # of inodes with ind/dind/tind blocks: 0/0/0
             Extent depth histogram: 80847/2
  2703993403 blocks used (14.14%, out of 19122880512)
           0 bad blocks
           0 large files       80709 regular files
         139 directories
           0 character device files
           0 block device files
           0 fifos
           0 links
           0 symbolic links (0 fast symbolic links)
           0 sockets
------------
       80848 files
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="235600" author="utopiabound" created="Fri, 26 Oct 2018 17:03:36 +0000"  >&lt;blockquote&gt;&lt;p&gt;1. Mount the OSTs that are having the issue as ldiskfs.&lt;/p&gt;

&lt;p&gt;2. copy last_rcvd&lt;/p&gt;

&lt;p&gt;3. rm last_rcvd from each ost&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;Just from the mounted OST that&apos;s having the issue&lt;/p&gt;
&lt;blockquote&gt;&lt;p&gt;4. umount ost&lt;/p&gt;

&lt;p&gt;5. and remount as lustre.&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;Correct.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;With regards to free space.&#160; I would check &lt;tt&gt;lfs df&lt;/tt&gt; from a client to see where lustre shows free space.&#160; If that doesn&apos;t clear things up, please open a separate ticket.&lt;/p&gt;</comment>
                            <comment id="235613" author="mhanafi" created="Fri, 26 Oct 2018 20:48:17 +0000"  >&lt;p&gt;This workaround work. The prio can be lowered&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                            <comment id="235614" author="adilger" created="Fri, 26 Oct 2018 21:42:43 +0000"  >&lt;p&gt;Also, have you run a full e2fsck after the RAID problems?  If not, then it is good to save the output (i.e. run under &quot;&lt;tt&gt;script&lt;/tt&gt;&quot; or similar).&lt;/p&gt;</comment>
                            <comment id="235906" author="mhanafi" created="Tue, 30 Oct 2018 12:04:01 +0000"  >&lt;p&gt;yes we did run full fsck.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;

&lt;p&gt;Please close the case.&lt;/p&gt;

&lt;p&gt;&#160;&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i0056n:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10020"><![CDATA[1]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>