<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:02:22 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-6687] ALL osp-sync in D state</title>
                <link>https://jira.whamcloud.com/browse/LU-6687</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;After we reboot and mount mdt we see all osp-sync threads in D state and the following errors.&lt;/p&gt;

&lt;p&gt;Jun  3 14:38:03 nbp8-mds1 kernel: LustreError: 5838:0:(osp_sync.c:487:osp_sync_new_setattr_job()) nbp8-OST009e-osc-MDT0000: invalid setattr record, lsr_valid:100&lt;br/&gt;
Jun  3 14:38:03 nbp8-mds1 kernel: LustreError: 5838:0:(osp_sync.c:487:osp_sync_new_setattr_job()) Skipped 1554778 previous similar messages&lt;br/&gt;
Jun  3 14:40:11 nbp8-mds1 kernel: LustreError: 6043:0:(osp_sync.c:487:osp_sync_new_setattr_job()) nbp8-OST010e-osc-MDT0000: invalid setattr record, lsr_valid:8191&lt;br/&gt;
Jun  3 14:40:11 nbp8-mds1 kernel: LustreError: 6043:0:(osp_sync.c:487:osp_sync_new_setattr_job()) Skipped 2608741 previous similar messages&lt;br/&gt;
Jun  3 14:44:27 nbp8-mds1 kernel: LustreError: 6342:0:(osp_sync.c:487:osp_sync_new_setattr_job()) nbp8-OST0133-osc-MDT0000: invalid setattr record, lsr_valid:8191&lt;br/&gt;
Jun  3 14:44:27 nbp8-mds1 kernel: LustreError: 6342:0:(osp_sync.c:487:osp_sync_new_setattr_job()) Skipped 10802201 previous similar messages&lt;br/&gt;
Jun  3 14:47:16 nbp8-mds1 pcp-pmie&lt;span class=&quot;error&quot;&gt;&amp;#91;4713&amp;#93;&lt;/span&gt;: High 1-minute load average 321load@nbp8-mds1&lt;br/&gt;
Jun  3 14:52:59 nbp8-mds1 kernel: LustreError: 6220:0:(osp_sync.c:487:osp_sync_new_setattr_job()) nbp8-OST0045-osc-MDT0000: invalid setattr record, lsr_valid:68&lt;br/&gt;
Jun  3 14:52:59 nbp8-mds1 kernel: LustreError: 6220:0:(osp_sync.c:487:osp_sync_new_setattr_job()) Skipped 58201221 previous similar messages&lt;/p&gt;</description>
                <environment></environment>
        <key id="30497">LU-6687</key>
            <summary>ALL osp-sync in D state</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="2">Won&apos;t Fix</resolution>
                                        <assignee username="niu">Niu Yawei</assignee>
                                    <reporter username="mhanafi">Mahmoud Hanafi</reporter>
                        <labels>
                    </labels>
                <created>Wed, 3 Jun 2015 21:54:06 +0000</created>
                <updated>Fri, 16 Oct 2015 04:26:14 +0000</updated>
                            <resolved>Fri, 16 Oct 2015 04:26:14 +0000</resolved>
                                    <version>Lustre 2.5.3</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="117381" author="pjones" created="Thu, 4 Jun 2015 03:46:58 +0000"  >&lt;p&gt;Niu&lt;/p&gt;

&lt;p&gt;Could you please advise?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="117404" author="niu" created="Thu, 4 Jun 2015 06:42:36 +0000"  >&lt;p&gt;Looks the related patch was applied correctly in your 2.5.3 tree (&lt;a href=&quot;https://github.com/jlan/lustre-nas/commit/fb970b342a7fac22a17b4932e11febb6963b3dff&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://github.com/jlan/lustre-nas/commit/fb970b342a7fac22a17b4932e11febb6963b3dff&lt;/a&gt;) &lt;/p&gt;

&lt;p&gt;Is this an upgraded system? and it&apos;s the first mount after upgrading? I&apos;m wondering if these invalid records were some leftover from old system.&lt;/p&gt;</comment>
                            <comment id="117406" author="niu" created="Thu, 4 Jun 2015 06:47:04 +0000"  >&lt;p&gt;BTW, because of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5297&quot; title=&quot;osp_sync_thread can&amp;#39;t handle invalid record gracefully&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5297&quot;&gt;&lt;del&gt;LU-5297&lt;/del&gt;&lt;/a&gt;, osp sync thread can&apos;t handle invalid records properly.&lt;/p&gt;</comment>
                            <comment id="117408" author="mhanafi" created="Thu, 4 Jun 2015 07:37:56 +0000"  >&lt;p&gt;This is a upgraded system to 2.5.3. This happens every time when the mdt is mounted. How do we go about fixing the invalid records?&lt;/p&gt;

&lt;p&gt;We are going and cleanup a lot of mismatch between object UID/GID and MDT records. These occurred most likely due to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6574&quot; title=&quot;chown doesn&amp;#39;t update object ownership on ost&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6574&quot;&gt;&lt;del&gt;LU-6574&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="117428" author="niu" created="Thu, 4 Jun 2015 13:02:14 +0000"  >&lt;p&gt;Hmm, fixing these invalid records manually will be troublesome (there isn&apos;t any llog edit tool, so you have to use hex edit to modify the records...)&lt;/p&gt;

&lt;p&gt;Actually, if there are only few leftover records, we can just delete all of them by removing the llog files, then we can move on to mount the MDT.&lt;/p&gt;

&lt;p&gt;1. lctl --device $MDTDEV llog_catlist to show all the catalogs for the unlink/setattr records;&lt;br/&gt;
2. mount mdt as ldiskfs and find all the catalogs under /O;&lt;br/&gt;
3. use llog_reader to show plain logs belong to these catalogs;&lt;br/&gt;
4. use llog_reader to see how many leftover records in the plain logs;&lt;br/&gt;
5. remove all the plain logs and catalogs, they will be recreated on next mount;&lt;/p&gt;</comment>
                            <comment id="117489" author="mhanafi" created="Thu, 4 Jun 2015 19:30:32 +0000"  >&lt;p&gt;nbp8-mds1 ~ # lctl --device 6 llog_catlist&lt;br/&gt;
OBD_IOC_CATLOGLIST failed: Operation not supported&lt;/p&gt;

&lt;p&gt;it looks like there are a lot of records. I am not sure If i understand item #4 and #5&lt;/p&gt;</comment>
                            <comment id="117543" author="niu" created="Fri, 5 Jun 2015 03:13:32 +0000"  >&lt;p&gt;Hmm, looks llog_catlist is only available in master now.&lt;/p&gt;

&lt;p&gt;Ok, each chown &amp;amp; unlink on MDT will generate a llog record in llog file, and this record will be used to sync the operations to OST objects, once the sync to OST done, the record will be removed from the llog file. Usually after a clean shutdown, there won&apos;t be any leftover records in the llog files. However, in your case, there are some invalid records which can&apos;t be processed and not removed at the end.&lt;/p&gt;

&lt;p&gt;Let&apos;s look at the on disk structure of llog files:&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;There is a global CATLOGLIST file which storing all the catalog IDs for each OST; (&quot;lctl llog_catlist&quot; can print the content of this file)&lt;/li&gt;
	&lt;li&gt;Each catalog is a plain log index, which stores plain log IDs;&lt;/li&gt;
	&lt;li&gt;Each plain log stores the unlink/setattr records I mentioned above;&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;I think that could make it easier for understanding the #4 &amp;amp; #5 of my previous comment?&lt;/p&gt;

&lt;p&gt;Given the &quot;lctl llog_catlist&quot; isn&apos;t supported in 2.5, you can remove all the leftover records as following: mount mdt as ldiskfs, find all the files which name is numerical under /O/1/ and remove them all. (it&apos;s better to backup these files)&lt;/p&gt;</comment>
                            <comment id="130574" author="mhanafi" created="Fri, 16 Oct 2015 00:08:08 +0000"  >&lt;p&gt;Please close this case&lt;/p&gt;</comment>
                            <comment id="130591" author="pjones" created="Fri, 16 Oct 2015 04:26:14 +0000"  >&lt;p&gt;ok Mahmoud&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="25440">LU-5297</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxevr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>