<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:29:44 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2959] ASSERTION( cli-&gt;cl_mgc_configs_dir ) for 200 osts x 2 oss</title>
                <link>https://jira.whamcloud.com/browse/LU-2959</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Hit LBUG on one of the oss.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LustreError: 16207:0:(mgc_request.c:1682:mgc_copy_llog()) ASSERTION( cli-&amp;gt;cl_mgc_configs_dir ) failed:
LustreError: 16207:0:(mgc_request.c:1682:mgc_copy_llog()) LBUG
Pid: 16207, comm: ll_cfg_requeue

Call Trace:
 [&amp;lt;ffffffffa04e2895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
 [&amp;lt;ffffffffa04e2e97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
 [&amp;lt;ffffffffa0dbfe2b&amp;gt;] mgc_process_cfg_log+0x134b/0x15c0 [mgc]
 [&amp;lt;ffffffffa0dc2093&amp;gt;] mgc_process_log+0x463/0x1390 [mgc]
 [&amp;lt;ffffffff814ead1a&amp;gt;] ? schedule_timeout+0x19a/0x2e0
 [&amp;lt;ffffffffa0dbca30&amp;gt;] ? mgc_blocking_ast+0x0/0x770 [mgc]
 [&amp;lt;ffffffffa07aed40&amp;gt;] ? ldlm_completion_ast+0x0/0x960 [ptlrpc]
 [&amp;lt;ffffffff81090d4c&amp;gt;] ? remove_wait_queue+0x3c/0x50
 [&amp;lt;ffffffffa0dc3973&amp;gt;] mgc_requeue_thread+0x1a3/0x750 [mgc]
 [&amp;lt;ffffffff8105fa40&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffffa0dc37d0&amp;gt;] ? mgc_requeue_thread+0x0/0x750 [mgc]
 [&amp;lt;ffffffff8100c0ca&amp;gt;] child_rip+0xa/0x20
 [&amp;lt;ffffffffa0dc37d0&amp;gt;] ? mgc_requeue_thread+0x0/0x750 [mgc]
 [&amp;lt;ffffffffa0dc37d0&amp;gt;] ? mgc_requeue_thread+0x0/0x750 [mgc]
 [&amp;lt;ffffffff8100c0c0&amp;gt;] ? child_rip+0x0/0x20

LustreError: dumping log to /tmp/lustre-log.1363211399.16207
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>while testing wide striping with 1mds, 2 oss. Each oss is creating 200 osts; and mount at the same time using a script.</environment>
        <key id="17885">LU-2959</key>
            <summary>ASSERTION( cli-&gt;cl_mgc_configs_dir ) for 200 osts x 2 oss</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="utopiabound">Nathaniel Clark</assignee>
                                    <reporter username="mdiep">Minh Diep</reporter>
                        <labels>
                            <label>LB</label>
                    </labels>
                <created>Wed, 13 Mar 2013 18:33:49 +0000</created>
                <updated>Fri, 20 Jun 2014 07:19:16 +0000</updated>
                            <resolved>Fri, 18 Oct 2013 16:51:29 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                    <fixVersion>Lustre 2.5.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="54777" author="mdiep" created="Mon, 25 Mar 2013 18:01:19 +0000"  >&lt;p&gt;I haven&apos;t been able to reproduce this on latest master. We need to check with Alex to see if there&apos;s fix recently about this&lt;/p&gt;</comment>
                            <comment id="54788" author="mdiep" created="Mon, 25 Mar 2013 19:49:36 +0000"  >&lt;p&gt;I used &lt;a href=&quot;https://build.whamcloud.com/job/lustre-master/1337/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://build.whamcloud.com/job/lustre-master/1337/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="54861" author="bzzz" created="Tue, 26 Mar 2013 18:02:13 +0000"  >&lt;p&gt;sorry, i&apos;m not aware of any fix in this area.&lt;/p&gt;</comment>
                            <comment id="54866" author="utopiabound" created="Tue, 26 Mar 2013 20:50:17 +0000"  >&lt;p&gt;I&apos;m inclined to believe this is a race between lustre_fill_super() and LDLM_CB_CANCELING. I think it&apos;s a real bug, but I think it&apos;s very hard to get to. Many OST&apos;s on a single OSS helps exacerbate the race.  I haven&apos;t nailed down the exact chain, yet.&lt;/p&gt;</comment>
                            <comment id="54939" author="pjones" created="Wed, 27 Mar 2013 17:49:52 +0000"  >&lt;p&gt;Dropping priority based on rarity&lt;/p&gt;</comment>
                            <comment id="54952" author="utopiabound" created="Wed, 27 Mar 2013 19:15:47 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/5860&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5860&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="55161" author="liwei" created="Mon, 1 Apr 2013 03:38:12 +0000"  >&lt;p&gt;It seems the MGC &quot;fs&quot; stuff, like cl_mgc_configs_dir, are time-shared among different OSTs.  Each OST sets &quot;fs&quot; and cleans &quot;fs&quot; up in server_start_targets().  mgc_requeue_thread() should set up the &quot;fs&quot; itself, but does not seem to be doing so.&lt;/p&gt;</comment>
                            <comment id="55166" author="tappro" created="Mon, 1 Apr 2013 06:29:22 +0000"  >&lt;p&gt;it doesn&apos;t but it checks that fs is proper one and don&apos;t process local log otherwise mgc_process_cfg_log():&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;	&lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (lctxt &amp;amp;&amp;amp; lsi &amp;amp;&amp;amp; IS_SERVER(lsi) &amp;amp;&amp;amp;
            (lsi-&amp;gt;lsi_srv_mnt == cli-&amp;gt;cl_mgc_vfsmnt) &amp;amp;&amp;amp;
	    !IS_MGS(lsi) &amp;amp;&amp;amp; lsi-&amp;gt;lsi_srv_mnt) {
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="55168" author="bzzz" created="Mon, 1 Apr 2013 06:58:39 +0000"  >&lt;p&gt;I think we do not update backups on a subsequent requeue (after initial during service startup). so, we don&apos;t need to setup/clear fs in requeue.&lt;br/&gt;
but given lsi_srv_mnt is unique, we shouldn&apos;t be calling mgc_copy_llog() on requeue.&lt;/p&gt;

&lt;p&gt;interesting thing is that mgc_fs_setup() takes cl_mgc_sem to serialize setup procedues and llog processing for different services, then mgc_fs_cleanup() release it. but requeue doesn&apos;t do this.&lt;/p&gt;</comment>
                            <comment id="56363" author="cliffw" created="Tue, 16 Apr 2013 00:02:54 +0000"  >&lt;p&gt;I have hit this assertion on Hyperion, testing 2.3.64 tag. Error occurred when attempting to mount a freshly formatted disk:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Apr 15 16:55:57 hyperion-dit31 kernel: Lustre: lustre-OST0032: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-450
Apr 15 16:55:57 hyperion-dit31 kernel: LustreError: 5369:0:(mgc_request.c:1686:mgc_copy_llog()) ASSERTION( cli-&amp;gt;cl_mgc_configs_dir ) failed:
Apr 15 16:55:57 hyperion-dit31 kernel: LustreError: 5369:0:(mgc_request.c:1686:mgc_copy_llog()) ASSERTION( cli-&amp;gt;cl_mgc_configs_dir ) failed:
Apr 15 16:55:57 hyperion-dit31 kernel: LustreError: 5369:0:(mgc_request.c:1686:mgc_copy_llog()) LBUG
Apr 15 16:55:57 hyperion-dit31 kernel: LustreError: 5369:0:(mgc_request.c:1686:mgc_copy_llog()) LBUG
Apr 15 16:55:57 hyperion-dit31 kernel: Pid: 5369, comm: ll_cfg_requeue
Apr 15 16:55:57 hyperion-dit31 kernel:
Apr 15 16:55:57 hyperion-dit31 kernel: Call Trace:
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa054c895&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa054ce97&amp;gt;] lbug_with_loc+0x47/0xb0 [libcfs]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa1075ecb&amp;gt;] mgc_process_cfg_log+0x134b/0x15c0 [mgc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa1078133&amp;gt;] mgc_process_log+0x463/0x1390 [mgc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffff8150eb5a&amp;gt;] ? schedule_timeout+0x19a/0x2e0
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa1072a60&amp;gt;] ? mgc_blocking_ast+0x0/0x7e0 [mgc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa0acce00&amp;gt;] ? ldlm_completion_ast+0x0/0x960 [ptlrpc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffff8109705c&amp;gt;] ? remove_wait_queue+0x3c/0x50
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa1079a13&amp;gt;] mgc_requeue_thread+0x1a3/0x750 [mgc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffff81063310&amp;gt;] ? default_wake_function+0x0/0x20
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa1079870&amp;gt;] ? mgc_requeue_thread+0x0/0x750 [mgc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffff8100c0ca&amp;gt;] child_rip+0xa/0x20
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa1079870&amp;gt;] ? mgc_requeue_thread+0x0/0x750 [mgc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffffa1079870&amp;gt;] ? mgc_requeue_thread+0x0/0x750 [mgc]
Apr 15 16:55:57 hyperion-dit31 kernel: [&amp;lt;ffffffff8100c0c0&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;The rest of the stack was lost.&lt;/p&gt;</comment>
                            <comment id="64597" author="utopiabound" created="Tue, 20 Aug 2013 13:43:19 +0000"  >&lt;p&gt;Fixed by &lt;a href=&quot;http://review.whamcloud.com/5049&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5049&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="87144" author="niu" created="Fri, 20 Jun 2014 07:19:16 +0000"  >&lt;p&gt;This bug is hit on several 2.4 sites, could someone backport it to 2.4 ?&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="16192">LU-2059</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="16192">LU-2059</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvlbj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>7216</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>