<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:30:58 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-3101] Interop 1.8.9&lt;-&gt;2.4 failure on test suite replay-single test_61d: cannot restart mgs</title>
                <link>https://jira.whamcloud.com/browse/LU-3101</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/a0617196-9725-11e2-9ec7-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/a0617196-9725-11e2-9ec7-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_61d failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;cannot restart mgs&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;MDS console shows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00:09:37:Lustre: DEBUG MARKER: == replay-single test 61d: error in llog_setup should cleanup the llog context correctly == 00:09:35 (1364368175)
00:09:37:Lustre: DEBUG MARKER: grep -c /mnt/mds&apos; &apos; /proc/mounts
00:09:37:Lustre: DEBUG MARKER: umount -d /mnt/mds
00:09:49:Lustre: DEBUG MARKER: lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
00:09:50:Lustre: DEBUG MARKER: lctl set_param fail_loc=0x80000605
00:09:50:Lustre: DEBUG MARKER: mkdir -p /mnt/mds
00:09:50:Lustre: DEBUG MARKER: mkdir -p /mnt/mds; mount -t lustre -o loop  /dev/lvm-MDS/P1 /mnt/mds
00:09:50:LDISKFS-fs (loop0): mounted filesystem with ordered data mode. quota=on. Opts: 
00:09:50:Lustre: *** cfs_fail_loc=605, val=0***
00:09:50:LustreError: 5059:0:(llog_obd.c:207:llog_setup()) MGS: ctxt 0 lop_setup=ffffffffa0631ce0 failed: rc = -95
00:09:50:LustreError: 5059:0:(obd_config.c:572:class_setup()) setup MGS failed (-95)
00:09:50:LustreError: 5059:0:(obd_mount.c:378:lustre_start_simple()) MGS setup error -95
00:09:50:LustreError: 15e-a: Failed to start MGS &apos;MGS&apos; (-95). Is the &apos;mgs&apos; module loaded?
00:09:50:LustreError: 5059:0:(obd_mount.c:1379:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0000: Can&apos;t end config log lustre-client.
00:09:50:LustreError: 5059:0:(obd_mount.c:2115:server_put_super()) lustre-MDT0000: failed to disconnect lwp. (rc=-2)
00:09:50:LustreError: 5059:0:(obd_mount.c:2145:server_put_super()) no obd lustre-MDT0000
00:09:51:LustreError: 5059:0:(obd_mount.c:139:server_deregister_mount()) lustre-MDT0000 not registered
00:09:51:LustreError: 5059:0:(obd_mount.c:2989:lustre_fill_super()) Unable to mount /dev/loop0 (-95)
00:09:51:Lustre: DEBUG MARKER: lctl set_param fail_loc=0
00:09:51:Lustre: DEBUG MARKER: mkdir -p /mnt/mds
00:09:51:Lustre: DEBUG MARKER: mkdir -p /mnt/mds; mount -t lustre -o loop  /dev/lvm-MDS/P1 /mnt/mds
00:09:51:LustreError: 15d-9: The MGS service was already started from server
00:09:51:LustreError: 5228:0:(obd_mount.c:1379:lustre_disconnect_lwp()) lustre-MDT0000-lwp-MDT0000: Can&apos;t end config log lustre-client.
00:09:51:LustreError: 5228:0:(obd_mount.c:2115:server_put_super()) lustre-MDT0000: failed to disconnect lwp. (rc=-2)
00:09:51:LustreError: 5228:0:(obd_mount.c:2145:server_put_super()) no obd lustre-MDT0000
00:09:51:LustreError: 5228:0:(obd_mount.c:139:server_deregister_mount()) lustre-MDT0000 not registered
00:09:51:LustreError: 5228:0:(obd_mount.c:2989:lustre_fill_super()) Unable to mount  (-114)
00:09:51:Lustre: DEBUG MARKER: /usr/sbin/lctl mark  replay-single test_61d: @@@@@@ FAIL: cannot restart mgs 
00:09:51:Lustre: DEBUG MARKER: replay-single test_61d: @@@@@@ FAIL: cannot restart mgs
00:09:51:Lustre: DEBUG MARKER: /usr/sbin/lctl dk &amp;gt; /logdir/test_logs/2013-03-26/lustre-master-el6-x86_64-vs-lustre-b1_8-el6-x86_64--full--2_4_1__1346__-70011898121780-141237/replay-single.test_61d.debug_log.$(hostname -s).1364368184.log;
00:09:51:         dmesg &amp;gt; /logdir/test_logs/2013-03-26/lu
00:09:51:Lustre: DEBUG MARKER: lctl set_param -n fail_loc=0 2&amp;gt;/dev/null || true
00:09:51:Lustre: DEBUG MARKER: rc=$([ -f /proc/sys/lnet/catastrophe ] &amp;amp;&amp;amp; echo $(&amp;lt; /proc/sys/lnet/catastrophe) || echo 0);
00:09:51:if [ $rc -ne 0 ]; then echo $(hostname): $rc; fi
00:09:51:exit $rc;
00:09:51:Lustre: DEBUG MARKER: /usr/sbin/lctl mark == replay-single test 62: don\&apos;t mis-drop resent replay == 00:09:46 \(1364368186\)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>client: 1.8.9&lt;br/&gt;
server: lustre-master build #1346</environment>
        <key id="18232">LU-3101</key>
            <summary>Interop 1.8.9&lt;-&gt;2.4 failure on test suite replay-single test_61d: cannot restart mgs</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="hongchao.zhang">Hongchao Zhang</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Wed, 3 Apr 2013 20:26:54 +0000</created>
                <updated>Mon, 19 Aug 2013 14:37:24 +0000</updated>
                            <resolved>Mon, 19 Aug 2013 14:37:24 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                    <fixVersion>Lustre 2.4.1</fixVersion>
                    <fixVersion>Lustre 2.5.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="55523" author="pjones" created="Thu, 4 Apr 2013 17:32:17 +0000"  >&lt;p&gt;Hongchao&lt;/p&gt;

&lt;p&gt;Could you please investigate?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="56186" author="hongchao.zhang" created="Fri, 12 Apr 2013 04:04:37 +0000"  >&lt;p&gt;the issue is reproduced on master locally, and it&apos;s caused by the wrong cleanup after MGS failed to start up.&lt;br/&gt;
the patch is tracked at &lt;a href=&quot;http://review.whamcloud.com/#change,6035&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,6035&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="64241" author="yujian" created="Wed, 14 Aug 2013 13:44:05 +0000"  >&lt;p&gt;Lustre client build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b1_8/258/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b1_8/258/&lt;/a&gt; (1.8.9-wc1)&lt;br/&gt;
Lustre server build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_4/31/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_4/31/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;replay-single test 61d hit the same failure:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/cf9987d6-0486-11e3-90ba-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/cf9987d6-0486-11e3-90ba-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hi Oleg,&lt;br/&gt;
Could you please cherry-pick the patch to Lustre b2_4 branch? Thanks.&lt;/p&gt;</comment>
                            <comment id="64454" author="hongchao.zhang" created="Mon, 19 Aug 2013 09:27:21 +0000"  >&lt;p&gt;the patch is landed on master&lt;/p&gt;</comment>
                            <comment id="64467" author="yujian" created="Mon, 19 Aug 2013 14:37:09 +0000"  >&lt;p&gt;The patch was also cherry-picked to Lustre b2_4 branch.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvn1b:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>7538</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>