<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:26:38 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-9487] mmp test_2: test_2 failed with 22</title>
                <link>https://jira.whamcloud.com/browse/LU-9487</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/adcc6618-5344-4cda-b96f-04bee182c49b&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/adcc6618-5344-4cda-b96f-04bee182c49b&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;New for tag 2.9.57 - not seen in past 6 months (Nov &apos;16 - May &apos;17).&lt;/p&gt;

&lt;p&gt;The same failure message is seen with 7 subsequent subtests.&lt;/p&gt;

&lt;p&gt;From test_log:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: trevis-45vm3 mkdir -p /mnt/lustre-mds1failover; mount -t lustre   		                   /dev/lvm-Role_MDS/P1 /mnt/lustre-mds1failover
trevis-45vm3 (MDS1-1): mount.lustre: according to /etc/mtab /dev/mapper/lvm--Role_MDS-P1 is already mounted on /mnt/lustre-mds1failover
Start of /dev/lvm-Role_MDS/P1 on mds1failover failed 17
trevis-45vm7 (MDS1-2): mount.lustre: mount /dev/mapper/lvm--Role_MDS-P1 at /mnt/lustre-mds1 failed: Invalid argument
trevis-45vm7: This may have multiple causes.
trevis-45vm7: Are the mount options correct?
trevis-45vm7: Check the syslog for more info.
Start of /dev/lvm-Role_MDS/P1 on mds1 failed 22
 mmp test_2: @@@@@@ FAIL: mount failure on failover pair mds1,mds1failover 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4927:error_noexit()
  = /usr/lib64/lustre/tests/mmp.sh:284:mount_after_interval_sub()
  = /usr/lib64/lustre/tests/mmp.sh:296:mount_after_interval()
  = /usr/lib64/lustre/tests/mmp.sh:446:test_2()
  = /usr/lib64/lustre/tests/test-framework.sh:5215:run_one()
  = /usr/lib64/lustre/tests/test-framework.sh:5254:run_one_logged()
  = /usr/lib64/lustre/tests/test-framework.sh:5101:run_test()
  = /usr/lib64/lustre/tests/mmp.sh:449:main()
Dumping lctl log to /test_logs/2017-05-08/lustre-master-el7-x86_64--failover--1_17_1__3575__-70321399293120-051547/mmp.test_2.*.1494260419.log
CMD: trevis-45vm1.trevis.hpdd.intel.com,trevis-45vm5,trevis-45vm6,trevis-45vm7,trevis-45vm8 /usr/sbin/lctl dk &amp;gt; /test_logs/2017-05-08/lustre-master-el7-x86_64--failover--1_17_1__3575__-70321399293120-051547/mmp.test_2.debug_log.\$(hostname -s).1494260419.log;
         dmesg &amp;gt; /test_logs/2017-05-08/lustre-master-el7-x86_64--failover--1_17_1__3575__-70321399293120-051547/mmp.test_2.dmesg.\$(hostname -s).1494260419.log
Resetting fail_loc on all nodes...CMD: trevis-45vm1.trevis.hpdd.intel.com,trevis-45vm5,trevis-45vm6,trevis-45vm7,trevis-45vm8 lctl set_param -n fail_loc=0 	    fail_val=0 2&amp;gt;/dev/null
done.
 mmp test_2: @@@@@@ FAIL: test_2 failed with 22 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:4939:error()
  = /usr/lib64/lustre/tests/test-framework.sh:5215:run_one()
  = /usr/lib64/lustre/tests/test-framework.sh:5254:run_one_logged()
  = /usr/lib64/lustre/tests/test-framework.sh:5101:run_test()
  = /usr/lib64/lustre/tests/mmp.sh:449:main()
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>trevis-45, failover&lt;br/&gt;
&amp;nbsp;&amp;nbsp;EL7, master branch, v2.9.57, b3575, ldiskfs&lt;br/&gt;
</environment>
        <key id="46011">LU-9487</key>
            <summary>mmp test_2: test_2 failed with 22</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="jcasper">James Casper</reporter>
                        <labels>
                    </labels>
                <created>Wed, 10 May 2017 15:37:04 +0000</created>
                <updated>Sat, 3 Jun 2017 04:30:50 +0000</updated>
                            <resolved>Sat, 3 Jun 2017 04:30:50 +0000</resolved>
                                    <version>Lustre 2.10.0</version>
                                    <fixVersion>Lustre 2.10.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="195276" author="casperjx" created="Wed, 10 May 2017 15:51:41 +0000"  >&lt;p&gt;The test set that follows mmp (replay-single) also fails with the same failure message.  replay-single is the last test set in the failover group.&lt;/p&gt;</comment>
                            <comment id="195536" author="adilger" created="Thu, 11 May 2017 17:49:14 +0000"  >&lt;p&gt;It looks like the MDS didn&apos;t clean up properly at the end of the previous test?  This needs to be investigated more closely.&lt;/p&gt;</comment>
                            <comment id="195555" author="casperjx" created="Thu, 11 May 2017 19:32:47 +0000"  >&lt;p&gt;A clarification on this statement: &quot;New for tag 2.9.57 - not seen in past 6 months (Nov &apos;16 - May &apos;17).&quot;&lt;/p&gt;

&lt;p&gt;This has not been seen in tag testing since November 10, 2016 (master branch, tags 51-56).  There are 55 occurrences outside of tag testing (all in  failover group testing).&lt;/p&gt;</comment>
                            <comment id="196022" author="pjones" created="Tue, 16 May 2017 15:34:47 +0000"  >&lt;p&gt;Alex is looking into this one&lt;/p&gt;</comment>
                            <comment id="196215" author="bzzz" created="Wed, 17 May 2017 17:42:54 +0000"  >&lt;p&gt;the log looks a bit strange..&lt;br/&gt;
[  307.590154] Lustre: DEBUG MARKER: ----&lt;del&gt;============= acceptance-small: mmp ============&lt;/del&gt;---- Mon May 8 16:18:34 UTC 2017&lt;br/&gt;
[  307.956938] Lustre: DEBUG MARKER: /usr/sbin/lctl mark == mmp test complete, duration -o sec ================================================================ 16:18:34 &amp;#40;1494260314&amp;#41;&lt;br/&gt;
[  308.111498] Lustre: DEBUG MARKER: == mmp test complete, duration -o sec ================================================================ 16:18:34 (1494260314)&lt;br/&gt;
[  336.896363] Lustre: DEBUG MARKER: /usr/sbin/lctl mark == mmp test 1: two mounts at the same time =========================================================== 16:19:03 &amp;#40;1494260343&amp;#41;&lt;/p&gt;</comment>
                            <comment id="196350" author="bzzz" created="Thu, 18 May 2017 14:56:04 +0000"  >&lt;p&gt;== mmp test 1: two mounts at the same time =========================================================== 16:19:03 (1494260343)&lt;br/&gt;
Starting mds1failover:   /dev/lvm-Role_MDS/P1 /mnt/lustre-mds1failover&lt;br/&gt;
Starting mds1:   /dev/lvm-Role_MDS/P1 /mnt/lustre-mds1&lt;br/&gt;
Started lustre-MDT0000&lt;br/&gt;
Start of /dev/lvm-Role_MDS/P1 on mds1 failed 22&lt;br/&gt;
Starting ost1:   /dev/lvm-Role_OSS/P1 /mnt/lustre-ost1&lt;br/&gt;
Starting ost1failover:   /dev/lvm-Role_OSS/P1 /mnt/lustre-ost1failover&lt;br/&gt;
Started lustre-OST0000&lt;br/&gt;
Start of /dev/lvm-Role_OSS/P1 on ost1failover failed 22&lt;br/&gt;
Stopping /mnt/lustre-ost1 (opts&lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt; on trevis-45vm8&lt;br/&gt;
PASS 1 (52s)&lt;/p&gt;

&lt;p&gt;for whatever reason MDS wasn&apos;t stopped by test 1, though I see in the scripts this:&lt;br/&gt;
    stop $mds_facet $opts || return ${PIPESTATUS&lt;span class=&quot;error&quot;&gt;&amp;#91;0&amp;#93;&lt;/span&gt;}&lt;br/&gt;
    stop $oss_facet $opts || return ${PIPESTATUS&lt;span class=&quot;error&quot;&gt;&amp;#91;0&amp;#93;&lt;/span&gt;}&lt;/p&gt;

&lt;p&gt;going to add a debugging&lt;/p&gt;</comment>
                            <comment id="196353" author="gerrit" created="Thu, 18 May 2017 15:02:50 +0000"  >&lt;p&gt;Alex Zhuravlev (alexey.zhuravlev@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/27192&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/27192&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9487&quot; title=&quot;mmp test_2: test_2 failed with 22&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9487&quot;&gt;&lt;del&gt;LU-9487&lt;/del&gt;&lt;/a&gt;: debugging patch&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 7881652b82308ef5193efa2cb85292ea540e1723&lt;/p&gt;</comment>
                            <comment id="196727" author="bzzz" created="Tue, 23 May 2017 11:21:27 +0000"  >&lt;p&gt;I think the root cause is that test #2 expects the primary MDS to be mounted:&lt;br/&gt;
    stop_services primary || return ${PIPESTATUS&lt;span class=&quot;error&quot;&gt;&amp;#91;0&amp;#93;&lt;/span&gt;}&lt;br/&gt;
but actually the mount process in mount_after_interval_sub() seem to be racy by desing:&lt;br/&gt;
    start $facet $device $opts &amp;amp;&lt;br/&gt;
    ...&lt;br/&gt;
    start $failover_facet $device $opts&lt;/p&gt;

&lt;p&gt;the test was developed by Jian Yu, it&apos;d helpful to get any hints on this.&lt;/p&gt;</comment>
                            <comment id="196782" author="gerrit" created="Tue, 23 May 2017 18:49:18 +0000"  >&lt;p&gt;Jian Yu (jian.yu@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/27253&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/27253&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9487&quot; title=&quot;mmp test_2: test_2 failed with 22&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9487&quot;&gt;&lt;del&gt;LU-9487&lt;/del&gt;&lt;/a&gt; tests: stop primary and failover services in mmp test 1&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 0f5520c5413e32caa55275c4dc7310e1a03a60a5&lt;/p&gt;</comment>
                            <comment id="196783" author="yujian" created="Tue, 23 May 2017 18:53:23 +0000"  >&lt;p&gt;Yes, Alex. In mmp.sh test_1(), it&apos;s likely the failover service is started before the primary one, I just pushed a patch to check and stop the failover service. &lt;/p&gt;</comment>
                            <comment id="198010" author="gerrit" created="Sat, 3 Jun 2017 03:57:24 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/27253/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/27253/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-9487&quot; title=&quot;mmp test_2: test_2 failed with 22&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-9487&quot;&gt;&lt;del&gt;LU-9487&lt;/del&gt;&lt;/a&gt; tests: stop primary and failover services in mmp test 1&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: a23bb09a3bc447e4286218416d1dd5d096724bfc&lt;/p&gt;</comment>
                            <comment id="198041" author="pjones" created="Sat, 3 Jun 2017 04:30:50 +0000"  >&lt;p&gt;Landed for 2.10&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzzcdz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>