<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:25:00 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2415] recovery-mds-scale test_failover_mds: lustre:MDT0000/recovery_status found no match</title>
                <link>https://jira.whamcloud.com/browse/LU-2415</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/8106fea4-3a9d-11e2-b2e6-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/8106fea4-3a9d-11e2-b2e6-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_failover_mds failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;test_failover_mds returned 7&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;test log shows&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== recovery-mds-scale test failover_mds: failover MDS == 13:08:23 (1354136903)
Started client load: dd on client-28vm5
CMD: client-28vm5 PATH=/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/openmpi/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin: MOUNT=/mnt/lustre ERRORS_OK= BREAK_ON_ERROR= END_RUN_FILE=/home/autotest/.autotest/shared_dir/2012-11-28/112808-70113261691540/end_run_file LOAD_PID_FILE=/tmp/client-load.pid TESTLOG_PREFIX=/logdir/test_logs/2012-11-28/lustre-master-el6-x86_64-fo__1065__-70113261691540-112807/recovery-mds-scale TESTNAME=test_failover_mds DBENCH_LIB=/usr/share/doc/dbench/loadfiles DBENCH_SRC= run_dd.sh
Started client load: tar on client-28vm6
CMD: client-28vm6 PATH=/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/openmpi/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin: MOUNT=/mnt/lustre ERRORS_OK= BREAK_ON_ERROR= END_RUN_FILE=/home/autotest/.autotest/shared_dir/2012-11-28/112808-70113261691540/end_run_file LOAD_PID_FILE=/tmp/client-load.pid TESTLOG_PREFIX=/logdir/test_logs/2012-11-28/lustre-master-el6-x86_64-fo__1065__-70113261691540-112807/recovery-mds-scale TESTNAME=test_failover_mds DBENCH_LIB=/usr/share/doc/dbench/loadfiles DBENCH_SRC= run_tar.sh
client loads pids:
CMD: client-28vm5,client-28vm6 cat /tmp/client-load.pid
client-28vm6: 4127
client-28vm5: 4080
==== Checking the clients loads BEFORE failover -- failure NOT OK              ELAPSED=0 DURATION=86400 PERIOD=900
CMD: client-28vm5 rc=\$([ -f /proc/sys/lnet/catastrophe ] &amp;amp;&amp;amp; echo \$(&amp;lt; /proc/sys/lnet/catastrophe) || echo 0);
		if [ \$rc -ne 0 ]; then echo \$(hostname): \$rc; fi
		exit \$rc;
CMD: client-28vm5 ps auxwww | grep -v grep | grep -q run_dd.sh
CMD: client-28vm6 rc=\$([ -f /proc/sys/lnet/catastrophe ] &amp;amp;&amp;amp; echo \$(&amp;lt; /proc/sys/lnet/catastrophe) || echo 0);
		if [ \$rc -ne 0 ]; then echo \$(hostname): \$rc; fi
		exit \$rc;
CMD: client-28vm6 ps auxwww | grep -v grep | grep -q run_tar.sh
Wait mds1 recovery complete before doing next failover...
CMD: client-28vm1.lab.whamcloud.com lctl get_param -n at_max
affected facets: mds1
CMD: client-28vm3 PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/openmpi/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin: NAME=autotest_config sh rpc.sh _wait_recovery_complete *.lustre:MDT0000.recovery_status 662 
client-28vm3: error: get_param: /proc/{fs,sys}/{lnet,lustre}/*/lustre:MDT0000/recovery_status: Found no match
mds1 recovery is not completed!
2012-11-28 13:08:32 Terminating clients loads ...
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="16828">LU-2415</key>
            <summary>recovery-mds-scale test_failover_mds: lustre:MDT0000/recovery_status found no match</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="6" iconUrl="https://jira.whamcloud.com/images/icons/statuses/closed.png" description="The issue is considered finished, the resolution is correct. Issues which are closed can be reopened.">Closed</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="bzzz">Alex Zhuravlev</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Fri, 30 Nov 2012 13:48:16 +0000</created>
                <updated>Fri, 8 Jul 2016 18:23:54 +0000</updated>
                            <resolved>Fri, 19 Apr 2013 22:31:27 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="48616" author="adilger" created="Fri, 30 Nov 2012 14:50:11 +0000"  >&lt;p&gt;It looks like the root of the problem is that it is looking for &quot;&lt;tt&gt;lctl get_param *.lustre:MDT0000.recovery_status&lt;/tt&gt;&quot; and &quot;&lt;tt&gt;lctl get_param *.lustre:OST0000.recovery_status&lt;/tt&gt;&quot; (note &apos;:&apos; instead of &apos;-&apos; in the device name).&lt;/p&gt;

&lt;p&gt;Somewhere the test-framework.sh is either finding or caching the wrong device name, or the device name was not updated on disk, or it actually has the wrong name in /proc.&lt;/p&gt;</comment>
                            <comment id="48719" author="bzzz" created="Tue, 4 Dec 2012 00:17:22 +0000"  >&lt;p&gt;just tried with llmount.sh:&lt;/p&gt;

&lt;p&gt;Setup mgs, mdt, osts&lt;br/&gt;
Starting mds1:   -o loop /tmp/lustre-mdt1 /mnt/mds1&lt;br/&gt;
Started lustre-MDT0000&lt;/p&gt;

&lt;p&gt;...&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;debugfs  -R stats /tmp/lustre-mdt1 |grep &apos;volume name&apos;&lt;br/&gt;
debugfs 1.42.3.wc3 (15-Aug-2012)&lt;br/&gt;
Filesystem volume name:   lustre-MDT0000&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;...&lt;/p&gt;

&lt;ol&gt;
	&lt;li&gt;ls /proc/fs/lustre/mdt/&lt;br/&gt;
lustre-MDT0000  num_refs&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;going through test-framework.sh now..&lt;/p&gt;</comment>
                            <comment id="55991" author="yujian" created="Wed, 10 Apr 2013 11:07:34 +0000"  >&lt;p&gt;The issue was fixed by patch &lt;a href=&quot;http://review.whamcloud.com/#change,5867&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,5867&lt;/a&gt; in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2008&quot; title=&quot;After hardware reboot (using pm) the node cannot be accessed&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2008&quot;&gt;&lt;del&gt;LU-2008&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="16082">LU-2008</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="31448">LU-6992</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvd4n:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5730</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>