<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:35:28 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-10479] recovery-mds-scale test failover_mds fails with &apos;test_failover_mds returned 4&apos; </title>
                <link>https://jira.whamcloud.com/browse/LU-10479</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;recovery-mds-scale is failing in test_failover_mds. From the test_log on the client, we can see that the clients exited immediately and no failovers took place:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Client load failed on node trevis-10vm4, rc=1
2018-01-08 21:34:46 Terminating clients loads ...
Duration:               86400
Server failover period: 1200 seconds
Exited after:           0 seconds
Number of failovers before exit:
mds1: 0 times
ost1: 0 times
ost2: 0 times
ost3: 0 times
ost4: 0 times
ost5: 0 times
ost6: 0 times
ost7: 0 times
Status: FAIL: rc=4
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The client load/jobs are terminated because tar is failing due to not enough (any?) free space available. From the run_tar_debug log, strangely, free space is blank:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;++ du -s /etc
++ awk &apos;{print $1}&apos;
+ USAGE=30784
+ /usr/sbin/lctl set_param &apos;llite.*.lazystatfs=0&apos;
+ df /mnt/lustre/d0.tar-trevis-10vm4.trevis.hpdd.intel.com
+ sleep 2
++ df /mnt/lustre/d0.tar-trevis-10vm4.trevis.hpdd.intel.com
++ awk &apos;/:/ { print $4 }&apos;
+ FREE_SPACE=
+ AVAIL=0
+ &apos;[&apos; 0 -lt 30784 &apos;]&apos;
+ echoerr &apos;no enough free disk space: need 30784, avail 0&apos;
+ echo &apos;no enough free disk space: need 30784, avail 0&apos;
no enough free disk space: need 30784, avail 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;There is nothing obviously wrong in the console and dmesg logs. &lt;/p&gt;

&lt;p&gt;So far, this failure is only seen on b2_10, but for several builds. Here are the b2_10 build numbers and links to logs for some of the failures:&lt;br/&gt;
build #17 CentOS 6.9 clients/ CentOS 6.9 ldiskfs servers - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/72399370-881b-11e7-b3ca-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/72399370-881b-11e7-b3ca-5254006e85c2&lt;/a&gt;&lt;br/&gt;
build #19 CentOS 6.9 clients/ CentOS 6.9 ZFS servers - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/c4fabe7e-937c-11e7-b722-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/c4fabe7e-937c-11e7-b722-5254006e85c2&lt;/a&gt;&lt;br/&gt;
build #30 CentOS 6.9 clients/ CentOS 7 ldiskfs servers - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/abaa60d2-a862-11e7-bb19-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/abaa60d2-a862-11e7-bb19-5254006e85c2&lt;/a&gt;&lt;br/&gt;
build #45 CentOS 6.9 clients/ CentOS 7  ldiskfs servers - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/29beec4e-caa1-11e7-9840-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/29beec4e-caa1-11e7-9840-52540065bddc&lt;/a&gt;&lt;br/&gt;
build #52 CentOS 6.9 clients/ CentOS 7  ldiskfs servers - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/76d51ac8-df3f-11e7-8027-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/76d51ac8-df3f-11e7-8027-52540065bddc&lt;/a&gt;&lt;br/&gt;
build #68 CentOS 6.9 clients/ CentOS 7  ldiskfs servers - &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/e5214052-f52d-11e7-8c23-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/e5214052-f52d-11e7-8c23-52540065bddc&lt;/a&gt;&lt;/p&gt;</description>
                <environment></environment>
        <key id="50117">LU-10479</key>
            <summary>recovery-mds-scale test failover_mds fails with &apos;test_failover_mds returned 4&apos; </summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="jamesanunez">James Nunez</reporter>
                        <labels>
                    </labels>
                <created>Tue, 9 Jan 2018 23:55:18 +0000</created>
                <updated>Thu, 14 Mar 2019 15:47:28 +0000</updated>
                                            <version>Lustre 2.10.1</version>
                    <version>Lustre 2.11.0</version>
                    <version>Lustre 2.10.2</version>
                    <version>Lustre 2.10.3</version>
                    <version>Lustre 2.10.4</version>
                    <version>Lustre 2.10.5</version>
                    <version>Lustre 2.10.6</version>
                    <version>Lustre 2.10.7</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>2</watches>
                                                                            <comments>
                            <comment id="223533" author="sarah" created="Tue, 13 Mar 2018 17:08:41 +0000"  >&lt;p&gt;also hit this on 2.11 tag-2.10.59 SLES12SP3 failover testing&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/bc2c657a-26cc-11e8-b74b-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/bc2c657a-26cc-11e8-b74b-52540065bddc&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="228102" author="sarah" created="Thu, 17 May 2018 18:37:15 +0000"  >&lt;p&gt;+1 on b2_10 &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/b757638a-58e8-11e8-b303-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/b757638a-58e8-11e8-b303-52540065bddc&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="231970" author="jamesanunez" created="Wed, 15 Aug 2018 15:10:41 +0000"  >&lt;p&gt;Note: we still see this issue since using RHEL 6.10; &lt;a href=&quot;https://testing.whamcloud.com/test_sets/a6e0f10c-a081-11e8-8ee3-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/a6e0f10c-a081-11e8-8ee3-52540065bddc&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Also, recovery-mds-scale may not be lceaning up after itself very well since recovery-random-scale test fail_client_mds and recovery-double-scale test pairwise_fail fails quickly&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;trevis-12vm4: 19732
trevis-12vm3: 20187
Found the END_RUN_FILE file: /autotest/trevis/2018-08-14/lustre-b2_10-el7-x86_64-vs-lustre-b2_10-el6_10-x86_64--failover--1_24_1__135___df1701ec-c51e-4131-bb8d-62fac9f7b291/shared_dir/end_run_file
trevis-12vm4.trevis.whamcloud.com
Client load  failed on node trevis-12vm4.trevis.whamcloud.com:
/autotest/trevis/2018-08-14/lustre-b2_10-el7-x86_64-vs-lustre-b2_10-el6_10-x86_64--failover--1_24_1__135___df1701ec-c51e-4131-bb8d-62fac9f7b291/recovery-random-scale.test_fail_client_mds.run__stdout.trevis-12vm4.trevis.whamcloud.com.log
/autotest/trevis/2018-08-14/lustre-b2_10-el7-x86_64-vs-lustre-b2_10-el6_10-x86_64--failover--1_24_1__135___df1701ec-c51e-4131-bb8d-62fac9f7b291/recovery-random-scale.test_fail_client_mds.run__debug.trevis-12vm4.trevis.whamcloud.com.log
2018-08-15 00:21:08 Terminating clients loads ...
Duration:               86400
Server failover period: 1200 seconds
Exited after:           0 seconds
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzzqr3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>