<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:20:56 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-1931] Test failure on test suite conf-sanity, subtest test_5d</title>
                <link>https://jira.whamcloud.com/browse/LU-1931</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Li Wei &amp;lt;liwei@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/0be5975e-f9b4-11e1-b8d8-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/0be5975e-f9b4-11e1-b8d8-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_5d failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;test failed to respond and timed out&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;From the test output:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;== conf-sanity test 5d: mount with ost down ========================================================== 00:57:50 (1347004670)
Loading modules from /usr/lib64/lustre
detected 1 online CPUs by getconf
libcfs will create CPU partition based on online CPUs
../libcfs/libcfs/libcfs options: &apos;cpu_npartitions=0&apos;
debug=-1
subsystem_debug=0xffb7e3ff
../lnet/lnet/lnet options: &apos;accept=all&apos;
gss/krb5 is not supported
loading modules on client-19vm2,client-19vm3,client-19vm4
CMD: client-19vm2,client-19vm3,client-19vm4 PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/usr/lib64/lustre/tests//usr/lib64/lustre/tests:/usr/lib64/lustre/tests:/usr/lib64/lustre/tests/../utils:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/openmpi/1.4-gcc/bin:/usr/kerberos/sbin:/usr/kerberos/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin: NAME=autotest_config sh rpc.sh load_modules_local 
client-19vm3: Loading modules from /usr/lib64/lustre
client-19vm3: detected 1 online CPUs by sysfs
client-19vm3: libcfs will create CPU partition based on online CPUs
client-19vm3: ../libcfs/libcfs/libcfs options: &apos;cpu_npartitions=0&apos;
client-19vm2: Loading modules from /usr/lib64/lustre
client-19vm2: detected 1 online CPUs by getconf
client-19vm2: libcfs will create CPU partition based on online CPUs
client-19vm2: ../libcfs/libcfs/libcfs options: &apos;cpu_npartitions=0&apos;
client-19vm4: Loading modules from /usr/lib64/lustre
client-19vm4: detected 1 online CPUs by sysfs
client-19vm4: libcfs will create CPU partition based on online CPUs
client-19vm4: ../libcfs/libcfs/libcfs options: &apos;cpu_npartitions=0&apos;
client-19vm2: debug=0x33f0404
client-19vm3: debug=0x33f0404
client-19vm2: subsystem_debug=0xffb7e3ff
client-19vm3: subsystem_debug=0xffb7e3ff
client-19vm2: ../lnet/lnet/lnet options: &apos;accept=all&apos;
client-19vm4: debug=0x33f0404
client-19vm4: subsystem_debug=0xffb7e3ff
client-19vm2: gss/krb5 is not supported
client-19vm3: gss/krb5 is not supported
client-19vm4: gss/krb5 is not supported
start ost1 service on client-19vm4
CMD: client-19vm4 mkdir -p /mnt/ost1
Starting ost1:   lustre-ost1/ost1 /mnt/ost1
CMD: client-19vm4 mkdir -p /mnt/ost1; mount -t lustre   		                   lustre-ost1/ost1 /mnt/ost1
client-19vm4: mount.lustre: mount lustre-ost1/ost1 at /mnt/ost1 failed: Input/output error
client-19vm4: Is the MGS running?
Start of lustre-ost1/ost1 on ost1 failed 5
start mds service on client-19vm3
CMD: client-19vm3 mkdir -p /mnt/mds1
CMD: client-19vm3 test -b /dev/lvm-MDS/P1
Starting mds1: -o user_xattr,acl  /dev/lvm-MDS/P1 /mnt/mds1
CMD: client-19vm3 mkdir -p /mnt/mds1; mount -t lustre -o user_xattr,acl  		                   /dev/lvm-MDS/P1 /mnt/mds1
CMD: client-19vm3 PATH=/usr/lib64/lustre/tests:/usr/lib/lustre/tests:/usr/lib64/lustre/tests:/opt/iozone/bin:/usr/lib64/lustre/tests//usr/lib64/lustre/tests:/usr/lib64/lustre/tests:/usr/lib64/lustre/tests/../utils:/opt/iozone/bin:/usr/lib64/lustre/tests/mpi:/usr/lib64/lustre/tests/racer:/usr/lib64/lustre/../lustre-iokit/sgpdd-survey:/usr/lib64/lustre/tests:/usr/lib64/lustre/utils/gss:/usr/lib64/lustre/utils:/usr/lib64/openmpi/1.4-gcc/bin:/usr/kerberos/sbin:/usr/kerberos/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin: NAME=autotest_config sh rpc.sh set_default_debug \&quot;-1\&quot; \&quot; 0xffb7e3ff\&quot; 32 
CMD: client-19vm3 e2label /dev/lvm-MDS/P1 2&amp;gt;/dev/null
Started lustre-MDT0000
stop ost1 service on client-19vm4
CMD: client-19vm4 grep -c /mnt/ost1&apos; &apos; /proc/mounts
CMD: client-19vm4 lsmod | grep lnet &amp;gt; /dev/null &amp;amp;&amp;amp; lctl dl | grep &apos; ST &apos;
mount lustre on /mnt/lustre.....
Starting client: client-19vm1.lab.whamcloud.com: -o user_xattr,acl,flock client-19vm3@tcp:/lustre /mnt/lustre
CMD: client-19vm1.lab.whamcloud.com mkdir -p /mnt/lustre
CMD: client-19vm1.lab.whamcloud.com mount -t lustre -o user_xattr,acl,flock client-19vm3@tcp:/lustre /mnt/lustre
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Info required for matching: conf-sanity 5d&lt;/p&gt;</description>
                <environment></environment>
        <key id="15949">LU-1931</key>
            <summary>Test failure on test suite conf-sanity, subtest test_5d</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                            <label>zfs</label>
                    </labels>
                <created>Thu, 13 Sep 2012 09:06:00 +0000</created>
                <updated>Tue, 2 Apr 2013 21:21:04 +0000</updated>
                            <resolved>Tue, 2 Apr 2013 21:21:04 +0000</resolved>
                                    <version>Lustre 2.3.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="44800" author="keith" created="Thu, 13 Sep 2012 12:40:48 +0000"  >&lt;p&gt;client-19vm2 (Client 2) rebooted at the start of this test for some unknown reason. &lt;/p&gt;

&lt;p&gt;We need dmesg /var/log/messages logs from the system.  &lt;/p&gt;</comment>
                            <comment id="45771" author="mdiep" created="Sat, 29 Sep 2012 11:45:11 +0000"  >&lt;p&gt;Keith, I can reproduce this in my vm (client-12vm&lt;span class=&quot;error&quot;&gt;&amp;#91;1-4&amp;#93;&lt;/span&gt;). it&apos;s hung at the same state right now. Please take a look or let me know when you want to look and I will reproduce it again. The way to do that is set USE_OFD=yes and OSTFSTYPE=zfs, LOAD_REMOTE_MODULE=true&lt;/p&gt;</comment>
                            <comment id="45801" author="mdiep" created="Mon, 1 Oct 2012 11:10:43 +0000"  >&lt;p&gt;this seems to be ORI-740, patch &lt;a href=&quot;http://review.whamcloud.com/#change,3678&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,3678&lt;/a&gt;. it needs to land on 2.3&lt;/p&gt;</comment>
                            <comment id="45814" author="mdiep" created="Mon, 1 Oct 2012 15:30:20 +0000"  >&lt;p&gt;== conf-sanity test 19b: start/stop OSTs without MDS == 09:45:26 (1349109926)&lt;br/&gt;
start ost1 service on client-12vm4&lt;br/&gt;
CMD: client-12vm4 mkdir -p /mnt/ost1&lt;br/&gt;
Starting ost1:   lustre-ost1/ost1 /mnt/ost1&lt;br/&gt;
CMD: client-12vm4 mkdir -p /mnt/ost1; mount -t lustre                                      lustre-ost1/ost1 /mnt/ost1&lt;br/&gt;
client-12vm4: mount.lustre: mount lustre-ost1/ost1 at /mnt/ost1 failed: Input/output error&lt;br/&gt;
client-12vm4: Is the MGS running?&lt;br/&gt;
Start of lustre-ost1/ost1 on ost1 failed 5&lt;br/&gt;
 conf-sanity test_19b: @@@@@@ FAIL: test_19b failed with 1 &lt;br/&gt;
  Trace dump:&lt;br/&gt;
  = /usr/lib64/lustre/tests/test-framework.sh:3640:error_noexit()&lt;br/&gt;
  = /usr/lib64/lustre/tests/test-framework.sh:3662:error()&lt;br/&gt;
  = /usr/lib64/lustre/tests/test-framework.sh:3898:run_one()&lt;br/&gt;
  = /usr/lib64/lustre/tests/test-framework.sh:3928:run_one_logged()&lt;br/&gt;
  = /usr/lib64/lustre/tests/test-framework.sh:3799:run_test()&lt;br/&gt;
  = /usr/lib64/lustre/tests/conf-sanity.sh:655:main()&lt;br/&gt;
Dumping lctl log to /tmp/test_logs/2012-10-01/092312/conf-sanity.test_19b.*.1349109952.log&lt;br/&gt;
CMD: client-12vm1.lab.whamcloud.com,client-12vm2,client-12vm3,client-12vm4 /usr/sbin/lctl dk &amp;gt; /tmp/test_logs/2012-10-01/092312/conf-sanity.test_19b.debug_log.\$(hostname -s).1349109952.log;&lt;br/&gt;
         dmesg &amp;gt; /tmp/test_logs/2012-10-01/092312/conf-sanity.test_19b.dmesg.\$(hostname -s).1349109952.log&lt;br/&gt;
CMD: client-12vm1.lab.whamcloud.com,client-12vm2,client-12vm3,client-12vm4 rsync -az /tmp/test_logs/2012-10-01/092312/conf-sanity.test_19b.*.1349109952.log client-12vm1.lab.whamcloud.com:/tmp/test_logs/2012-10-01/092312&lt;br/&gt;
client-12vm1: Host key verification failed.&lt;br/&gt;
client-12vm1: rsync: connection unexpectedly closed (0 bytes received so far) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender&amp;#93;&lt;/span&gt;&lt;br/&gt;
client-12vm1: rsync error: unexplained error (code 255) at io.c(600) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender=3.0.6&amp;#93;&lt;/span&gt;&lt;br/&gt;
client-12vm3: Host key verification failed.&lt;br/&gt;
client-12vm3: rsync: connection unexpectedly closed (0 bytes received so far) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender&amp;#93;&lt;/span&gt;&lt;br/&gt;
client-12vm3: rsync error: error in rsync protocol data stream (code 12) at io.c(600) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender=3.0.6&amp;#93;&lt;/span&gt;&lt;br/&gt;
client-12vm2: Host key verification failed.&lt;br/&gt;
client-12vm2: rsync: connection unexpectedly closed (0 bytes received so far) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender&amp;#93;&lt;/span&gt;&lt;br/&gt;
client-12vm2: rsync error: error in rsync protocol data stream (code 12) at io.c(600) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender=3.0.6&amp;#93;&lt;/span&gt;&lt;br/&gt;
client-12vm4: Host key verification failed.&lt;br/&gt;
client-12vm4: rsync: connection unexpectedly closed (0 bytes received so far) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender&amp;#93;&lt;/span&gt;&lt;br/&gt;
client-12vm4: rsync error: error in rsync protocol data stream (code 12) at io.c(600) &lt;span class=&quot;error&quot;&gt;&amp;#91;sender=3.0.6&amp;#93;&lt;/span&gt;&lt;br/&gt;
test_19b returned 1&lt;br/&gt;
FAIL 19b (28s)&lt;/p&gt;</comment>
                            <comment id="45815" author="bzzz" created="Mon, 1 Oct 2012 15:34:21 +0000"  >&lt;p&gt;this is a dup of &lt;a href=&quot;http://jira.whamcloud.com/browse/LU-2059&quot; class=&quot;external-link&quot; rel=&quot;nofollow&quot;&gt;http://jira.whamcloud.com/browse/LU-2059&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="45816" author="bzzz" created="Mon, 1 Oct 2012 15:35:23 +0000"  >&lt;p&gt;basically, all the tests where OST starts w/o MGS running should be disabled with ZFS untill &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2059&quot; title=&quot;mgc to backup configuration on osd-based llogs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2059&quot;&gt;&lt;del&gt;LU-2059&lt;/del&gt;&lt;/a&gt; is done.&lt;/p&gt;
</comment>
                            <comment id="55336" author="utopiabound" created="Tue, 2 Apr 2013 21:21:04 +0000"  >&lt;p&gt;Per Alex&apos;s comment, closing this as duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2059&quot; title=&quot;mgc to backup configuration on osd-based llogs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2059&quot;&gt;&lt;del&gt;LU-2059&lt;/del&gt;&lt;/a&gt;.&lt;br/&gt;
This test was excepted by patch &lt;a href=&quot;http://review.whamcloud.com/4234&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4234&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="16192">LU-2059</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv3yv:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4171</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>