<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:21:43 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2023] Test failure on test suite parallel-scale-nfsv3</title>
                <link>https://jira.whamcloud.com/browse/LU-2023</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/b2d86c02-0658-11e2-9b17-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/b2d86c02-0658-11e2-9b17-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;From the test report, all sub tests passed while at the end of suite log shows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;client-26vm6: mount.lustre: mount client-26vm3@tcp:/lustre at /mnt/lustre failed: File exists
 parallel-scale-nfsv3 : @@@@@@ FAIL: failed to mount lustre after nfs test 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:3645:error_noexit()
  = /usr/lib64/lustre/tests/parallel-scale-nfs.sh:106:main()
== parallel-scale-nfsv3 parallel-scale-nfs.sh test complete, duration 4952 sec == 00:49:32 (1348472972)
/usr/lib64/lustre/tests/parallel-scale-nfs.sh: FAIL:  failed to mount lustre after nfs test
NFSCLIENT mode: setup, cleanup, check config skipped
CMD: client-26vm5,client-26vm6.lab.whamcloud.com echo \$(hostname); grep &apos; &apos;/mnt/lustre&apos; &apos; /proc/mounts
client-26vm5.lab.whamcloud.com
10.10.4.150@tcp:/lustre /mnt/lustre lustre rw,flock,user_xattr 0 0
client-26vm6.lab.whamcloud.com
 parallel-scale-nfsv3 : @@@@@@ FAIL: NFSCLIENT=true mode, but no NFS export found! 
  Trace dump:
  = /usr/lib64/lustre/tests/test-framework.sh:3645:error_noexit()
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>Server/Client lustre-b2_3-RC1 RHEL6</environment>
        <key id="16108">LU-2023</key>
            <summary>Test failure on test suite parallel-scale-nfsv3</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="mdiep">Minh Diep</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Mon, 24 Sep 2012 17:20:26 +0000</created>
                <updated>Fri, 2 Dec 2016 17:02:38 +0000</updated>
                            <resolved>Fri, 2 Dec 2016 17:02:38 +0000</resolved>
                                    <version>Lustre 2.3.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="45488" author="pjones" created="Tue, 25 Sep 2012 00:44:52 +0000"  >&lt;p&gt;Minh&lt;/p&gt;

&lt;p&gt;Could you please advise on this one?&lt;/p&gt;

&lt;p&gt;thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="45491" author="mdiep" created="Tue, 25 Sep 2012 02:06:05 +0000"  >&lt;p&gt;found this on the mds/nfs server (client-26vm3) console.&lt;/p&gt;

&lt;p&gt;00:39:05:Lustre: DEBUG MARKER: == parallel-scale-nfsv3 test iorfpp: iorfpp == 00:38:56 (1348472336)&lt;br/&gt;
00:39:05:Lustre: DEBUG MARKER: lfs setstripe /mnt/lustre/d0.ior.fpp -c -1&lt;br/&gt;
00:42:57:rpc-srv/tcp: nfsd: got error -32 when sending 140 bytes - shutting down socket   &amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&lt;br/&gt;
00:49:29:Lustre: DEBUG MARKER: lctl set_param -n fail_loc=0 2&amp;gt;/dev/null || true&lt;br/&gt;
00:49:29:Lustre: DEBUG MARKER: rc=$([ -f /proc/sys/lnet/catastrophe ] &amp;amp;&amp;amp; echo $(&amp;lt; /proc/sys/lnet/catastrophe) || echo 0);&lt;br/&gt;
00:49:29:if [ $rc -ne 0 ]; then echo $(hostname): $rc; fi&lt;br/&gt;
00:49:29:exit $rc;&lt;br/&gt;
00:49:30:Lustre: DEBUG MARKER: service nfs stop&lt;/p&gt;

&lt;p&gt;I think this is nfs issue under stress test. This results in nfs shutdown was not completed, hence when lustre mount would fail to do (-17) file exist.&lt;/p&gt;

&lt;p&gt;I would like to get a second opinion on this.&lt;/p&gt;


</comment>
                            <comment id="45492" author="mdiep" created="Tue, 25 Sep 2012 02:09:44 +0000"  >&lt;p&gt;FanYong,&lt;/p&gt;

&lt;p&gt;could you comment?&lt;/p&gt;</comment>
                            <comment id="45496" author="yong.fan" created="Tue, 25 Sep 2012 03:24:23 +0000"  >&lt;p&gt;I do not think so. I found some logs on the Lustre client (client2, client-26vm6):&lt;/p&gt;


&lt;p&gt;=====================================&lt;br/&gt;
00:49:41:LustreError: 11188:0:(genops.c:309:class_newdev()) Device lustre-MDT0000-mdc-ffff88005073c800 already exists at 13, won&apos;t add&lt;br/&gt;
00:49:41:LustreError: 11188:0:(obd_config.c:365:class_attach()) Cannot create device lustre-MDT0000-mdc-ffff88005073c800 of type mdc : -17&lt;br/&gt;
00:49:41:LustreError: 11188:0:(obd_config.c:1499:class_config_llog_handler()) Err -17 on cfg command:&lt;br/&gt;
00:49:41:Lustre:    cmd=cf001 0:lustre-MDT0000-mdc  1:mdc  2:lustre-clilmv_UUID  &lt;br/&gt;
00:49:41:LustreError: 15c-8: MGC10.10.4.150@tcp: The configuration from log &apos;lustre-client&apos; failed (-17). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
00:49:41:LustreError: 11185:0:(llite_lib.c:998:ll_fill_super()) Unable to process log: -17&lt;br/&gt;
00:49:41:Lustre: Unmounted lustre-client&lt;br/&gt;
00:49:41:LustreError: 11185:0:(ldlm_request.c:1166:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
00:49:41:LustreError: 11185:0:(ldlm_request.c:1166:ldlm_cli_cancel_req()) Skipped 1 previous similar message&lt;br/&gt;
00:49:41:LustreError: 11185:0:(ldlm_request.c:1792:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
00:49:41:LustreError: 11185:0:(ldlm_request.c:1792:ldlm_cli_cancel_list()) Skipped 1 previous similar message&lt;br/&gt;
00:49:41:LustreError: 11185:0:(obd_mount.c:2569:lustre_fill_super()) Unable to mount  (-17)&lt;br/&gt;
00:49:41:Lustre: DEBUG MARKER: /usr/sbin/lctl mark  parallel-scale-nfsv3 : @@@@@@ FAIL: failed to mount lustre after nfs test&lt;br/&gt;
=====================================&lt;/p&gt;


&lt;p&gt;I do not know the system configuration exactly. But I guess the client2 is NFS server, right? If yes, then just because the Lustre client fail to mount client-26vm3 (MDS&amp;amp;MGS), then cause NFS server re-export failure. As for why client2 failed to mount, it seems related with former uncleaned environment:&lt;/p&gt;

&lt;p&gt;===================================&lt;br/&gt;
struct obd_device *class_newdev(const char *type_name, const char *name)&lt;br/&gt;
{&lt;br/&gt;
...&lt;/p&gt;

&lt;p&gt;                if (obd &amp;amp;&amp;amp; obd-&amp;gt;obd_name &amp;amp;&amp;amp;&lt;br/&gt;
                    (strcmp(name, obd-&amp;gt;obd_name) == 0)) {&lt;br/&gt;
                        CERROR(&quot;Device %s already exists at %d, won&apos;t add\n&quot;,&lt;br/&gt;
                               name, i);&lt;br/&gt;
                        if (result) &lt;/p&gt;
{
                                LASSERTF(result-&amp;gt;obd_magic == OBD_DEVICE_MAGIC,
                                         &quot;%p obd_magic %08x != %08x\n&quot;, result,
                                         result-&amp;gt;obd_magic, OBD_DEVICE_MAGIC);
                                LASSERTF(result-&amp;gt;obd_minor == new_obd_minor,
                                         &quot;%p obd_minor %d != %d\n&quot;, result,
                                         result-&amp;gt;obd_minor, new_obd_minor);

                                obd_devs[result-&amp;gt;obd_minor] = NULL;
                                result-&amp;gt;obd_name[0]=&apos;\0&apos;;
                         }
&lt;p&gt;                        result = ERR_PTR(-EEXIST);&lt;br/&gt;
                        break;&lt;br/&gt;
                }&lt;br/&gt;
...&lt;br/&gt;
}&lt;br/&gt;
===================================&lt;/p&gt;

&lt;p&gt;We need the lustre kernel log to find why the environment is not cleaned. Without such log, we cannot know much.&lt;/p&gt;</comment>
                            <comment id="45515" author="mdiep" created="Tue, 25 Sep 2012 10:36:51 +0000"  >&lt;p&gt;Fanyoung, you can find lustre debug log here on brent&lt;br/&gt;
/home/autotest/logdir/test_logs/2012-09-22/lustre-b2_3-el6-x86_64_&lt;em&gt;24&lt;/em&gt;_-7f2b7f4b9360/parallel-scale-nfsv3..debug_log*&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv3sn:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4142</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>