<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:43:13 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-11363] sanity-sec test 31 fails with  &apos;unable to remount client&apos; </title>
                <link>https://jira.whamcloud.com/browse/LU-11363</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;sanity-sec test_31 was added by the patch at &lt;a href=&quot;https://review.whamcloud.com/#/c/32590/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/#/c/32590/&lt;/a&gt; and merged with master on September 10, 2018. So far, the test is either failing or crashing for review-dne-zfs-part-2 only. &lt;/p&gt;

&lt;p&gt;Looking at the logs for the failure &lt;a href=&quot;https://testing.whamcloud.com/test_sets/c7881c1e-b5b7-11e8-8c12-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/c7881c1e-b5b7-11e8-8c12-52540065bddc&lt;/a&gt;, from the test_log, for every target, we see a problem when tunefs is called&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: trevis-5vm8 tunefs.lustre --quiet --writeconf lustre-mdt1/mdt1
trevis-5vm8: 
trevis-5vm8: tunefs.lustre FATAL: Device lustre-mdt1/mdt1 has not been formatted with mkfs.lustre
trevis-5vm8: tunefs.lustre: exiting with 19 (No such device)
checking for existing Lustre data: not found
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;From there, we see a variety of other errors&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Started lustre-MDT0003
CMD: trevis-5vm9 lctl get_param -n mdt.lustre-MDT0003.identity_upcall
/usr/lib64/lustre/tests/test-framework.sh: line 4452: mdt.lustre-MDT0000.identity_upcall: command not found
CMD: trevis-5vm9 lctl set_param -n mdt.lustre-MDT0003.identity_upcall &quot;NONE&quot;
CMD: trevis-5vm9 lctl set_param -n mdt/lustre-MDT0003/identity_flush=-1
&#8230;
CMD: trevis-5vm5.trevis.whamcloud.com lctl dl | grep &apos; IN osc &apos; 2&amp;gt;/dev/null | wc -l
error: get_param: param_path &apos;mdc/*/connect_flags&apos;: No such file or directory
jobstats not supported by server
disable quota as required
CMD: trevis-5vm8 /usr/sbin/lctl list_nids | grep tcp999
Starting client: trevis-5vm5.trevis.whamcloud.com:  -o user_xattr,flock,network=tcp999 10.9.5.8@tcp999:/lustre /mnt/lustre
CMD: trevis-5vm5.trevis.whamcloud.com mkdir -p /mnt/lustre
CMD: trevis-5vm5.trevis.whamcloud.com mount -t lustre -o user_xattr,flock,network=tcp999 10.9.5.8@tcp999:/lustre /mnt/lustre
mount.lustre: mount 10.9.5.8@tcp999:/lustre at /mnt/lustre failed: Invalid argument
This may have multiple causes.
Is &apos;lustre&apos; the correct filesystem name?
Are the mount options correct?
Check the syslog for more info.
unconfigure:
    - lnet:
          errno: -16
          descr: &quot;LNet unconfigure error: Device or resource busy&quot;
Starting client: trevis-5vm5.trevis.whamcloud.com:  -o user_xattr,flock,network=tcp999 10.9.5.8@tcp999:/lustre /mnt/lustre
CMD: trevis-5vm5.trevis.whamcloud.com mkdir -p /mnt/lustre
CMD: trevis-5vm5.trevis.whamcloud.com mount -t lustre -o user_xattr,flock,network=tcp999 10.9.5.8@tcp999:/lustre /mnt/lustre
mount.lustre: mount 10.9.5.8@tcp999:/lustre at /mnt/lustre failed: No such file or directory
Is the MGS specification correct?
Is the filesystem name correct?
If upgrading, is the copied client log valid? (see upgrade docs)
 sanity-sec test_31: @@@@@@ FAIL: unable to remount client 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The following are links to logs for other test session failures for this test&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/6d51eee0-b54f-11e8-b86b-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/6d51eee0-b54f-11e8-b86b-52540065bddc&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/a0a5d418-b555-11e8-a7de-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/a0a5d418-b555-11e8-a7de-52540065bddc&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/6070a87e-b59f-11e8-8c12-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/6070a87e-b59f-11e8-8c12-52540065bddc&lt;/a&gt;&lt;/p&gt;


&lt;p&gt;When sanity-sec test_31 crashes, we see the following in the kernel-crash log&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[ 9311.019503] Lustre: DEBUG MARKER: mount -t lustre -o user_xattr,flock,network=tcp999 10.2.8.122@tcp999:/lustre /mnt/lustre
[ 9311.029516] LustreError: 21790:0:(obd_mount.c:1422:lmd_parse()) LNet Dynamic Peer Discovery is enabled on this node. &apos;network&apos; mount option cannot be taken into account.
[ 9311.031037] LustreError: 21790:0:(obd_mount.c:1520:lmd_parse()) Bad mount options user_xattr,flock,network=tcp999,device=10.2.8.122@tcp999:/lustre
[ 9311.032361] LustreError: 21790:0:(obd_mount.c:1608:lustre_fill_super()) Unable to mount  (-22)
[ 9312.035556] LNet: Removed LNI 10.2.8.119@tcp999
[ 9312.170496] Key type lgssc unregistered
[ 9312.171026] Lustre: 21892:0:(gss_mech_switch.c:80:lgss_mech_unregister()) Unregister krb5 mechanism
[ 9314.495561] LNet: Removed LNI 10.2.8.119@tcp
[ 9314.657567] LNet: HW NUMA nodes: 1, HW CPU cores: 2, npartitions: 1
[ 9314.661048] alg: No test for adler32 (adler32-zlib)
[ 9315.459156] Lustre: Lustre: Build Version: 2.11.54_104_gd365ea2
[ 9315.529642] LNet: Added LNI 10.2.8.119@tcp [8/256/0/180]
[ 9315.530284] LNet: Accept all, port 7988
[ 9315.537592] LNet: Added LNI 10.2.8.119@tcp999 [8/256/0/180]
[ 9315.541706] Lustre: DEBUG MARKER: mkdir -p /mnt/lustre
[ 9315.550513] Lustre: DEBUG MARKER: mount -t lustre -o user_xattr,flock,network=tcp999 10.2.8.122@tcp999:/lustre /mnt/lustre
[ 9315.605193] LustreError: 22006:0:(ldlm_lib.c:492:client_obd_setup()) can&apos;t add initial connection
[ 9315.606173] LustreError: 22006:0:(obd_config.c:559:class_setup()) setup lustre-MDT0000-mdc-ffff8c373b3f5000 failed (-2)
[ 9315.607252] LustreError: 22006:0:(obd_config.c:1835:class_config_llog_handler()) MGC10.2.8.122@tcp999: cfg command failed: rc = -2
[ 9315.608409] Lustre:    cmd=cf003 0:lustre-MDT0000-mdc  1:lustre-MDT0000_UUID  2:10.2.8.122@tcp  

[ 9315.609546] LustreError: 108:0:(connection.c:96:ptlrpc_connection_put()) ASSERTION( atomic_read(&amp;amp;conn-&amp;gt;c_refcount) &amp;gt; 1 ) failed: 
[ 9315.609934] LustreError: 15c-8: MGC10.2.8.122@tcp999: The configuration from log &apos;lustre-client&apos; failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
[ 9315.613151] LustreError: 108:0:(connection.c:96:ptlrpc_connection_put()) LBUG
[ 9315.613864] Pid: 108, comm: kworker/1:2 3.10.0-862.9.1.el7.x86_64 #1 SMP Mon Jul 16 16:29:36 UTC 2018
[ 9315.614783] Call Trace:
[ 9315.615088]  [&amp;lt;ffffffffc07847cc&amp;gt;] libcfs_call_trace+0x8c/0xc0 [libcfs]
[ 9315.615779]  [&amp;lt;ffffffffc078487c&amp;gt;] lbug_with_loc+0x4c/0xa0 [libcfs]
[ 9315.616419]  [&amp;lt;ffffffffc0a7aac3&amp;gt;] ptlrpc_connection_put+0x213/0x220 [ptlrpc]
[ 9315.617180]  [&amp;lt;ffffffffc08b4c15&amp;gt;] obd_zombie_imp_cull+0x65/0x3e0 [obdclass]
[ 9315.617705] LustreError: 21994:0:(obd_config.c:610:class_cleanup()) Device 3 not setup
[ 9315.617739] Lustre: Unmounted lustre-client
[ 9315.619443]  [&amp;lt;ffffffffbd8b35ef&amp;gt;] process_one_work+0x17f/0x440
[ 9315.620210]  [&amp;lt;ffffffffbd8b4686&amp;gt;] worker_thread+0x126/0x3c0
[ 9315.620798]  [&amp;lt;ffffffffbd8bb621&amp;gt;] kthread+0xd1/0xe0
[ 9315.621336]  [&amp;lt;ffffffffbdf205f7&amp;gt;] ret_from_fork_nospec_end+0x0/0x39
[ 9315.622164]  [&amp;lt;ffffffffffffffff&amp;gt;] 0xffffffffffffffff
[ 9315.622720] Kernel panic - not syncing: LBUG
[ 9315.623235] CPU: 1 PID: 108 Comm: kworker/1:2 Kdump: loaded Tainted: G           OE  ------------   3.10.0-862.9.1.el7.x86_64 #1
[ 9315.624371] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
[ 9315.624956] Workqueue: obd_zombid obd_zombie_imp_cull [obdclass]
[ 9315.625577] Call Trace:
[ 9315.625859]  [&amp;lt;ffffffffbdf0e84e&amp;gt;] dump_stack+0x19/0x1b
[ 9315.626383]  [&amp;lt;ffffffffbdf08b50&amp;gt;] panic+0xe8/0x21f
[ 9315.626868]  [&amp;lt;ffffffffc07848cb&amp;gt;] lbug_with_loc+0x9b/0xa0 [libcfs]
[ 9315.627502]  [&amp;lt;ffffffffc0a7aac3&amp;gt;] ptlrpc_connection_put+0x213/0x220 [ptlrpc]
[ 9315.628222]  [&amp;lt;ffffffffc08b4c15&amp;gt;] obd_zombie_imp_cull+0x65/0x3e0 [obdclass]
[ 9315.628918]  [&amp;lt;ffffffffbd8b35ef&amp;gt;] process_one_work+0x17f/0x440
[ 9315.629498]  [&amp;lt;ffffffffbd8b4686&amp;gt;] worker_thread+0x126/0x3c0
[ 9315.630059]  [&amp;lt;ffffffffbd8b4560&amp;gt;] ? manage_workers.isra.24+0x2a0/0x2a0
[ 9315.630732]  [&amp;lt;ffffffffbd8bb621&amp;gt;] kthread+0xd1/0xe0
[ 9315.631234]  [&amp;lt;ffffffffbd8bb550&amp;gt;] ? insert_kthread_work+0x40/0x40
[ 9315.631839]  [&amp;lt;ffffffffbdf205f7&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[ 9315.632490]  [&amp;lt;ffffffffbd8bb550&amp;gt;] ? insert_kthread_work+0x40/0x40
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Logs for when sanity-sec test 31 crashes are at&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/4ec4717a-b5b6-11e8-b86b-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/4ec4717a-b5b6-11e8-b86b-52540065bddc&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/fe8c7708-b569-11e8-a7de-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/fe8c7708-b569-11e8-a7de-52540065bddc&lt;/a&gt;&lt;/p&gt;</description>
                <environment>DNE/ZFS</environment>
        <key id="53274">LU-11363</key>
            <summary>sanity-sec test 31 fails with  &apos;unable to remount client&apos; </summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="jamesanunez">James Nunez</reporter>
                        <labels>
                            <label>DNE</label>
                            <label>zfs</label>
                    </labels>
                <created>Tue, 11 Sep 2018 16:01:39 +0000</created>
                <updated>Sat, 15 Dec 2018 18:05:59 +0000</updated>
                            <resolved>Tue, 16 Oct 2018 15:53:52 +0000</resolved>
                                    <version>Lustre 2.12.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="233344" author="tappro" created="Tue, 11 Sep 2018 17:39:10 +0000"  >&lt;p&gt;+1 on master&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/4e6f80fa-b5c2-11e8-b86b-52540065bddc&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/4e6f80fa-b5c2-11e8-b86b-52540065bddc&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="233446" author="tappro" created="Thu, 13 Sep 2018 06:27:39 +0000"  >&lt;p&gt;It has more than 60% failure rate now&lt;/p&gt;</comment>
                            <comment id="233461" author="jamesanunez" created="Thu, 13 Sep 2018 14:47:01 +0000"  >&lt;p&gt;Patch &lt;a href=&quot;https://review.whamcloud.com/#/c/33139/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/#/c/33139/&lt;/a&gt; reverted the patch that added sanity-sec test 31. Thus, all patches should be rebased to get this update.&lt;/p&gt;</comment>
                            <comment id="234015" author="sebastien" created="Wed, 26 Sep 2018 14:08:09 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;In patch &lt;a href=&quot;https://review.whamcloud.com/33189&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/33189&lt;/a&gt; , I modified writeconf_all() so that it uses vdev instead of dev name.&lt;br/&gt;
Unfortunately, test_31 still fails, with following messages:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: trevis-33vm7 tunefs.lustre --quiet --writeconf /dev/lvm-Role_OSS/P1
trevis-33vm7: 
trevis-33vm7: tunefs.lustre FATAL: Device /dev/lvm-Role_OSS/P1 has not been formatted with mkfs.lustre
trevis-33vm7: tunefs.lustre: exiting with 19 (No such device)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;However, target was formatted with:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;CMD: trevis-33vm7 mkfs.lustre --mgsnode=trevis-33vm8@tcp --fsname=lustre --ost --index=0 --param=sys.timeout=20 --backfstype=zfs --device-size=9950986 --reformat lustre-ost1/ost1 /dev/lvm-Role_OSS/P1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;(as seen in lustre-initialization-1 logs).&lt;/p&gt;

&lt;p&gt;So /dev/lvm-Role_OSS/P1 should be a valid device.&lt;/p&gt;

&lt;p&gt;Or does the problem stem from the fact that tunefs.lustre cannot be used on targets using ZFS as backend?&lt;/p&gt;</comment>
                            <comment id="234055" author="jamesanunez" created="Thu, 27 Sep 2018 08:23:31 +0000"  >&lt;p&gt;I think I understand the issue with the original patch 32590. In sanity-sec test 31, we call stopall() to stop all servers which calls stop(). For ZFS, stop() will export the zpool and, thus, tunefs.lustre --writeconf will fail on exported zpools. We need to either set KEEP_ZPOOL to true or import the zpool after calling stopall(). &lt;/p&gt;</comment>
                            <comment id="234994" author="sebastien" created="Tue, 16 Oct 2018 15:51:26 +0000"  >&lt;p&gt;I think this ticket can be closed now that patch at &lt;a href=&quot;https://review.whamcloud.com/33189&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/33189&lt;/a&gt; has landed.&lt;/p&gt;</comment>
                            <comment id="234995" author="pjones" created="Tue, 16 Oct 2018 15:53:52 +0000"  >&lt;p&gt;ok sure.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="52400">LU-11057</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i0026v:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>