<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:49:02 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-5159] Lustre MGS/MDT fails to start using initscripts using 2.4.2 based packages</title>
                <link>https://jira.whamcloud.com/browse/LU-5159</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;I set up a small Lustre filesystem inside of a few VMs running our TOSS 2.2 packages, and the initscript is failing to mount the MGS and MDT when run after a reboot of the MGS. I think this might be a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1279&quot; title=&quot;failure trying to mount two targets at the same time after boot&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1279&quot;&gt;&lt;del&gt;LU-1279&lt;/del&gt;&lt;/a&gt;, so feel free to mark it a duplicate if that&apos;s the case.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;-bash-4.1# dmesg -c &amp;gt; /dev/null
-bash-4.1# time /etc/init.d/lustre start
Mounting stotch-mds1/mgs0 on /mnt/lustre/local/stotch-MGS0000
Mounting stotch-mds1/mdt0 on /mnt/lustre/local/stotch-MDT0000
mount.lustre: mount stotch-mds1/mgs0 at /mnt/lustre/local/stotch-MGS0000 failed: No such device
Are the lustre modules loaded?
Check /etc/modprobe.conf and /proc/filesystems
mount.lustre: mount stotch-mds1/mdt0 at /mnt/lustre/local/stotch-MDT0000 failed: Input/output error
Is the MGS running?

real    7m34.545s
user    0m0.427s
sys     0m0.173s

-bash-4.1# mount
/dev/mapper/VolGroup-lv_root on / type ext4 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw)
/dev/vda1 on /boot type ext4 (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)

-bash-4.1# dmesg
LNet: HW CPU cores: 4, npartitions: 1
alg: No test for crc32 (crc32-table)
alg: No test for adler32 (adler32-zlib)
padlock: VIA PadLock Hash Engine not detected.
Lustre: Lustre: Build Version: 2.4.2-11chaos-11chaos--PRISTINE-2.6.32-431.17.2.1chaos.ch5.2.x86_64
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol RQF_FLD_QUERY
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol req_capsule_server_pack
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol req_capsule_client_get
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol ptlrpc_queue_wait
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol req_capsule_fini
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol req_capsule_init
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol req_capsule_set
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol req_capsule_server_get
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol ptlrpc_at_set_req_timeout
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol ptlrpc_request_alloc_pack
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol RMF_FLD_OPC
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol ptlrpc_request_set_replen
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol RMF_FLD_MDFLD
fld: gave up waiting for init of module ptlrpc.
fld: Unknown symbol ptlrpc_req_finished
LNet: Added LNI 192.168.2.90@tcp [8/256/0/180]
LNet: Accept secure, port 988
LustreError: 2927:0:(client.c:1053:ptlrpc_import_delay_req()) @@@ send limit expired   req@ffff880068beb000 x1470206796890120/t0(0) o253-&amp;gt;MGC192.168.2.90@tcp@0@lo:26/25 lens 4768/4768 e 0 to 0 dl 0 ref 2 fl Rpc:W/0/ffffffff rc 0/-1
LustreError: 2927:0:(obd_mount_server.c:1140:server_register_target()) stotch-MDT0000: error registering with the MGS: rc = -5 (not fatal)
LustreError: 2927:0:(client.c:1053:ptlrpc_import_delay_req()) @@@ send limit expired   req@ffff880068beb000 x1470206796890124/t0(0) o101-&amp;gt;MGC192.168.2.90@tcp@0@lo:26/25 lens 328/344 e 0 to 0 dl 0 ref 2 fl Rpc:W/0/ffffffff rc 0/-1
LustreError: 2927:0:(client.c:1053:ptlrpc_import_delay_req()) @@@ send limit expired   req@ffff880068beb000 x1470206796890128/t0(0) o101-&amp;gt;MGC192.168.2.90@tcp@0@lo:26/25 lens 328/344 e 0 to 0 dl 0 ref 2 fl Rpc:W/0/ffffffff rc 0/-1
LustreError: 15c-8: MGC192.168.2.90@tcp: The configuration from log &apos;stotch-MDT0000&apos; failed (-5). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
LustreError: 2927:0:(obd_mount_server.c:1273:server_start_targets()) failed to start server stotch-MDT0000: -5
Lustre: stotch-MDT0000: Unable to start target: -5
LustreError: 2927:0:(obd_mount_server.c:865:lustre_disconnect_lwp()) stotch-MDT0000-lwp-MDT0000: Can&apos;t end config log stotch-client.
LustreError: 2927:0:(obd_mount_server.c:1442:server_put_super()) stotch-MDT0000: failed to disconnect lwp. (rc=-2)
LustreError: 2927:0:(obd_mount_server.c:1472:server_put_super()) no obd stotch-MDT0000
Lustre: server umount stotch-MDT0000 complete
LustreError: 2927:0:(obd_mount.c:1290:lustre_fill_super()) Unable to mount  (-5)

-bash-4.1# rpm -qa | grep lustre
lustre-tools-llnl-1.6-1.ch5.2.x86_64
lustre-osd-ldiskfs-2.4.2-11chaos_2.6.32_431.17.2.1chaos.ch5.2.ch5.2.x86_64
lustre-modules-2.4.2-11chaos_2.6.32_431.17.2.1chaos.ch5.2.ch5.2.x86_64
lustre-osd-zfs-2.4.2-11chaos_2.6.32_431.17.2.1chaos.ch5.2.ch5.2.x86_64
lustre-debuginfo-2.4.2-11chaos_2.6.32_431.17.2.1chaos.ch5.2.ch5.2.x86_64
lustre-2.4.2-11chaos_2.6.32_431.17.2.1chaos.ch5.2.ch5.2.x86_64

-bash-4.1# cat /etc/ldev.conf 
stotch-mds1 - stotch-MGS0000 zfs:stotch-mds1/mgs0
stotch-mds1 - stotch-MDT0000 zfs:stotch-mds1/mdt0
stotch-oss1 - stotch-OST0000 zfs:stotch-oss1/ost0
stotch-oss2 - stotch-OST0001 zfs:stotch-oss2/ost0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Is this expected behavior? I assume not.&lt;/p&gt;

&lt;p&gt;If I run the script a second time, everything mounts just fine (and &lt;b&gt;much&lt;/b&gt; faster):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;-bash-4.1# time /etc/init.d/lustre start
Mounting stotch-mds1/mgs0 on /mnt/lustre/local/stotch-MGS0000
Mounting stotch-mds1/mdt0 on /mnt/lustre/local/stotch-MDT0000

real    0m4.484s
user    0m0.439s
sys     0m0.228s

-bash-4.1# mount
/dev/mapper/VolGroup-lv_root on / type ext4 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw)
/dev/vda1 on /boot type ext4 (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
stotch-mds1/mgs0 on /mnt/lustre/local/stotch-MGS0000 type lustre (rw)
stotch-mds1/mdt0 on /mnt/lustre/local/stotch-MDT0000 type lustre (rw)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="25063">LU-5159</key>
            <summary>Lustre MGS/MDT fails to start using initscripts using 2.4.2 based packages</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="hongchao.zhang">Hongchao Zhang</assignee>
                                    <reporter username="prakash">Prakash Surya</reporter>
                        <labels>
                            <label>llnl</label>
                    </labels>
                <created>Sat, 7 Jun 2014 00:12:39 +0000</created>
                <updated>Fri, 29 Aug 2014 17:29:56 +0000</updated>
                            <resolved>Fri, 29 Aug 2014 17:29:56 +0000</resolved>
                                    <version>Lustre 2.4.2</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="86067" author="pjones" created="Sun, 8 Jun 2014 12:48:55 +0000"  >&lt;p&gt;Hongchao&lt;/p&gt;

&lt;p&gt;Could you please comment on this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="86564" author="hongchao.zhang" created="Fri, 13 Jun 2014 15:20:07 +0000"  >&lt;p&gt;I have reproduced this issue, and it is a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1279&quot; title=&quot;failure trying to mount two targets at the same time after boot&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1279&quot;&gt;&lt;del&gt;LU-1279&lt;/del&gt;&lt;/a&gt;, which &quot;modprobe&quot; regards the module with its state == &quot;MODULE_STATE_COMING&quot;&lt;br/&gt;
(in this case, is &quot;lnet&quot;) as a valid one, then it continues to load the following modules, which will take the &quot;module_mutex&quot; and wait the previous module (lnet)&lt;br/&gt;
to be complete (the state changes to &quot;MODULE_STATE_LIVE&quot;) in &quot;resolve_symbol&quot;,  but the mutex lock prevents its submodules (in this case, it&apos;s various&lt;br/&gt;
klnd modules) to be loaded, then it will fail with &quot;XXX1: gave up waiting for init of module XXX2. XXX1: Unknown symbol XXXXXX&quot;.&lt;/p&gt;</comment>
                            <comment id="92829" author="pjones" created="Fri, 29 Aug 2014 17:29:56 +0000"  >&lt;p&gt;Closing as a duplicate. It looks like there is a recently-landed patch for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1279&quot; title=&quot;failure trying to mount two targets at the same time after boot&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1279&quot;&gt;&lt;del&gt;LU-1279&lt;/del&gt;&lt;/a&gt; to try out.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="13847">LU-1279</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwnxr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>14229</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>