<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:39:50 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4120] Not a bug file system unavailable.</title>
                <link>https://jira.whamcloud.com/browse/LU-4120</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We have had a file system down to be upgraded since Monday.&lt;/p&gt;

&lt;p&gt;We can not bring the filesystem in to operation.&lt;/p&gt;

&lt;p&gt;We have been talking with DDN support since Tuesday.&lt;/p&gt;</description>
                <environment>ubuntu 12.04 LTS</environment>
        <key id="21501">LU-4120</key>
            <summary>Not a bug file system unavailable.</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="6">Not a Bug</resolution>
                                        <assignee username="bogl">Bob Glossman</assignee>
                                    <reporter username="james beal">James Beal</reporter>
                        <labels>
                    </labels>
                <created>Fri, 18 Oct 2013 13:52:00 +0000</created>
                <updated>Tue, 28 Oct 2014 20:46:23 +0000</updated>
                            <resolved>Tue, 28 Oct 2014 20:46:23 +0000</resolved>
                                    <version>Lustre 1.8.9</version>
                    <version>Lustre 2.4.1</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="69278" author="james beal" created="Fri, 18 Oct 2013 13:52:37 +0000"  >
&lt;p&gt;The system was using the default e2fsprogs until very recently. We have now installed 1.42.7.wc1 with the patch to get the quota working.&lt;/p&gt;

&lt;p&gt;The system originally had both  of the servers it could be on listed as failnode rather than just the other node in the pair.&lt;/p&gt;

&lt;p&gt;At one point we had 26 out of 30 OSS&apos;s available but the last 4 would not come on line ( this was after only those 4 were working ).&lt;/p&gt;

&lt;p&gt;We have replaced the kernel modules and user space tools after &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4111&quot; title=&quot;client kernel panic&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4111&quot;&gt;&lt;del&gt;LU-4111&lt;/del&gt;&lt;/a&gt;, however this configuration has worked previously on three other filesystems.&lt;/p&gt;

&lt;p&gt;At present when we mount an OST it does not become up in the MDT&apos;s lctl  device list but stays in AT eg.&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# lctl dl&lt;br/&gt;
  0 UP mgs MGS MGS 7&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp 3f2155e6-d9ad-156f-d431-406faeb78ef5 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
  5 AT osc lus04-OST0000-osc lus04-mdtlov_UUID 1&lt;/p&gt;

&lt;p&gt;We have downgraded the system to 1.8.9wc1 and the system performs in the same way.&lt;/p&gt;</comment>
                            <comment id="69279" author="james beal" created="Fri, 18 Oct 2013 13:58:36 +0000"  >&lt;p&gt;MGS/MDT mount logs&lt;br/&gt;
Oct 18 14:29:06 lus04-mds1 kernel: [ 7549.035620] Lustre: Build Version: v1_8_9_WC1sanger1--PRISTINE-2.6.32.59-sles-lustre-1.8.8wc1&lt;br/&gt;
Oct 18 14:29:06 lus04-mds1 kernel: [ 7549.081427] Lustre: Added LNI 172.17.148.4@tcp &lt;span class=&quot;error&quot;&gt;&amp;#91;8/256/0/180&amp;#93;&lt;/span&gt;&lt;br/&gt;
Oct 18 14:29:06 lus04-mds1 kernel: [ 7549.081520] Lustre: Accept secure, port 988&lt;br/&gt;
Oct 18 14:29:06 lus04-mds1 kernel: [ 7549.135974] Lustre: Lustre Client File System; &lt;a href=&quot;http://www.lustre.org/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://www.lustre.org/&lt;/a&gt;&lt;br/&gt;
Oct 18 14:29:07 lus04-mds1 kernel: [ 7549.145508] LDISKFS-fs (dm-1): barriers disabled&lt;br/&gt;
Oct 18 14:29:07 lus04-mds1 kernel: [ 7549.147409] LDISKFS-fs (dm-1): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 18 14:29:07 lus04-mds1 kernel: [ 7549.261222] LDISKFS-fs (dm-1): barriers disabled&lt;br/&gt;
Oct 18 14:29:07 lus04-mds1 kernel: [ 7549.262804] LDISKFS-fs (dm-1): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 18 14:29:07 lus04-mds1 kernel: [ 7549.289975] Lustre: MGS MGS started&lt;br/&gt;
Oct 18 14:29:07 lus04-mds1 kernel: [ 7549.290696] Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.133538] LDISKFS-fs (dm-0): barriers disabled&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.136647] LDISKFS-fs (dm-0): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.265245] LDISKFS-fs (dm-0): barriers disabled&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.267566] LDISKFS-fs (dm-0): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.298400] LustreError: 13c-e: Client log lus04-client has disappeared! Regenerating all logs.&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.298589] Lustre: MGS: Logs for fs lus04 were removed by user request.  All servers must be restarted in order to regenerate the logs.&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.313147] Lustre: Enabling user_xattr&lt;br/&gt;
Oct 18 14:29:16 lus04-mds1 kernel: [ 7558.343685] Lustre: lus04-MDT0000: Now serving lus04-MDT0000 on /dev/mapper/lus04--mdt0-lus04 with recovery enabled&lt;br/&gt;
Oct 18 14:29:47 lus04-mds1 kernel: [ 7590.037063] Lustre: MGS: Regenerating lus04-OST0000 log by user request.&lt;br/&gt;
Oct 18 14:29:51 lus04-mds1 kernel: [ 7593.422391] LustreError: 21629:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
Oct 18 14:29:51 lus04-mds1 kernel: [ 7593.422561] LustreError: 21629:0:(obd_config.c:372:class_setup()) setup lus04-OST0000-osc failed (-2)&lt;br/&gt;
Oct 18 14:29:51 lus04-mds1 kernel: [ 7593.422695] LustreError: 21629:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
Oct 18 14:29:51 lus04-mds1 kernel: [ 7593.422832] Lustre:    cmd=cf003 0:lus04-OST0000-osc  1:lus04-OST0000_UUID  2:0@&amp;lt;0:0&amp;gt; &lt;/p&gt;

&lt;p&gt;OSS mount logs&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.183166&amp;#93;&lt;/span&gt; Lustre: Build Version: v1_8_9_WC1sanger1--PRISTINE-2.6.32.59-sles-lustre-1.8.8wc1&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.229796&amp;#93;&lt;/span&gt; Lustre: Added LNI 172.17.148.6@tcp &lt;span class=&quot;error&quot;&gt;&amp;#91;8/256/0/180&amp;#93;&lt;/span&gt;&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.229842&amp;#93;&lt;/span&gt; Lustre: Accept secure, port 988&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.285325&amp;#93;&lt;/span&gt; Lustre: Lustre Client File System; &lt;a href=&quot;http://www.lustre.org/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://www.lustre.org/&lt;/a&gt;&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.370273&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): barriers disabled&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.392095&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.578277&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): barriers disabled&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.587805&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.601175&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 18 14:29:47 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11025.644539&amp;#93;&lt;/span&gt; Lustre: Filtering OBD driver; &lt;a href=&quot;http://wiki.whamcloud.com/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://wiki.whamcloud.com/&lt;/a&gt;&lt;br/&gt;
Oct 18 14:29:48 lus04-oss1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;11026.007326&amp;#93;&lt;/span&gt; Lustre: lus04-OST0000: Now serving lus04-OST0000 on /dev/mapper/vd00 with recovery enabled&lt;br/&gt;
root@lus04-oss1:/# cat /proc/fs/lustre/obdfilter/lus04-OST0000/recovery_status &lt;br/&gt;
status: INACTIVE&lt;/p&gt;</comment>
                            <comment id="69299" author="jlevi" created="Fri, 18 Oct 2013 16:24:10 +0000"  >&lt;p&gt;Bob,&lt;br/&gt;
Could you please have a look at this one?&lt;br/&gt;
Thank you!&lt;/p&gt;</comment>
                            <comment id="69307" author="bogl" created="Fri, 18 Oct 2013 17:05:26 +0000"  >&lt;p&gt;Looking into possible causes.  I suspect it may be a side effect of upgrading/downgrading between 1.8.9 and 2.4.1.  Hope to have more and better information to give you soon.&lt;/p&gt;</comment>
                            <comment id="69309" author="james beal" created="Fri, 18 Oct 2013 17:13:16 +0000"  >&lt;p&gt;Bob we had the same symptoms when the system was running 2.4.1&lt;/p&gt;

&lt;p&gt;I can try and generate a log of the discussion with DDN and email or work out how to attach it as a private file.&lt;/p&gt;</comment>
                            <comment id="69321" author="bogl" created="Fri, 18 Oct 2013 19:08:10 +0000"  >&lt;p&gt;what state is your fs in now?  all umounted?  all mounted but some OSTs inactive? just MDT mounted?&lt;/p&gt;</comment>
                            <comment id="69328" author="james beal" created="Fri, 18 Oct 2013 20:08:25 +0000"  >&lt;p&gt;We have the MGS and MDT mounted and a single OSS.&lt;/p&gt;

&lt;p&gt;We have stopped as lctl dl shows on the MGS/MDT.&lt;/p&gt;

&lt;p&gt;  0 UP mgs MGS MGS 7&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp 3f2155e6-d9ad-156f-d431-406faeb78ef5 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
  5 AT osc lus04-OST0000-osc lus04-mdtlov_UUID 1&lt;/p&gt;

&lt;p&gt;Note the OSS is stuck in AT not up.&lt;/p&gt;</comment>
                            <comment id="69329" author="james beal" created="Fri, 18 Oct 2013 20:11:01 +0000"  >&lt;p&gt;Additionally when we try and mount the file system from am client&lt;/p&gt;

&lt;p&gt;root@isg-disc-mon-05:~# dmesg&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.147231&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.149699&amp;#93;&lt;/span&gt; LustreError: 2312:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.149912&amp;#93;&lt;/span&gt; LustreError: 2312:0:(obd_config.c:372:class_setup()) setup lus04-MDT0000-mdc-ffff8806181bcc00 failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.150204&amp;#93;&lt;/span&gt; LustreError: 2312:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.151628&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-MDT0000-mdc  1:lus04-MDT0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.151957&amp;#93;&lt;/span&gt; LustreError: 15c-8: MGC172.17.148.4@tcp: The configuration from log &apos;lus04-client&apos; failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.152374&amp;#93;&lt;/span&gt; LustreError: 2302:0:(llite_lib.c:1099:ll_fill_super()) Unable to process log: -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.152695&amp;#93;&lt;/span&gt; LustreError: 2302:0:(obd_config.c:443:class_cleanup()) Device 2 not setup&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.152975&amp;#93;&lt;/span&gt; LustreError: 2302:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.153170&amp;#93;&lt;/span&gt; LustreError: 2302:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.155666&amp;#93;&lt;/span&gt; Lustre: client lus04-client(ffff8806181bcc00) umount complete&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;127708.155791&amp;#93;&lt;/span&gt; LustreError: 2302:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-2)&lt;/p&gt;</comment>
                            <comment id="69335" author="bogl" created="Fri, 18 Oct 2013 20:54:02 +0000"  >&lt;p&gt;A couple of things to try and verify.  Some of these may be repeats of things you&apos;ve already done.&lt;/p&gt;

&lt;p&gt;Confirm you have the same version of e2fsprogs on all your servers.&lt;/p&gt;

&lt;p&gt;Unmount everything.&lt;/p&gt;

&lt;p&gt;Bring up at least lnet everywhere with &apos;modprobe lnet&apos;, should already be loaded on nodes where you have attempted mounts.  get the list of nids on every server with &apos;lctl list_nids&apos;.  see if all the nids shown look real and sensible.  The error msg Lustre: cmd=cf003 0:lus04-OST0000-osc 1:lus04-OST0000_UUID 2:0@&amp;lt;0:0&amp;gt; in particular suggests something wrong with nids. 2:0 doesn&apos;t look like a sensible value.&lt;/p&gt;

&lt;p&gt;Examine the config log with llog_reader.&lt;br/&gt;
Mount the MGS with -t ldiskfs.&lt;br/&gt;
do &apos;llog_reader &amp;lt;mountpoint&amp;gt;/CONFIGS/$FSNAME-client&apos;  where $FSNAME in your fs name (lus04?).&lt;br/&gt;
ummount it again.&lt;/p&gt;

&lt;p&gt;Do a fresh round of tunefs.lustre --writeconf on all your MDTs and OSTs.&lt;/p&gt;

&lt;p&gt;Try mounting again, MGS/MDT first, then OSTs in index order.&lt;/p&gt;

&lt;p&gt;Worst case if we can&apos;t repair this fs would it be possible to reformat?  Seems like the most drastic solution, but most likely to work.&lt;/p&gt;
</comment>
                            <comment id="69337" author="james beal" created="Fri, 18 Oct 2013 21:13:45 +0000"  >&lt;p&gt;&quot;A couple of things to try and verify. Some of these may be repeats of things you&apos;ve already done.&lt;br/&gt;
Confirm you have the same version of e2fsprogs on all your servers.&quot;&lt;/p&gt;

&lt;p&gt;lus04-mds1: e2fsck 1.42.7.wc1 (12-Apr-2013)&lt;br/&gt;
lus04-mds1: 	Using EXT2FS Library version 1.42.7.wc1, 12-Apr-2013&lt;br/&gt;
lus04-mds2: e2fsck 1.42.7.wc1 (12-Apr-2013)&lt;br/&gt;
lus04-mds2: 	Using EXT2FS Library version 1.42.7.wc1, 12-Apr-2013&lt;br/&gt;
lus04-oss1: e2fsck 1.42.7.wc1 (12-Apr-2013)&lt;br/&gt;
lus04-oss1: 	Using EXT2FS Library version 1.42.7.wc1, 12-Apr-2013&lt;br/&gt;
lus04-oss2: e2fsck 1.42.7.wc1 (12-Apr-2013)&lt;br/&gt;
lus04-oss2: 	Using EXT2FS Library version 1.42.7.wc1, 12-Apr-2013&lt;br/&gt;
lus04-oss3: e2fsck 1.42.7.wc1 (12-Apr-2013)&lt;br/&gt;
lus04-oss3: 	Using EXT2FS Library version 1.42.7.wc1, 12-Apr-2013&lt;br/&gt;
lus04-oss4: e2fsck 1.42.7.wc1 (12-Apr-2013)&lt;br/&gt;
lus04-oss4: 	Using EXT2FS Library version 1.42.7.wc1, 12-Apr-2013&lt;/p&gt;

&lt;p&gt;One thing I have noticed is that I can&apos;t list the nids if I only modprobe lnet rather than lustre.&lt;/p&gt;

&lt;p&gt;oot@it-admin:~# dsh -M -g lus04 &quot;modprobe lnet&quot;&lt;br/&gt;
root@it-admin:~# dsh -M -g lus04 &quot;lctl list_nids&quot;&lt;br/&gt;
lus04-mds1: 172.17.148.4@tcp&lt;br/&gt;
lus04-mds2: IOC_LIBCFS_GET_NI error 100: Network is down&lt;br/&gt;
lus04-oss1: 172.17.148.6@tcp&lt;br/&gt;
lus04-oss2: IOC_LIBCFS_GET_NI error 100: Network is down&lt;br/&gt;
lus04-oss3: IOC_LIBCFS_GET_NI error 100: Network is down&lt;br/&gt;
lus04-oss4: IOC_LIBCFS_GET_NI error 100: Network is down&lt;br/&gt;
root@it-admin:~# dsh -M -g lus04 &quot;modprobe lustre&quot;&lt;br/&gt;
root@it-admin:~# dsh -M -g lus04 &quot;lctl list_nids&quot;&lt;br/&gt;
lus04-mds1: 172.17.148.4@tcp&lt;br/&gt;
lus04-mds2: 172.17.148.5@tcp&lt;br/&gt;
lus04-oss1: 172.17.148.6@tcp&lt;br/&gt;
lus04-oss2: 172.17.148.7@tcp&lt;br/&gt;
lus04-oss3: 172.17.148.8@tcp&lt;br/&gt;
lus04-oss4: 172.17.148.9@tcp&lt;/p&gt;

&lt;p&gt;llog_reader CONFIGS/lus04-client &lt;br/&gt;
Header size : 8192&lt;br/&gt;
Time : Fri Oct 18 14:29:16 2013&lt;br/&gt;
Number of records: 18&lt;br/&gt;
Target uuid : config_uuid &lt;br/&gt;
-----------------------&lt;br/&gt;
#01 (224)marker   3 (flags=0x01, v1.8.9.0) lus04-clilov    &apos;lov setup&apos; Fri Oct 18 14:29:16 2013-&lt;br/&gt;
#02 (120)attach    0:lus04-clilov  1:lov  2:lus04-clilov_UUID  &lt;br/&gt;
#03 (168)lov_setup 0:lus04-clilov  1:(struct lov_desc)&lt;br/&gt;
		uuid=lus04-clilov_UUID  stripe:cnt=1 size=1048576 offset=18446744073709551615 pattern=0x1&lt;br/&gt;
#04 (224)marker   3 (flags=0x02, v1.8.9.0) lus04-clilov    &apos;lov setup&apos; Fri Oct 18 14:29:16 2013-&lt;br/&gt;
#05 (224)marker   4 (flags=0x01, v1.8.9.0) lus04-MDT0000   &apos;add mdc&apos; Fri Oct 18 14:29:16 2013-&lt;br/&gt;
#06 (128)attach    0:lus04-MDT0000-mdc  1:mdc  2:lus04-MDT0000-mdc_UUID  &lt;br/&gt;
#07 (128)setup     0:lus04-MDT0000-mdc  1:lus04-MDT0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
#08 (088)add_uuid  nid=172.17.148.5@tcp(0x20000ac119405)  0:  1:172.17.148.5@tcp  &lt;br/&gt;
#09 (112)add_conn  0:lus04-MDT0000-mdc  1:172.17.148.5@tcp  &lt;br/&gt;
#10 (128)mount_option 0:  1:lus04-client  2:lus04-clilov  3:lus04-MDT0000-mdc  &lt;br/&gt;
#11 (224)marker   4 (flags=0x02, v1.8.9.0) lus04-MDT0000   &apos;add mdc&apos; Fri Oct 18 14:29:16 2013-&lt;br/&gt;
#12 (224)marker   7 (flags=0x01, v1.8.9.0) lus04-OST0000   &apos;add osc&apos; Fri Oct 18 14:29:47 2013-&lt;br/&gt;
#13 (128)attach    0:lus04-OST0000-osc  1:osc  2:lus04-clilov_UUID  &lt;br/&gt;
#14 (128)setup     0:lus04-OST0000-osc  1:lus04-OST0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
#15 (088)add_uuid  nid=172.17.148.7@tcp(0x20000ac119407)  0:  1:172.17.148.7@tcp  &lt;br/&gt;
#16 (112)add_conn  0:lus04-OST0000-osc  1:172.17.148.7@tcp  &lt;br/&gt;
#17 (128)lov_modify_tgts add 0:lus04-clilov  1:lus04-OST0000_UUID  2:0  3:1  &lt;br/&gt;
#18 (224)marker   7 (flags=0x02, v1.8.9.0) lus04-OST0000   &apos;add osc&apos; Fri Oct 18 14:29:47 2013-&lt;/p&gt;

&lt;p&gt;I will do the &quot;tunefs.lustre --writeconf&quot; and report back later.&lt;/p&gt;

&lt;p&gt;While the filesystem has not got irreplaceable information and could be regenerated, the file system was in production. And we would prefer that we could copy the data off to another system before reformatting it.&lt;/p&gt;</comment>
                            <comment id="69338" author="bogl" created="Fri, 18 Oct 2013 21:18:45 +0000"  >&lt;p&gt;Oh, sorry. I left out a step.  After doing &apos;modprobe lnet&apos; do &apos;lctl net up&apos;. That should enable the following lctl cmd to work.&lt;/p&gt;</comment>
                            <comment id="69341" author="james beal" created="Fri, 18 Oct 2013 21:46:23 +0000"  >&lt;p&gt;&quot;Oh, sorry. I left out a step. After doing &apos;modprobe lnet&apos; do &apos;lctl net up&apos;. That should enable the following lctl cmd to work.&quot;&lt;/p&gt;

&lt;p&gt;root@it-admin:~# dsh -M -g lus04 &quot;lustre_rmmod&quot;&lt;br/&gt;
root@it-admin:~# dsh -M -g lus04 &quot;modprobe lnet&quot;&lt;br/&gt;
root@it-admin:~# dsh -M -g lus04 &quot;lctl net up&quot;&lt;br/&gt;
lus04-mds1: LNET configured&lt;br/&gt;
lus04-mds2: LNET configured&lt;br/&gt;
lus04-oss1: LNET configured&lt;br/&gt;
lus04-oss2: LNET configured&lt;br/&gt;
lus04-oss3: LNET configured&lt;br/&gt;
lus04-oss4: LNET configured&lt;br/&gt;
root@it-admin:~# dsh -M -g lus04 &quot;lctl list_nids&quot;&lt;br/&gt;
lus04-mds1: 172.17.148.4@tcp&lt;br/&gt;
lus04-mds2: 172.17.148.5@tcp&lt;br/&gt;
lus04-oss1: 172.17.148.6@tcp&lt;br/&gt;
lus04-oss2: 172.17.148.7@tcp&lt;br/&gt;
lus04-oss3: 172.17.148.8@tcp&lt;br/&gt;
lus04-oss4: 172.17.148.9@tcp&lt;/p&gt;

&lt;p&gt;I will start off with the &quot;tunefs.lustre --writeconf &quot; You say MDTs and OSS&apos;s do you want the MGS done as well ?&lt;/p&gt;</comment>
                            <comment id="69342" author="bogl" created="Fri, 18 Oct 2013 21:51:33 +0000"  >&lt;p&gt;yes, please, all devices.  MGS as well.  You have a separate MGS and MDT?  I was just assuming the MGS/MDT was a combo device.&lt;/p&gt;</comment>
                            <comment id="69343" author="james beal" created="Fri, 18 Oct 2013 22:03:51 +0000"  >&lt;p&gt;Commands run and sample output for MGS/MDT/OSS writeconf&lt;/p&gt;

&lt;p&gt;tunefs.lustre --writeconf   --mgs  --failnode 172.17.148.5  /dev/lus04-mgs0/lus04&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: failover.node=172.17.148.5@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: failover.node=172.17.148.5@tcp failover.node=172.17.148.5@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;tunefs.lustre  --writeconf --erase-params  --mgsnode 172.17.148.4@tcp    --mgsnode 172.17.148.5@tcp  --failnode=172.17.148.5@tcp  /dev/lus04-mdt0/lus04&lt;/p&gt;

&lt;p&gt;Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1001&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.5@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1141&lt;br/&gt;
              (MDT update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.5@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd15&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd17&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd19&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd21&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd23&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd25&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd27&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.8 --ost /dev/mapper/vd29&lt;/p&gt;

&lt;p&gt;tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.9 --ost /dev/mapper/vd16&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.9 --ost /dev/mapper/vd18&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.9 --ost /dev/mapper/vd20&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.9 --ost /dev/mapper/vd22&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.9 --ost /dev/mapper/vd24&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.9 --ost /dev/mapper/vd26&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.9 --ost /dev/mapper/vd28&lt;/p&gt;

&lt;p&gt;tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.6 --ost /dev/mapper/vd01&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.6 --ost /dev/mapper/vd03&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.6 --ost /dev/mapper/vd05&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.6 --ost /dev/mapper/vd07&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.6 --ost /dev/mapper/vd09&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.6 --ost /dev/mapper/vd11&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.6 --ost /dev/mapper/vd13&lt;/p&gt;

&lt;p&gt;tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd00&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd02&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd04&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd06&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd08&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd10&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd12&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp --failnode 172.17.148.7 --ost /dev/mapper/vd14&lt;/p&gt;

&lt;p&gt;Sample output&lt;/p&gt;

&lt;p&gt;checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04=OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1102&lt;br/&gt;
              (OST writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.7@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1142&lt;br/&gt;
              (OST update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.8@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;</comment>
                            <comment id="69345" author="bogl" created="Fri, 18 Oct 2013 22:12:33 +0000"  >&lt;p&gt;Possibly need a second opinion, but those params don&apos;t look at all right to me.  I see 2 --mgsnode options + a --failnode option.  Think there should be only 1 --mdsnode + ! --failnode.&lt;/p&gt;

&lt;p&gt;Just for simplification purposes could you try just setting 1 --mgsnode and no --failnode?  I understand this probably isn&apos;t your desired target config, but in the interest of getting the fs back up to possibly copy files off simpler is better.  If we can get that to work we can work on turning failover back on again later.&lt;/p&gt;</comment>
                            <comment id="69346" author="james beal" created="Fri, 18 Oct 2013 22:12:49 +0000"  >&lt;p&gt;oot@lus04-mds1:~# dmesg -c&lt;br/&gt;
root@lus04-mds1:~# mount -t lustre /dev/lus04-mgs0/lus04 /export/MGS&lt;br/&gt;
root@lus04-mds1:~# mount -t lustre /dev/lus04-mdt0/lus04 /export/MDT0&lt;br/&gt;
root@lus04-mds1:~# dmesg&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.319550&amp;#93;&lt;/span&gt; Lustre: Build Version: v1_8_9_WC1sanger1--PRISTINE-2.6.32.59-sles-lustre-1.8.8wc1&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.415542&amp;#93;&lt;/span&gt; Lustre: Lustre Client File System; &lt;a href=&quot;http://www.lustre.org/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://www.lustre.org/&lt;/a&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.424833&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.426341&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.524900&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.526415&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.553022&amp;#93;&lt;/span&gt; Lustre: MGS MGS started&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38471.553767&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38484.993123&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38484.996208&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38485.108837&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38485.111140&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38485.142781&amp;#93;&lt;/span&gt; Lustre: MGS: Logs for fs lus04 were removed by user request.  All servers must be restarted in order to regenerate the logs.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38485.156338&amp;#93;&lt;/span&gt; Lustre: Enabling user_xattr&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38485.170944&amp;#93;&lt;/span&gt; Lustre: lus04-MDT0000: Now serving lus04-MDT0000 on /dev/mapper/lus04--mdt0-lus04 with recovery enabled&lt;br/&gt;
root@lus04-mds1:~# &lt;/p&gt;

&lt;p&gt;And the kernel logfile of the server with the MGS and MDT&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;38634.935325&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST0000 log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38639.574055&amp;#93;&lt;/span&gt; LustreError: 12608:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38639.574234&amp;#93;&lt;/span&gt; LustreError: 12608:0:(obd_config.c:372:class_setup()) setup lus04-OST0000-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38639.574391&amp;#93;&lt;/span&gt; LustreError: 12608:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38639.574563&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0000-osc  1:lus04-OST0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38644.876011&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST0001 log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38648.094895&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST0002 log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38649.236649&amp;#93;&lt;/span&gt; LustreError: 12609:0:(obd_config.c:611:class_add_conn()) try to add conn on immature client dev&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38649.236820&amp;#93;&lt;/span&gt; Lustre: 12609:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0000_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38649.246930&amp;#93;&lt;/span&gt; LustreError: 12609:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0000_UUID not set up&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38649.247075&amp;#93;&lt;/span&gt; LustreError: 12609:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0000_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38649.247213&amp;#93;&lt;/span&gt; LustreError: 12609:0:(obd_config.c:1199:class_config_llog_handler()) Err -22 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38649.247349&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0000_UUID  2:0  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38650.948357&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST0003 log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38655.146204&amp;#93;&lt;/span&gt; LustreError: 12622:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38655.146368&amp;#93;&lt;/span&gt; LustreError: 12622:0:(obd_config.c:372:class_setup()) setup lus04-OST0001-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38655.146503&amp;#93;&lt;/span&gt; LustreError: 12622:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38655.146641&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0001-osc  1:lus04-OST0001_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38655.453621&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST0004 log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38659.318763&amp;#93;&lt;/span&gt; LustreError: 12624:0:(obd_config.c:611:class_add_conn()) try to add conn on immature client dev&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38659.318917&amp;#93;&lt;/span&gt; Lustre: 12624:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0001_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38659.339303&amp;#93;&lt;/span&gt; LustreError: 12624:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0001_UUID not set up&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38659.339447&amp;#93;&lt;/span&gt; LustreError: 12624:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0001_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38659.339588&amp;#93;&lt;/span&gt; LustreError: 12624:0:(obd_config.c:1199:class_config_llog_handler()) Err -22 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38659.339722&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0001_UUID  2:1  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38662.002368&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST0006 log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38662.002372&amp;#93;&lt;/span&gt; Lustre: Skipped 1 previous similar message&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38666.809608&amp;#93;&lt;/span&gt; LustreError: 12633:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38666.809768&amp;#93;&lt;/span&gt; LustreError: 12633:0:(obd_config.c:372:class_setup()) setup lus04-OST0002-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38666.809896&amp;#93;&lt;/span&gt; LustreError: 12633:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38666.810024&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0002-osc  1:lus04-OST0002_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38672.517992&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST0009 log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38672.517996&amp;#93;&lt;/span&gt; Lustre: Skipped 2 previous similar messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.545529&amp;#93;&lt;/span&gt; LustreError: 12634:0:(obd_config.c:611:class_add_conn()) try to add conn on immature client dev&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.545678&amp;#93;&lt;/span&gt; Lustre: 12634:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0002_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.604532&amp;#93;&lt;/span&gt; LustreError: 12635:0:(llog_lvfs.c:612:llog_lvfs_create()) error looking up logfile 0x1e3:0x0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.604684&amp;#93;&lt;/span&gt; LustreError: 12635:0:(llog_cat.c:174:llog_cat_id2handle()) error opening log id 0x1e3:0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.604826&amp;#93;&lt;/span&gt; LustreError: 12635:0:(llog_obd.c:291:cat_cancel_cb()) Cannot find handle for log 0x1e3&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.605968&amp;#93;&lt;/span&gt; LustreError: 12634:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0002_UUID not set up&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.606117&amp;#93;&lt;/span&gt; LustreError: 12634:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0002_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.606258&amp;#93;&lt;/span&gt; LustreError: 12634:0:(obd_config.c:1199:class_config_llog_handler()) Err -22 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38673.606401&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0002_UUID  2:2  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38680.560870&amp;#93;&lt;/span&gt; LustreError: 12638:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38680.561031&amp;#93;&lt;/span&gt; LustreError: 12638:0:(obd_config.c:372:class_setup()) setup lus04-OST0003-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38680.561162&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0003-osc  1:lus04-OST0003_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.703831&amp;#93;&lt;/span&gt; LustreError: 12651:0:(obd_config.c:611:class_add_conn()) try to add conn on immature client dev&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.703987&amp;#93;&lt;/span&gt; Lustre: 12651:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0003_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.748801&amp;#93;&lt;/span&gt; LustreError: 12652:0:(llog_lvfs.c:612:llog_lvfs_create()) error looking up logfile 0x1e8:0x0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.748936&amp;#93;&lt;/span&gt; LustreError: 12652:0:(llog_cat.c:174:llog_cat_id2handle()) error opening log id 0x1e8:0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.749062&amp;#93;&lt;/span&gt; LustreError: 12652:0:(llog_obd.c:291:cat_cancel_cb()) Cannot find handle for log 0x1e8&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.750193&amp;#93;&lt;/span&gt; LustreError: 12651:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0003_UUID not set up&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.750328&amp;#93;&lt;/span&gt; LustreError: 12651:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0003_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.750468&amp;#93;&lt;/span&gt; LustreError: 12651:0:(obd_config.c:1199:class_config_llog_handler()) Err -22 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.750608&amp;#93;&lt;/span&gt; LustreError: 12651:0:(obd_config.c:1199:class_config_llog_handler()) Skipped 1 previous similar message&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38685.750748&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0003_UUID  2:3  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38689.533835&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST000e log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38689.533839&amp;#93;&lt;/span&gt; Lustre: Skipped 4 previous similar messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38693.310511&amp;#93;&lt;/span&gt; LustreError: 12659:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38693.310678&amp;#93;&lt;/span&gt; LustreError: 12659:0:(obd_config.c:372:class_setup()) setup lus04-OST0004-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38693.310824&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0004-osc  1:lus04-OST0004_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38724.566711&amp;#93;&lt;/span&gt; Lustre: MGS: Regenerating lus04-OST000f log by user request.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.657075&amp;#93;&lt;/span&gt; LustreError: 12691:0:(obd_config.c:611:class_add_conn()) try to add conn on immature client dev&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.657250&amp;#93;&lt;/span&gt; Lustre: 12691:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0004_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.697633&amp;#93;&lt;/span&gt; LustreError: 12692:0:(llog_lvfs.c:612:llog_lvfs_create()) error looking up logfile 0x1e7:0x0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.697806&amp;#93;&lt;/span&gt; LustreError: 12692:0:(llog_cat.c:174:llog_cat_id2handle()) error opening log id 0x1e7:0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.697988&amp;#93;&lt;/span&gt; LustreError: 12692:0:(llog_obd.c:291:cat_cancel_cb()) Cannot find handle for log 0x1e7&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.699077&amp;#93;&lt;/span&gt; LustreError: 12691:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0004_UUID not set up&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.699234&amp;#93;&lt;/span&gt; LustreError: 12691:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0004_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.699394&amp;#93;&lt;/span&gt; LustreError: 12691:0:(obd_config.c:1199:class_config_llog_handler()) Err -22 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.699546&amp;#93;&lt;/span&gt; LustreError: 12691:0:(obd_config.c:1199:class_config_llog_handler()) Skipped 1 previous similar message&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38727.699700&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0004_UUID  2:4  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38739.668093&amp;#93;&lt;/span&gt; LustreError: 12693:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38739.668302&amp;#93;&lt;/span&gt; LustreError: 12693:0:(obd_config.c:372:class_setup()) setup lus04-OST0005-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38739.668467&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0005-osc  1:lus04-OST0005_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.892767&amp;#93;&lt;/span&gt; LustreError: 12705:0:(obd_config.c:611:class_add_conn()) try to add conn on immature client dev&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.892918&amp;#93;&lt;/span&gt; Lustre: 12705:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0005_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.936667&amp;#93;&lt;/span&gt; LustreError: 12706:0:(llog_lvfs.c:612:llog_lvfs_create()) error looking up logfile 0x1ec:0x0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.936816&amp;#93;&lt;/span&gt; LustreError: 12706:0:(llog_cat.c:174:llog_cat_id2handle()) error opening log id 0x1ec:0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.936956&amp;#93;&lt;/span&gt; LustreError: 12706:0:(llog_obd.c:291:cat_cancel_cb()) Cannot find handle for log 0x1ec&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.938176&amp;#93;&lt;/span&gt; LustreError: 12705:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0005_UUID not set up&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.938322&amp;#93;&lt;/span&gt; LustreError: 12705:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0005_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38745.938463&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0005_UUID  2:5  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38750.820202&amp;#93;&lt;/span&gt; LustreError: 12707:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38750.820379&amp;#93;&lt;/span&gt; LustreError: 12707:0:(obd_config.c:372:class_setup()) setup lus04-OST0006-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38750.820521&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0006-osc  1:lus04-OST0006_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.362456&amp;#93;&lt;/span&gt; LustreError: 12715:0:(obd_config.c:611:class_add_conn()) try to add conn on immature client dev&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.362609&amp;#93;&lt;/span&gt; Lustre: 12715:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0006_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.402883&amp;#93;&lt;/span&gt; LustreError: 12716:0:(llog_lvfs.c:612:llog_lvfs_create()) error looking up logfile 0x1eb:0x0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.403024&amp;#93;&lt;/span&gt; LustreError: 12716:0:(llog_cat.c:174:llog_cat_id2handle()) error opening log id 0x1eb:0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.403159&amp;#93;&lt;/span&gt; LustreError: 12716:0:(llog_obd.c:291:cat_cancel_cb()) Cannot find handle for log 0x1eb&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.404281&amp;#93;&lt;/span&gt; LustreError: 12715:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0006_UUID not set up&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.404420&amp;#93;&lt;/span&gt; LustreError: 12715:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0006_UUID&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38758.404559&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0006_UUID  2:6  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38762.319547&amp;#93;&lt;/span&gt; LustreError: 12718:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38762.319692&amp;#93;&lt;/span&gt; LustreError: 12718:0:(obd_config.c:1199:class_config_llog_handler()) Skipped 4 previous similar messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38762.319835&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0007-osc  1:lus04-OST0007_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38769.310496&amp;#93;&lt;/span&gt; LustreError: 12721:0:(llog_lvfs.c:612:llog_lvfs_create()) error looking up logfile 0x1f0:0x0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38769.310651&amp;#93;&lt;/span&gt; LustreError: 12721:0:(llog_cat.c:174:llog_cat_id2handle()) error opening log id 0x1f0:0: rc -2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38769.310794&amp;#93;&lt;/span&gt; LustreError: 12721:0:(llog_obd.c:291:cat_cancel_cb()) Cannot find handle for log 0x1f0&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38769.311941&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0007_UUID  2:7  3:1  &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38775.699590&amp;#93;&lt;/span&gt; LustreError: 12735:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38775.699736&amp;#93;&lt;/span&gt; LustreError: 12735:0:(ldlm_lib.c:333:client_obd_setup()) Skipped 1 previous similar message&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38775.699899&amp;#93;&lt;/span&gt; LustreError: 12735:0:(obd_config.c:372:class_setup()) setup lus04-OST0008-osc failed (-2)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38775.700035&amp;#93;&lt;/span&gt; LustreError: 12735:0:(obd_config.c:372:class_setup()) Skipped 1 previous similar message&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;38775.700175&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-OST0008-osc  1:lus04-OST0008_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;/p&gt;

&lt;p&gt;Kernel messages from one of the OSS&apos;s&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.117243&amp;#93;&lt;/span&gt; Lustre: Build Version: v1_8_9_WC1sanger1--PRISTINE-2.6.32.59-sles-lustre-1.8.8wc1&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.215106&amp;#93;&lt;/span&gt; Lustre: Lustre Client File System; &lt;a href=&quot;http://www.lustre.org/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://www.lustre.org/&lt;/a&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.269066&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.290981&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.480698&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.490424&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-14): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.503841&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.561443&amp;#93;&lt;/span&gt; Lustre: Filtering OBD driver; &lt;a href=&quot;http://wiki.whamcloud.com/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://wiki.whamcloud.com/&lt;/a&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42070.592412&amp;#93;&lt;/span&gt; Lustre: lus04-OST0000: Now serving lus04-OST0000 on /dev/mapper/vd00 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42083.497821&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-2): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42083.518965&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-2): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42083.681216&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-2): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42083.690062&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-2): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42083.972670&amp;#93;&lt;/span&gt; Lustre: lus04-OST0002: Now serving lus04-OST0002 on /dev/mapper/vd02 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42090.839376&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-5): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42090.859872&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-5): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42091.040129&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-5): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42091.048959&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-5): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42091.389277&amp;#93;&lt;/span&gt; Lustre: lus04-OST0004: Now serving lus04-OST0004 on /dev/mapper/vd04 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42097.403696&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42097.425964&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42097.588491&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42097.597649&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42098.006214&amp;#93;&lt;/span&gt; Lustre: lus04-OST0006: Now serving lus04-OST0006 on /dev/mapper/vd06 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42104.273626&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-6): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42104.295832&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-6): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42104.452328&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-6): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42104.461200&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-6): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42104.809165&amp;#93;&lt;/span&gt; Lustre: lus04-OST0008: Now serving lus04-OST0008 on /dev/mapper/vd08 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42111.294964&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42111.317643&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42111.483790&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42111.493072&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-1): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42111.836007&amp;#93;&lt;/span&gt; Lustre: lus04-OST000a: Now serving lus04-OST000a on /dev/mapper/vd10 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42117.397322&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-3): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42117.417872&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-3): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42117.577000&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-3): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42117.585901&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-3): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42124.916269&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-12): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42124.940165&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-12): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42125.119575&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-12): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42125.128708&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-12): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42125.516470&amp;#93;&lt;/span&gt; Lustre: lus04-OST000e: Now serving lus04-OST000e on /dev/mapper/vd14 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;42125.516474&amp;#93;&lt;/span&gt; Lustre: Skipped 1 previous similar message&lt;br/&gt;
root@lus04-oss1:~# &lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# lctl dl&lt;br/&gt;
  0 UP mgs MGS MGS 13&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp ec0779df-4709-87c5-fbb3-c1370bccaab5 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
  5 AT osc lus04-OST0000-osc lus04-mdtlov_UUID 1&lt;br/&gt;
  6 AT osc lus04-OST0001-osc lus04-mdtlov_UUID 1&lt;br/&gt;
  7 AT osc lus04-OST0002-osc lus04-mdtlov_UUID 1&lt;br/&gt;
  8 AT osc lus04-OST0003-osc lus04-mdtlov_UUID 1&lt;br/&gt;
  9 AT osc lus04-OST0004-osc lus04-mdtlov_UUID 1&lt;br/&gt;
 10 AT osc lus04-OST0005-osc lus04-mdtlov_UUID 1&lt;br/&gt;
 11 AT osc lus04-OST0006-osc lus04-mdtlov_UUID 1&lt;br/&gt;
 12 AT osc lus04-OST0007-osc lus04-mdtlov_UUID 1&lt;br/&gt;
 13 AT osc lus04-OST0008-osc lus04-mdtlov_UUID 1&lt;/p&gt;</comment>
                            <comment id="69347" author="james beal" created="Fri, 18 Oct 2013 22:25:15 +0000"  >&lt;p&gt;&quot;Possibly need a second opinion, but those params don&apos;t look at all right to me. I see 2 --mgsnode options + a --failnode option. Think there should be only 1 --mdsnode + ! --failnode.&lt;/p&gt;

&lt;p&gt;Just for simplification purposes could you try just setting 1 --mgsnode and no --failnode? I understand this probably isn&apos;t your desired target config, but in the interest of getting the fs back up to possibly copy files off simpler is better. If we can get that to work we can work on turning failover back on again later.&quot;&lt;/p&gt;

&lt;p&gt;Of course, I also missed the --erase-params on the MGS. Right now we only want to get the file system back for a few days so we can copy the data off and reformat it ( so a day or 2 for 240TB ).&lt;/p&gt;</comment>
                            <comment id="69348" author="james beal" created="Fri, 18 Oct 2013 22:32:40 +0000"  >&lt;p&gt;After unmounting all the OSS&apos;s and attempting to unmount the MDT&lt;/p&gt;

&lt;p&gt;Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.645245&amp;#93;&lt;/span&gt; Lustre: 12744:0:(mds_lov.c:1114:mds_notify()) MDS lus04-MDT0000: add target lus04-OST0008_UUID&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.645247&amp;#93;&lt;/span&gt; Lustre: 12744:0:(mds_lov.c:1114:mds_notify()) Skipped 1 previous similar message&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.686343&amp;#93;&lt;/span&gt; LustreError: 12745:0:(llog_lvfs.c:612:llog_lvfs_create()) error looking up logfile 0x1ef:0x0: rc -2&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.686490&amp;#93;&lt;/span&gt; LustreError: 12745:0:(llog_cat.c:174:llog_cat_id2handle()) error opening log id 0x1ef:0: rc -2&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.686625&amp;#93;&lt;/span&gt; LustreError: 12745:0:(llog_obd.c:291:cat_cancel_cb()) Cannot find handle for log 0x1ef&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.687672&amp;#93;&lt;/span&gt; LustreError: 12744:0:(lov_obd.c:289:lov_connect_obd()) Target lus04-OST0008_UUID not set up&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.687816&amp;#93;&lt;/span&gt; LustreError: 12744:0:(lov_obd.c:289:lov_connect_obd()) Skipped 1 previous similar message&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.687952&amp;#93;&lt;/span&gt; LustreError: 12744:0:(lov_obd.c:727:lov_add_target()) connect or notify failed (-22) for lus04-OST0008_UUID&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.688093&amp;#93;&lt;/span&gt; LustreError: 12744:0:(lov_obd.c:727:lov_add_target()) Skipped 1 previous similar message&lt;br/&gt;
Oct 18 23:10:38 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;38784.688235&amp;#93;&lt;/span&gt; Lustre:    cmd=cf00d 0:lus04-mdtlov  1:lus04-OST0008_UUID  2:8  3:1  &lt;br/&gt;
Oct 18 23:26:35 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39740.309228&amp;#93;&lt;/span&gt; Lustre: Failing over lus04-MDT0000&lt;br/&gt;
Oct 18 23:26:35 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39740.310959&amp;#93;&lt;/span&gt; LustreError: 13468:0:(lov_obd.c:1012:lov_cleanup()) lov tgt 0 not cleaned! deathrow=0, lovrc=1&lt;br/&gt;
Oct 18 23:27:05 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39770.254590&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 30 secs.&lt;br/&gt;
Oct 18 23:27:35 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39800.201341&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 60 secs.&lt;br/&gt;
Oct 18 23:28:05 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39830.148112&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 90 secs.&lt;br/&gt;
Oct 18 23:28:35 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39860.094852&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 120 secs.&lt;br/&gt;
Oct 18 23:29:05 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39890.041601&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 150 secs.&lt;br/&gt;
Oct 18 23:29:35 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39919.988359&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 180 secs.&lt;br/&gt;
Oct 18 23:30:05 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39949.935117&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 210 secs.&lt;br/&gt;
Oct 18 23:30:35 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;39979.881872&amp;#93;&lt;/span&gt; Lustre: Mount still busy with 14 refs after 240 secs.&lt;/p&gt;

&lt;p&gt;How should I proceed on the unmount of the MDT&lt;/p&gt;</comment>
                            <comment id="69349" author="bogl" created="Fri, 18 Oct 2013 22:41:50 +0000"  >&lt;p&gt;Try umount -f on the MDT.  If that doesn&apos;t work then just force reboot the server &amp;amp; bring it up again.&lt;/p&gt;

&lt;p&gt;Looking back over the history here I suspect something is wrong with the failover config.&lt;/p&gt;

&lt;p&gt;All the tunefs params show up with no_primnode flags set.  Also show a failnode param set.  Some of the earlier logs suggest that MGS &amp;amp; MDT is attempting to communicate with the failover nid when you are mounting OSTs on the primary server. Think we may get a bit farther without failover for now.&lt;/p&gt;</comment>
                            <comment id="69350" author="james beal" created="Fri, 18 Oct 2013 22:49:18 +0000"  >&lt;p&gt;I am going to be doing that in an instant, but I thought it would be worth mentioning that this filesystem was originally formatted 3+years ago with both possible servers specified as failover nids, not just the alternative host.&lt;/p&gt;

&lt;p&gt;This is from my notes from, note that we had two failnodes configured one of which was the &quot;primary&quot; node.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;15/10/2013 12:22:35&amp;#93;&lt;/span&gt; James Beal: &lt;br/&gt;
root@lus04-oss1:~#  tunefs.lustre   --erase-params  --writeconf --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp  --failnode 172.17.148.6@tcp --failnode 172.17.148.7@tcp --param ost.quota_type=ug   /dev/mapper/vd08&lt;br/&gt;
checking for existing Lustre data: found&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-OST0008&lt;br/&gt;
Index:      8&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x2&lt;br/&gt;
              (OST )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: mgsnode=172.17.148.5@tcp mgsnode=172.17.148.4@tcp failover.node=172.17.148.6@tcp failover.node=172.17.148.7@tcp ost.quota_type=ug2&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04=OST0008&lt;br/&gt;
Index:      8&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x142&lt;br/&gt;
              (OST update writeconf )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.6@tcp failover.node=172.17.148.7@tcp ost.quota_type=ug&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;</comment>
                            <comment id="69351" author="bogl" created="Fri, 18 Oct 2013 23:02:30 +0000"  >&lt;p&gt;hmm. I notice that there is never a --fsname= option in your tunefs cmds, but the LustreFS reported in the params being set is still lus04.  Suggests to me that there is only some partial erasure going on.&lt;/p&gt;

&lt;p&gt;Just to satisfy me could you do a round of tunefs where the only cmd line option is --erase-params then follow that with another round where you do --writeconf with all the desired params, including --fsname=lus04, but no --erase-params?&lt;/p&gt;</comment>
                            <comment id="69353" author="bogl" created="Fri, 18 Oct 2013 23:14:42 +0000"  >&lt;p&gt;It occurs to me to ask are you possibly running the same tunefs cmds for a given device on both the primary node and the backup node where that device is visible?  Think that would have very ill effects.&lt;/p&gt;</comment>
                            <comment id="69354" author="james beal" created="Fri, 18 Oct 2013 23:19:31 +0000"  >&lt;p&gt;&quot;It occurs to me to ask are you possibly running the same tunefs cmds for a given device on both the primary node and the backup node where that device is visible? Think that would have very ill effects.&quot;&lt;/p&gt;

&lt;p&gt;It is always worth asking those questions but no, I am only running the tunefs on the active node, the MGS and MDT I do by hand and the OSS I am keying off the commented mounts in /etc/fstab. I am just about to do this with the OSS(s) now.&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# tunefs.lustre --erase-params /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;br/&gt;
root@lus04-mds1:~# tunefs.lustre --erase-params /dev/lus04-mdt0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1001&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1041&lt;br/&gt;
              (MDT update no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;br/&gt;
root@lus04-mds1:~# tunefs.lustre --writeconf --mgs --fsname=lus04 /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;br/&gt;
root@lus04-mds1:~# tunefs.lustre --writeconf --mgsnode 172.17.148.4@tcp  --fsname=lus04 /dev/lus04-mdt0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1041&lt;br/&gt;
              (MDT update no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1141&lt;br/&gt;
              (MDT update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;br/&gt;
root@lus04-mds1:~# &lt;/p&gt;</comment>
                            <comment id="69355" author="james beal" created="Fri, 18 Oct 2013 23:26:04 +0000"  >&lt;p&gt;And a sample OSS&lt;/p&gt;

&lt;p&gt;tunefs.lustre  --erase-params &lt;/p&gt;

&lt;p&gt;Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1142&lt;br/&gt;
              (OST update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1142&lt;br/&gt;
              (OST update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;


&lt;p&gt;tunefs.lustre  --ost  --writeconf --mgsnode 172.17.148.4@tcp --fsname=lus04 &lt;/p&gt;

&lt;p&gt;Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1142&lt;br/&gt;
              (OST update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1142&lt;br/&gt;
              (OST update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;I havn&apos;t attempted to mount them yet as the unmounting of the MDT does not go well.&lt;/p&gt;</comment>
                            <comment id="69356" author="bogl" created="Fri, 18 Oct 2013 23:39:44 +0000"  >&lt;p&gt;I see all your nodes still have no_primnode flag.  Have been trying to get that set on a local test fs, can&apos;t figure out how.  Looking at the code of mkfs.lustre it seems to be associated with setting a failover, failnode, or servicenode option.  But none of those options are in your recent tunefs commands.  Seems like some things are not getting cleared by erase or writeconf.  I may need to consult wiser experts.&lt;/p&gt;</comment>
                            <comment id="69358" author="james beal" created="Fri, 18 Oct 2013 23:49:54 +0000"  >&lt;p&gt;In the last few days this file system has had  tunefs.lustre run a lot, we have used servicenode, failnode and had it having failover or no failover. I do have a record of what has happened with the file system but it is over 100 pages in word....&lt;/p&gt;

&lt;p&gt;It is now 12:45 in the morning local time for me, I appreciate your efforts. If there is anything more I can do then add it to the ticket and I will try and do it tomorrow family permitting. As I said this system was formatted with say on .4, failnode=.4, failnode=.5 which worked well enough however it feels that fixing that broke things but I have no evidence for that.&lt;/p&gt;</comment>
                            <comment id="69359" author="bogl" created="Fri, 18 Oct 2013 23:55:05 +0000"  >&lt;p&gt;Sorry, I didn&apos;t realize it was so late for you.  by all means quit for now.  I will try to check back on this ticket over the weekend, but can&apos;t promise.  Thanks for your quick response and turnaround on my requests.&lt;/p&gt;</comment>
                            <comment id="69360" author="james beal" created="Fri, 18 Oct 2013 23:58:56 +0000"  >&lt;p&gt;&quot;Sorry, I didn&apos;t realize it was so late for you. by all means quit for now. I will try to check back on this ticket over the weekend, but can&apos;t promise. Thanks for your quick response and turnaround on my requests.&quot;&lt;/p&gt;

&lt;p&gt;I realise that with you in the states that we will not have that much overlap so I am happy to stay up late to get the support. I completely understand about the weekend. Have a good weekend and if anything comes up then put it here and I will attend to it as promptly as I can manage &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;</comment>
                            <comment id="69369" author="james beal" created="Sat, 19 Oct 2013 12:34:08 +0000"  >&lt;p&gt;Part one of try everything on the failnode and see if it all works if a prime node is not defined.&lt;/p&gt;


&lt;p&gt;root@lus04-mds1:~#  tunefs.lustre --writeconf --erase-params --mgs --failnode 172.17.148.5@tcp --fsname=lus04 /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: failover.node=172.17.148.5@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: failover.node=172.17.148.5@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;


&lt;p&gt;root@lus04-mds1:~#  tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mgsnode=172.17.148.5@tcp   --failnode 172.17.148.5@tcp --fsname=lus04 /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: failover.node=172.17.148.5@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.5@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;


&lt;p&gt;Now to mount these on their partner node .5&lt;/p&gt;

&lt;p&gt;Oct 19 13:22:51 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94212.007840&amp;#93;&lt;/span&gt; Lustre: MGS MGS started&lt;br/&gt;
Oct 19 13:22:51 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94212.008444&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 19 13:23:05 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94225.421033&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): barriers disabled&lt;br/&gt;
Oct 19 13:23:05 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94225.424375&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 19 13:23:05 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94225.538080&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): barriers disabled&lt;br/&gt;
Oct 19 13:23:05 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94225.540969&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-0): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 19 13:23:05 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94225.557986&amp;#93;&lt;/span&gt; Lustre: MGS: Logs for fs lus04 were removed by user request.  All servers must be restarted in order to regenerate the logs.&lt;br/&gt;
Oct 19 13:23:05 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94225.574965&amp;#93;&lt;/span&gt; Lustre: Enabling user_xattr&lt;br/&gt;
Oct 19 13:23:05 lus04-mds2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;94225.604777&amp;#93;&lt;/span&gt; Lustre: lus04-MDT0000: Now serving lus04-MDT0000 on /dev/mapper/lus04--mdt0-lus04 with recovery enabled&lt;/p&gt;

&lt;p&gt;tunefs.lustre --ost --writeconf --mgsnode 172.17.148.5@tcp --fsname=lus04 --ost --failnode=172.17.148.7@tcp /dev/mapper/vd00&lt;/p&gt;


&lt;p&gt;And to mount on the partner pair, note &quot; sent from MGC172.17.148.4@tcp to NID 172.17.148.4@tcp 105s ago has timed out&quot;, The MGS is mounted on .5&lt;/p&gt;

&lt;p&gt;Oct 19 13:27:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93559.887640&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): barriers disabled&lt;br/&gt;
Oct 19 13:27:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93559.910142&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 19 13:27:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93560.104185&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): barriers disabled&lt;br/&gt;
Oct 19 13:27:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93560.113434&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 19 13:29:40 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93664.929907&amp;#93;&lt;/span&gt; Lustre: 32416:0:(client.c:1529:ptlrpc_expire_one_request()) @@@ Request x1449276099002414 sent from MGC172.17.148.4@tcp to NID 172.17.148.4@tcp 105s ago has timed out (105s prior to deadline).&lt;br/&gt;
Oct 19 13:29:40 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93664.929910&amp;#93;&lt;/span&gt;   req@ffff8802f5e27c00 x1449276099002414/t0 o250-&amp;gt;MGS@MGC172.17.148.4@tcp_0:26/25 lens 368/584 e 0 to 1 dl 1382185780 ref 1 fl Rpc:N/0/0 rc 0/0&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.895825&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.914641&amp;#93;&lt;/span&gt; LustreError: 2228:0:(llog.c:381:llog_process()) cannot start thread: -513&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.914671&amp;#93;&lt;/span&gt; LustreError: 2228:0:(mgc_request.c:1094:mgc_copy_llog()) Failed to copy remote log lus04-OST0000 (-513)&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.944136&amp;#93;&lt;/span&gt; LustreError: 2228:0:(llog.c:381:llog_process()) cannot start thread: -513&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.944173&amp;#93;&lt;/span&gt; LustreError: 15c-8: MGC172.17.148.4@tcp: The configuration from log &apos;lus04-OST0000&apos; failed (-513). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.944224&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:1143:server_start_targets()) failed to start server lus04-OST0000: -513&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.944253&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:1672:server_fill_super()) Unable to start targets: -513&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.944381&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:1455:server_put_super()) no obd lus04-OST0000&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.944531&amp;#93;&lt;/span&gt; LustreError: 2228:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93684.944662&amp;#93;&lt;/span&gt; LustreError: 2228:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93685.048913&amp;#93;&lt;/span&gt; Lustre: server umount lus04-OST0000 complete&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93685.048919&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-513)&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93685.076179&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): barriers disabled&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93685.085465&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93685.194214&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): barriers disabled&lt;br/&gt;
Oct 19 13:30:00 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93685.203402&amp;#93;&lt;/span&gt; LDISKFS-fs (dm-13): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 19 13:30:04 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93689.375697&amp;#93;&lt;/span&gt; LustreError: 137-5: UUID &apos;lus04-OST0000_UUID&apos; is not available for connect (no target)&lt;br/&gt;
Oct 19 13:30:04 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93689.375831&amp;#93;&lt;/span&gt; LustreError: 2735:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (&lt;del&gt;19)  req@ffff880614092000 x1449326212546592/t0 o8&lt;/del&gt;&amp;gt;&amp;lt;?&amp;gt;@&amp;lt;?&amp;gt;:0/0 lens 368/0 e 0 to 0 dl 1382185904 ref 1 fl Interpret:/0/0 rc -19/0&lt;br/&gt;
Oct 19 13:31:45 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93790.023923&amp;#93;&lt;/span&gt; Lustre: 32416:0:(client.c:1529:ptlrpc_expire_one_request()) @@@ Request x1449276099002422 sent from MGC172.17.148.4@tcp to NID 172.17.148.4@tcp 105s ago has timed out (105s prior to deadline).&lt;br/&gt;
Oct 19 13:31:45 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93790.023926&amp;#93;&lt;/span&gt;   req@ffff8802f5a62800 x1449276099002422/t0 o250-&amp;gt;MGS@MGC172.17.148.4@tcp_0:26/25 lens 368/584 e 0 to 1 dl 1382185905 ref 1 fl Rpc:N/0/0 rc 0/0&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.663025&amp;#93;&lt;/span&gt; LustreError: 2228:0:(mgc_request.c:365:mgc_requeue_add()) log lus04-OST0000: cannot start requeue thread (-513),no more log updates!&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.663175&amp;#93;&lt;/span&gt; LustreError: 2228:0:(mgc_request.c:638:mgc_blocking_ast()) cancel CB failed -513:&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.663308&amp;#93;&lt;/span&gt; LustreError: 2228:0:(mgc_request.c:639:mgc_blocking_ast()) ### MGC ast ns: MGC172.17.148.4@tcp lock: ffff8802a8324e00/0x85ea67077e7a36d8 lrc: 5/0,0 mode: --/CR res: 224151172460/0 rrc: 1 type: PLN flags: 0x4002c90 remote: 0x0 expref: -99 pid: 2228 timeout 0&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.664025&amp;#93;&lt;/span&gt; LustreError: 2228:0:(llog.c:381:llog_process()) cannot start thread: -513&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.664164&amp;#93;&lt;/span&gt; LustreError: 15c-8: MGC172.17.148.4@tcp: The configuration from log &apos;lus04-OST0000&apos; failed (-513). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.664426&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:1143:server_start_targets()) failed to start server lus04-OST0000: -513&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.664560&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:1672:server_fill_super()) Unable to start targets: -513&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.664741&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:1455:server_put_super()) no obd lus04-OST0000&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.801800&amp;#93;&lt;/span&gt; Lustre: server umount lus04-OST0000 complete&lt;br/&gt;
Oct 19 13:31:55 lus04-oss2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;93799.801806&amp;#93;&lt;/span&gt; LustreError: 2228:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-513)&lt;/p&gt;</comment>
                            <comment id="69370" author="james beal" created="Sat, 19 Oct 2013 12:51:08 +0000"  >&lt;p&gt;root@lus04-mds1:~#  tunefs.lustre --writeconf --erase-params --mgs  --fsname=lus04 /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.5@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;At this point I made a mistake, I wrote the MDT config into the MGS by accident.&lt;/p&gt;


&lt;p&gt;root@lus04-mds1:~#  tunefs.lustre --writeconf --erase-params --mgsnode=172.17.148.4@tcp --mdt  --fsname=lus04 /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1144&lt;br/&gt;
              (MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDTffff&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1145&lt;br/&gt;
              (MDT MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;I tried to reverse this.&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~#   tunefs.lustre --writeconf --erase-params  /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1145&lt;br/&gt;
              (MDT MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1145&lt;br/&gt;
              (MDT MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;I now have two mistakes I need to fix in the flags.&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~#   tunefs.lustre --writeconf --erase-params  --mgs  --fsname=lus04   /dev/lus04-mgs0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1145&lt;br/&gt;
              (MDT MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1145&lt;br/&gt;
              (MDT MGS update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;I will copy a backup I did of the MGS ( on thursday ) over to the machine locally and attempt to mount that as a loop device.&lt;/p&gt;
</comment>
                            <comment id="69371" author="james beal" created="Sat, 19 Oct 2013 15:55:55 +0000"  >&lt;p&gt;I can now get the OSS to mount and it looks ok however a client will not mount&lt;/p&gt;

&lt;p&gt;Start with a clean MGS&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:/#  dd if=/dev/zero bs=1M count=128 of=/home/MGS_DONOT_DELETE&lt;br/&gt;
128+0 records in&lt;br/&gt;
128+0 records out&lt;br/&gt;
134217728 bytes (134 MB) copied, 0.0937643 s, 1.4 GB/s&lt;br/&gt;
root@lus04-mds1:/# losetup /dev/loop2 /home/MGS_DONOT_DELETE&lt;br/&gt;
root@lus04-mds1:/# mkfs.lustre --mgs --fsname=lus04 /dev/loop2&lt;/p&gt;

&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x74&lt;br/&gt;
              (MGS needs_index first_time update )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;

&lt;p&gt;checking for existing Lustre data: not found&lt;br/&gt;
device size = 128MB&lt;br/&gt;
formatting backing filesystem ldiskfs on /dev/loop2&lt;br/&gt;
	target name  MGS&lt;br/&gt;
	4k blocks     32768&lt;br/&gt;
	options        -q -O uninit_bg,dir_nlink -E lazy_journal_init -F&lt;br/&gt;
mkfs_cmd = mke2fs -j -b 4096 -L MGS  -q -O uninit_bg,dir_nlink -E lazy_journal_init -F /dev/loop2 32768&lt;br/&gt;
Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# mount -t lustre /dev/loop2 /export/MGS&lt;/p&gt;



&lt;p&gt;And the MDT&lt;/p&gt;

&lt;p&gt;tunefs.lustre /dev/lus04-mdt0/lus04 &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1001&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1001&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;


&lt;p&gt;Do all the OSS mounts on the partners and look at the devices&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:/root# lctl dl&lt;br/&gt;
  0 UP mgs MGS MGS 5&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp 5024ddc3-8729-483f-148c-5bbfe6326be2 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
  5 UP osc lus04-OST0000-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  6 UP osc lus04-OST0001-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  7 UP osc lus04-OST0002-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  8 UP osc lus04-OST0003-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  9 UP osc lus04-OST0004-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 10 UP osc lus04-OST0005-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 11 UP osc lus04-OST0006-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 12 UP osc lus04-OST0007-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 13 UP osc lus04-OST0008-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 14 UP osc lus04-OST0009-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 15 UP osc lus04-OST000a-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 16 UP osc lus04-OST000b-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 17 UP osc lus04-OST000c-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 18 UP osc lus04-OST000d-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 19 UP osc lus04-OST000e-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 20 UP osc lus04-OST000f-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 21 UP osc lus04-OST0010-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 22 UP osc lus04-OST0011-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 23 UP osc lus04-OST0012-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 24 UP osc lus04-OST0013-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 25 UP osc lus04-OST0014-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 26 UP osc lus04-OST0015-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 27 UP osc lus04-OST0016-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 28 UP osc lus04-OST0017-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 29 UP osc lus04-OST0018-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 30 UP osc lus04-OST0019-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 31 UP osc lus04-OST001a-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 32 UP osc lus04-OST001b-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 33 UP osc lus04-OST001c-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 34 UP osc lus04-OST001d-osc lus04-mdtlov_UUID 5&lt;/p&gt;

&lt;p&gt;See that a client mount fails&lt;/p&gt;

&lt;p&gt;mount /lustre/scratch104&lt;br/&gt;
mount.lustre: mount lus04-mds1@tcp0:/lus04 at /lustre/scratch104 failed: No such file or directory&lt;br/&gt;
Is the MGS specification correct?&lt;br/&gt;
Is the filesystem name correct?&lt;br/&gt;
If upgrading, is the copied client log valid? (see upgrade docs)&lt;br/&gt;
root@bc-11-2-07:~# tail -f /var/log/kern.log&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.814834&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-MDT0000-mdc  1:lus04-MDT0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.815018&amp;#93;&lt;/span&gt; LustreError: 15c-8: MGC172.17.148.4@tcp: The configuration from log &apos;lus04-client&apos; failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.815345&amp;#93;&lt;/span&gt; LustreError: 11016:0:(llite_lib.c:1099:ll_fill_super()) Unable to process log: -2&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.815751&amp;#93;&lt;/span&gt; LustreError: 11016:0:(obd_config.c:443:class_cleanup()) Device 68 not setup&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.816019&amp;#93;&lt;/span&gt; LustreError: 11016:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.816192&amp;#93;&lt;/span&gt; LustreError: 11016:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Skipped 1 previous similar message&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.816344&amp;#93;&lt;/span&gt; LustreError: 11016:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.816497&amp;#93;&lt;/span&gt; LustreError: 11016:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) Skipped 1 previous similar message&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.817443&amp;#93;&lt;/span&gt; Lustre: client lus04-client(ffff88015c402800) umount complete&lt;br/&gt;
Oct 19 16:26:29 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17774424.817530&amp;#93;&lt;/span&gt; LustreError: 11016:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-2)&lt;/p&gt;


&lt;p&gt;Clean out the client config from the MGS&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:/# umount /export/MGS&lt;br/&gt;
root@lus04-mds1:/# mount -t ldiskfs /dev/loop1 /export/MGS&lt;br/&gt;
root@lus04-mds1:/# cd /export/MGS/CONFIGS&lt;br/&gt;
 root@lus04-mds1:/export/MGS/CONFIGS# mkdir ../CONFIGS_OLD1&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# mv lus04-client ../CONFIGS_OLD1&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# cd /&lt;br/&gt;
root@lus04-mds1:/# umount /export/MGS&lt;/p&gt;

&lt;p&gt;Now tunefs the OSS so that they believe they are owned by one server, but have a failover mid&lt;/p&gt;

&lt;p&gt;root@lus04-oss1:~# for i in `seq -w 00 07 ` ; do  tunefs.lustre --ost --writeconf --erase-params --mgsnode 172.17.148.4@tcp --fsname=lus04 --ost --failnode=172.17.148.7@tcp /dev/mapper/vd$i  ; done&lt;/p&gt;

&lt;p&gt;Now mount them on the partner node&lt;/p&gt;

&lt;p&gt;root@lus04-oss2:~# for i in `seq -w 00 07 ` ; do  mount -t lustre /dev/mapper/vd$i /export/vd$i; done&lt;/p&gt;

&lt;p&gt;See that the MGS/MDT looks ok&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# lctl dl&lt;br/&gt;
  0 UP mgs MGS MGS 13&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp 75489c46-b5a0-50ff-86e9-3f688e8a1de8 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
  5 UP osc lus04-OST0000-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  6 UP osc lus04-OST0001-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  7 UP osc lus04-OST0002-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  8 UP osc lus04-OST0003-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  9 UP osc lus04-OST0004-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 10 UP osc lus04-OST0005-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 11 UP osc lus04-OST0006-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 12 UP osc lus04-OST0007-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 13 UP osc lus04-OST0008-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 14 UP osc lus04-OST0009-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 15 UP osc lus04-OST000a-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 16 UP osc lus04-OST000b-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 17 UP osc lus04-OST000c-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 18 UP osc lus04-OST000d-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 19 UP osc lus04-OST000e-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 20 UP osc lus04-OST000f-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 21 UP osc lus04-OST0010-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 22 UP osc lus04-OST0011-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 23 UP osc lus04-OST0012-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 24 UP osc lus04-OST0013-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 25 UP osc lus04-OST0014-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 26 UP osc lus04-OST0015-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 27 UP osc lus04-OST0016-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 28 UP osc lus04-OST0017-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 29 UP osc lus04-OST0018-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 30 UP osc lus04-OST0019-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 31 UP osc lus04-OST001a-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 32 UP osc lus04-OST001b-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 33 UP osc lus04-OST001c-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 34 UP osc lus04-OST001d-osc lus04-mdtlov_UUID 5&lt;br/&gt;
root@lus04-mds1:~# &lt;/p&gt;

&lt;p&gt;Try and mount the client and get frustrated.&lt;br/&gt;
Oct 19 16:49:04 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775779.740161&amp;#93;&lt;/span&gt; Lustre: Removed LNI 172.17.115.32@tcp&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.590501&amp;#93;&lt;/span&gt; Lustre: Build Version: v1_8_9_WC1sanger1--PRISTINE-2.6.32.59-sles-lustre-1.8.8wc1&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.662391&amp;#93;&lt;/span&gt; Lustre: Added LNI 172.17.115.32@tcp &lt;span class=&quot;error&quot;&gt;&amp;#91;8/256/0/180&amp;#93;&lt;/span&gt;&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.662573&amp;#93;&lt;/span&gt; Lustre: Accept secure, port 988&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.752886&amp;#93;&lt;/span&gt; Lustre: Lustre Client File System; &lt;a href=&quot;http://www.lustre.org/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://www.lustre.org/&lt;/a&gt;&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.773139&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.775766&amp;#93;&lt;/span&gt; LustreError: 11961:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.775996&amp;#93;&lt;/span&gt; LustreError: 11961:0:(obd_config.c:372:class_setup()) setup lus04-MDT0000-mdc-ffff88041200d000 failed (-2)&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.776149&amp;#93;&lt;/span&gt; LustreError: 11961:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.776301&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-MDT0000-mdc  1:lus04-MDT0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.776486&amp;#93;&lt;/span&gt; LustreError: 15c-8: MGC172.17.148.4@tcp: The configuration from log &apos;lus04-client&apos; failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.776819&amp;#93;&lt;/span&gt; LustreError: 11891:0:(llite_lib.c:1099:ll_fill_super()) Unable to process log: -2&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.777235&amp;#93;&lt;/span&gt; LustreError: 11891:0:(obd_config.c:443:class_cleanup()) Device 2 not setup&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.777504&amp;#93;&lt;/span&gt; LustreError: 11891:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.777659&amp;#93;&lt;/span&gt; LustreError: 11891:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.786546&amp;#93;&lt;/span&gt; Lustre: client lus04-client(ffff88041200d000) umount complete&lt;br/&gt;
Oct 19 16:49:12 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775787.786635&amp;#93;&lt;/span&gt; LustreError: 11891:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-2)&lt;/p&gt;

&lt;p&gt;Try and reset the lus04-client log&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# umount /export/MGS&lt;br/&gt;
root@lus04-mds1:~# mount -t ldiskfs /dev/loop2 /export/MGS&lt;br/&gt;
root@lus04-mds1:~# cd /export/MGS/CONFIGS/&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# mkdir ../CONFIGS1&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# mv lus04-client ../CONFIGS1&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# cd /&lt;br/&gt;
root@lus04-mds1:/# umount /export/MGS&lt;br/&gt;
root@lus04-mds1:/# mount -t lustre /dev/loop2 /export/MGS&lt;/p&gt;

&lt;p&gt;Mount a client&lt;/p&gt;

&lt;p&gt;Oct 19 16:51:43 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775938.418669&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 19 16:51:43 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775938.421159&amp;#93;&lt;/span&gt; LustreError: 156-2: The client profile &apos;lus04-client&apos; could not be read from the MGS.  Does that filesystem exist?&lt;br/&gt;
Oct 19 16:51:43 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775938.421516&amp;#93;&lt;/span&gt; LustreError: 11971:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
Oct 19 16:51:43 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775938.421688&amp;#93;&lt;/span&gt; LustreError: 11971:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
Oct 19 16:51:43 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775938.422630&amp;#93;&lt;/span&gt; Lustre: client lus04-client(ffff8801c6b9d400) umount complete&lt;br/&gt;
Oct 19 16:51:43 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775938.422713&amp;#93;&lt;/span&gt; LustreError: 11971:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-22)&lt;/p&gt;

&lt;p&gt;Reintroduce the MDT &lt;/p&gt;

&lt;p&gt;root@lus04-mds1:/# umount /export/MDT0&lt;br/&gt;
root@lus04-mds1:/# mount -t lustre /dev/lus04-mdt0/lus04 /export/MDT0&lt;/p&gt;

&lt;p&gt;And try the client mount again.&lt;/p&gt;

&lt;p&gt;Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.316139&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.319088&amp;#93;&lt;/span&gt; LustreError: 11998:0:(ldlm_lib.c:333:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.319323&amp;#93;&lt;/span&gt; LustreError: 11998:0:(obd_config.c:372:class_setup()) setup lus04-MDT0000-mdc-ffff88040f93fc00 failed (-2)&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.319477&amp;#93;&lt;/span&gt; LustreError: 11998:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.319628&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-MDT0000-mdc  1:lus04-MDT0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.319815&amp;#93;&lt;/span&gt; LustreError: 15c-8: MGC172.17.148.4@tcp: The configuration from log &apos;lus04-client&apos; failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.320147&amp;#93;&lt;/span&gt; LustreError: 11988:0:(llite_lib.c:1099:ll_fill_super()) Unable to process log: -2&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.320558&amp;#93;&lt;/span&gt; LustreError: 11988:0:(obd_config.c:443:class_cleanup()) Device 2 not setup&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.320812&amp;#93;&lt;/span&gt; LustreError: 11988:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
Oct 19 16:52:23 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.320983&amp;#93;&lt;/span&gt; LustreError: 11988:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
Oct 19 16:52:24 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.329645&amp;#93;&lt;/span&gt; Lustre: client lus04-client(ffff88040f93fc00) umount complete&lt;br/&gt;
Oct 19 16:52:24 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17775979.329734&amp;#93;&lt;/span&gt; LustreError: 11988:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-2)&lt;/p&gt;</comment>
                            <comment id="69376" author="james beal" created="Sun, 20 Oct 2013 07:25:28 +0000"  >
&lt;p&gt;losetup /dev/loop3 /nfs/ssg_data01/jb23/lus04-mdt_in_use_talk_to_me_first &lt;br/&gt;
tune2fs -O ^quota /dev/loop3 &lt;br/&gt;
tune2fs 1.42.7.wc1 (12-Apr-2013)&lt;/p&gt;

&lt;p&gt;tunefs.lustre --dryrun /dev/loop3&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04=MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1101&lt;br/&gt;
              (MDT writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.4@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1101&lt;br/&gt;
              (MDT writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.4@tcp&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;

&lt;p&gt;tunefs.lustre --writeconf --erase-params --mgsnode 172.17.148.4@tcp --fsname=lus04 /dev/loop3&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04=MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1101&lt;br/&gt;
              (MDT writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.4@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1141&lt;br/&gt;
              (MDT update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;



&lt;p&gt;Unmount the OSS&apos;s, and reset the config,.&lt;/p&gt;

&lt;p&gt;root@lus04-oss1:~#  for i in `seq -w 08 14` ; do  tunefs.lustre --ost --writeconf --erase-params --mgsnode 172.17.148.4@tcp --fsname=lus04 --ost --failnode=172.17.148.7@tcp /dev/mapper/vd$i  ; done&lt;br/&gt;
root@lus04-oss2:~# for i in `seq -w 00 07 ` ; do  tunefs.lustre --ost --writeconf --erase-params --mgsnode 172.17.148.4@tcp --fsname=lus04 --ost --failnode=172.17.148.6@tcp /dev/mapper/vd$i  ; done&lt;br/&gt;
root@lus04-oss3:~# for i in `seq -w 15 22` ; do  tunefs.lustre --ost --writeconf --erase-params --mgsnode 172.17.148.4@tcp --fsname=lus04 --ost --failnode=172.17.148.9@tcp /dev/mapper/vd$i  ; done&lt;br/&gt;
root@lus04-oss4:~#  for i in `seq -w 23 29` ; do  tunefs.lustre --ost --writeconf --erase-params --mgsnode 172.17.148.4@tcp --fsname=lus04 --ost --failnode=172.17.148.8@tcp /dev/mapper/vd$i  ; done&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# mount | grep lus&lt;br/&gt;
/dev/loop2 on /export/MGS type lustre (rw)&lt;br/&gt;
root@lus04-mds1:~# mount -t lustre /dev/loop3 /export/MDT0&lt;br/&gt;
root@lus04-mds1:~# tail /var/log/kern.log&lt;br/&gt;
Oct 20 08:09:44 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66192.187279&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): barriers disabled&lt;br/&gt;
Oct 20 08:09:44 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66192.214036&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.710174&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): barriers disabled&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.712679&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.790005&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): barriers disabled&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.791717&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.803940&amp;#93;&lt;/span&gt; Lustre: MGS: Logs for fs lus04 were removed by user request.  All servers must be restarted in order to regenerate the logs.&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.806465&amp;#93;&lt;/span&gt; Lustre: Enabling user_xattr&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.889602&amp;#93;&lt;/span&gt; Lustre: Mounting lus04-MDT0000 at first time on 2.0 FS, remove all clients for interop needs&lt;br/&gt;
Oct 20 08:10:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66232.941098&amp;#93;&lt;/span&gt; Lustre: lus04-MDT0000: Now serving lus04-MDT0000 on /dev/loop3 with recovery enabled&lt;/p&gt;

&lt;p&gt;root@lus04-oss1:~# for i in `seq -w 00 07` ; do mount -t lustre /dev/mapper/vd$i /export/vd$i ; done&lt;br/&gt;
root@lus04-oss2:~# for i in `seq -w 08 14` ; do mount -t lustre /dev/mapper/vd$i /export/vd$i ; done&lt;br/&gt;
root@lus04-oss4:~# for i in `seq -w 15 22` ; do mount -t lustre /dev/mapper/vd$i /export/vd$i ; done&lt;br/&gt;
root@lus04-oss3:~# for i in `seq -w 23 29` ; do mount -t lustre /dev/mapper/vd$i /export/vd$i ; done&lt;/p&gt;

&lt;p&gt;Wait for a bit&lt;br/&gt;
grep -i statu /proc/fs/lustre/obdfilter/lus04-OST*/recovery_status&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0017/recovery_status:status: RECOVERING&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0018/recovery_status:status: RECOVERING&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0019/recovery_status:status: RECOVERING&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST001a/recovery_status:status: RECOVERING&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST001b/recovery_status:status: RECOVERING&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST001c/recovery_status:status: RECOVERING&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST001d/recovery_status:status: RECOVERING&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# lctl dl&lt;br/&gt;
  0 UP mgs MGS MGS 13&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp f450e477-6ad3-c2d7-72f2-ace4ad6d6513 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
  5 UP osc lus04-OST0000-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  6 UP osc lus04-OST0001-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  7 UP osc lus04-OST0002-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  8 UP osc lus04-OST0003-osc lus04-mdtlov_UUID 5&lt;br/&gt;
  9 UP osc lus04-OST0004-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 10 UP osc lus04-OST0005-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 11 UP osc lus04-OST0006-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 12 UP osc lus04-OST0007-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 13 UP osc lus04-OST0008-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 14 UP osc lus04-OST0009-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 15 UP osc lus04-OST000a-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 16 UP osc lus04-OST000b-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 17 UP osc lus04-OST000c-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 18 UP osc lus04-OST000d-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 19 UP osc lus04-OST000e-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 20 UP osc lus04-OST000f-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 21 UP osc lus04-OST0010-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 22 UP osc lus04-OST0011-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 23 UP osc lus04-OST0012-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 24 UP osc lus04-OST0013-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 25 UP osc lus04-OST0014-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 26 UP osc lus04-OST0015-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 27 UP osc lus04-OST0016-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 28 UP osc lus04-OST0017-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 29 UP osc lus04-OST0018-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 30 UP osc lus04-OST0019-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 31 UP osc lus04-OST001a-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 32 UP osc lus04-OST001b-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 33 UP osc lus04-OST001c-osc lus04-mdtlov_UUID 5&lt;br/&gt;
 34 UP osc lus04-OST001d-osc lus04-mdtlov_UUID 5&lt;br/&gt;
root@lus04-mds1:~# &lt;/p&gt;

&lt;p&gt;Mounting client fails again.&lt;/p&gt;

&lt;p&gt;Try and reset the client log&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:~# losetup -a&lt;br/&gt;
/dev/loop0: &lt;span class=&quot;error&quot;&gt;&amp;#91;001c&amp;#93;&lt;/span&gt;:5830 (/mnt/users/jb23/lus04-mgs_in_use)&lt;br/&gt;
/dev/loop1: &lt;span class=&quot;error&quot;&gt;&amp;#91;001c&amp;#93;&lt;/span&gt;:5830 (/mnt/users/jb23/lus04-mgs_in_use)&lt;br/&gt;
/dev/loop2: &lt;span class=&quot;error&quot;&gt;&amp;#91;0846&amp;#93;&lt;/span&gt;:134525730 (/home/MGS_DONOT_DELETE)&lt;br/&gt;
/dev/loop3: &lt;span class=&quot;error&quot;&gt;&amp;#91;0019&amp;#93;&lt;/span&gt;:10440828 (/nfs/ssg_data01/jb23/lus04-mdt_in_use_talk_to_me_first)&lt;br/&gt;
root@lus04-mds1:~# mount -t ldiskfs /dev/loop3 /export/MDT0&lt;br/&gt;
root@lus04-mds1:~# mount -t ldiskfs /dev/loop2 /export/MGS&lt;br/&gt;
root@lus04-mds1:~# cd /export/MDT0/&lt;br/&gt;
CATALOGS           lost+found/        oi.16.10           oi.16.20           oi.16.30           oi.16.40           oi.16.50           oi.16.60           REMOTE_PARENT_DIR/&lt;br/&gt;
changelog_catalog  lov_objid          oi.16.11           oi.16.21           oi.16.31           oi.16.41           oi.16.51           oi.16.61           ROOT/&lt;br/&gt;
changelog_users    lov_objseq         oi.16.12           oi.16.22           oi.16.32           oi.16.42           oi.16.52           oi.16.62           seq_ctl&lt;br/&gt;
CONFIGS/           lquota.group       oi.16.13           oi.16.23           oi.16.33           oi.16.43           oi.16.53           oi.16.63           seq_srv&lt;br/&gt;
CONFIGS_OLD/       lquota.user        oi.16.14           oi.16.24           oi.16.34           oi.16.44           oi.16.54           oi.16.7            &lt;br/&gt;
fld                lquota_v2.group    oi.16.15           oi.16.25           oi.16.35           oi.16.45           oi.16.55           oi.16.8            &lt;br/&gt;
health_check       lquota_v2.user     oi.16.16           oi.16.26           oi.16.36           oi.16.46           oi.16.56           oi.16.9            &lt;br/&gt;
last_rcvd          O/                 oi.16.17           oi.16.27           oi.16.37           oi.16.47           oi.16.57           OI_scrub           &lt;br/&gt;
lfsck_bookmark     OBJECTS/           oi.16.18           oi.16.28           oi.16.38           oi.16.48           oi.16.58           PENDING/           &lt;br/&gt;
lfsck_namespace    oi.16.0            oi.16.19           oi.16.29           oi.16.39           oi.16.49           oi.16.59           quota_master/      &lt;br/&gt;
LOGS/              oi.16.1            oi.16.2            oi.16.3            oi.16.4            oi.16.5            oi.16.6            quota_slave/       &lt;br/&gt;
root@lus04-mds1:~# cd /export/MDT0/CONFIGS&lt;br/&gt;
root@lus04-mds1:/export/MDT0/CONFIGS# mkdir ../C&lt;br/&gt;
CATALOGS     CONFIGS/     CONFIGS_OLD/ &lt;br/&gt;
root@lus04-mds1:/export/MDT0/CONFIGS# mkdir ../CONFIGS_OLD1&lt;br/&gt;
root@lus04-mds1:/export/MDT0/CONFIGS# mv lus04-client ../CONFIGS_OLD1&lt;br/&gt;
root@lus04-mds1:/export/MDT0/CONFIGS# cd /&lt;br/&gt;
root@lus04-mds1:/# umount /export/MDT0&lt;br/&gt;
root@lus04-mds1:/# cd /export/MGS/CONFIGS/&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# mkdir ../CONFIGS1&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# mv lus04-client ../CONFIGS1&lt;br/&gt;
root@lus04-mds1:/export/MGS/CONFIGS# cd /&lt;br/&gt;
root@lus04-mds1:/# umount /export/MGS&lt;br/&gt;
root@lus04-mds1:/# lustre_rmmod &lt;br/&gt;
fsfilt_ldiskfs         75227  0 &lt;br/&gt;
obdclass              582304  10 mds,fsfilt_ldiskfs,mgs,mgc,lustre,lov,mdc,lquota,osc,ptlrpc&lt;br/&gt;
lvfs                   43190  12 mds,fsfilt_ldiskfs,mgs,mgc,lustre,lov,mdc,lquota,osc,ptlrpc,obdclass&lt;br/&gt;
libcfs                248201  14 mds,fsfilt_ldiskfs,mgs,mgc,lustre,lov,mdc,lquota,osc,ksocklnd,ptlrpc,obdclass,lnet,lvfs&lt;br/&gt;
ldiskfs               291319  1 fsfilt_ldiskfs&lt;br/&gt;
mbcache                 8134  2 ldiskfs,ext4&lt;br/&gt;
jbd2                   63282  3 fsfilt_ldiskfs,ldiskfs,ext4&lt;br/&gt;
crc16                   1659  2 ldiskfs,ext4&lt;br/&gt;
root@lus04-mds1:/# lustre_rmmod &lt;br/&gt;
open /proc/sys/lnet/dump_kernel failed: No such file or directory&lt;br/&gt;
open(dump_kernel) failed: No such file or directory&lt;/p&gt;

&lt;p&gt;root@lus04-mds1:/# mount -t lustre /dev/loop2 /export/MGS&lt;br/&gt;
root@lus04-mds1:/# mount -t lustre /dev/loop3 /export/MDT0&lt;br/&gt;
root@lus04-mds1:/# tail /var/log/kern.log&lt;br/&gt;
Oct 20 08:22:09 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66935.739538&amp;#93;&lt;/span&gt; Lustre: MGS MGS started&lt;br/&gt;
Oct 20 08:22:09 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66935.740321&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 20 08:22:15 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.327177&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): barriers disabled&lt;br/&gt;
Oct 20 08:22:15 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.342193&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 20 08:22:16 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.480158&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): barriers disabled&lt;br/&gt;
Oct 20 08:22:16 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.481802&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): mounted filesystem with ordered data mode&lt;br/&gt;
Oct 20 08:22:16 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.511948&amp;#93;&lt;/span&gt; LustreError: 13c-e: Client log lus04-client has disappeared! Regenerating all logs.&lt;br/&gt;
Oct 20 08:22:16 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.514410&amp;#93;&lt;/span&gt; Lustre: MGS: Logs for fs lus04 were removed by user request.  All servers must be restarted in order to regenerate the logs.&lt;br/&gt;
Oct 20 08:22:16 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.516976&amp;#93;&lt;/span&gt; Lustre: Enabling user_xattr&lt;br/&gt;
Oct 20 08:22:16 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;66942.579581&amp;#93;&lt;/span&gt; Lustre: lus04-MDT0000: Now serving lus04-MDT0000 on /dev/loop3 with recovery enabled&lt;/p&gt;

&lt;p&gt;Tune the OSS and remount them&lt;/p&gt;

&lt;p&gt;Mount the client still fails.&lt;/p&gt;

&lt;p&gt;mount /lustre/scratch104&lt;br/&gt;
mount.lustre: mount lus04-mds1@tcp0:/lus04 at /lustre/scratch104 failed: No such file or directory&lt;br/&gt;
Is the MGS specification correct?&lt;br/&gt;
Is the filesystem name correct?&lt;br/&gt;
If upgrading, is the copied client log valid? (see upgrade docs)&lt;br/&gt;
root@bc-11-2-02:~# tail /var/log/kern.log&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.917419&amp;#93;&lt;/span&gt; LustreError: 23097:0:(obd_config.c:372:class_setup()) setup lus04-MDT0000-mdc-ffff88041f2c3400 failed (-2)&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.917572&amp;#93;&lt;/span&gt; LustreError: 23097:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.917742&amp;#93;&lt;/span&gt; Lustre:    cmd=cf003 0:lus04-MDT0000-mdc  1:lus04-MDT0000_UUID  2:0@&amp;lt;0:0&amp;gt;  &lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.917930&amp;#93;&lt;/span&gt; LustreError: 15c-8: MGC172.17.148.4@tcp: The configuration from log &apos;lus04-client&apos; failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.918272&amp;#93;&lt;/span&gt; LustreError: 23027:0:(llite_lib.c:1099:ll_fill_super()) Unable to process log: -2&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.918686&amp;#93;&lt;/span&gt; LustreError: 23027:0:(obd_config.c:443:class_cleanup()) Device 2 not setup&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.918942&amp;#93;&lt;/span&gt; LustreError: 23027:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.919097&amp;#93;&lt;/span&gt; LustreError: 23027:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.927817&amp;#93;&lt;/span&gt; Lustre: client lus04-client(ffff88041f2c3400) umount complete&lt;br/&gt;
Oct 20 08:24:16 bc-11-2-02 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;510327.927906&amp;#93;&lt;/span&gt; LustreError: 23027:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-2)&lt;/p&gt;</comment>
                            <comment id="69390" author="james beal" created="Mon, 21 Oct 2013 12:40:00 +0000"  >&lt;p&gt;We have managed to mount the file system. I will update the ticket with more information once I have started the operation to copy all the data off it &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;. Thanks for you help Bob&lt;/p&gt;</comment>
                            <comment id="69401" author="bogl" created="Mon, 21 Oct 2013 15:05:03 +0000"  >&lt;p&gt;James, Glad to hear you worked out a way to mount the file system successfully.  Will be curious to see exactly what combination of settings worked for you.  Looking forward to your post of more information.&lt;/p&gt;</comment>
                            <comment id="69507" author="james beal" created="Tue, 22 Oct 2013 11:21:24 +0000"  >&lt;p&gt;I have checked with Sven that he is happy with me posting this chat log.&lt;/p&gt;

&lt;p&gt;After breaking the MGS by adding the &#8211;mdt flag to it we formatted a new MGS, which is documented earlier in this bug &#8220;19/Oct/13&quot;&lt;/p&gt;

&lt;p&gt;mkfs.lustre --mgs --fsname=lus04 /dev/loop2&lt;/p&gt;

&lt;p&gt;We tried using service nodes rather than fail node, I also suspect that mounting things on the &quot;secondary&quot; node helped however I can&apos;t prove this as as it was working I was loathed to break the sequence to try experiments on it.&lt;/p&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:39:55&amp;#93;&lt;/span&gt; James Beal: root@lus04-mds1:/# tunefs.lustre  --writeconf --erase-params   --servicenode=172.17.148.4 --servicenode=172.17.148.5 --fsname=lus04 --mgs  /dev/loop2&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x74&lt;br/&gt;
              (MGS needs_index first_time update )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters:&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     MGS&lt;br/&gt;
Index:      unassigned&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1174&lt;br/&gt;
              (MGS needs_index first_time update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:51:55&amp;#93;&lt;/span&gt; James Beal: root@lus04-mds1:/# tunefs.lustre   --writeconf --erase-params   --mgsnode 172.17.148.4@tcp  --mgsnode 172.17.148.5@tcp     --servicenode=172.17.148.4 --servicenode=172.17.148.5 --fsname=lus04   /dev/loop3&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1001&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1141&lt;br/&gt;
              (MDT update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp failover.node=172.17.148.4@tcp failover.node=172.17.148.5@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;Unload the lustre modules on mds1&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:53:56&amp;#93;&lt;/span&gt; James Beal: root@lus04-mds1:/# mount -t lustre /dev/loop2 /export/MGS&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:54:02&amp;#93;&lt;/span&gt; James Beal: mount -t lustre /dev/loop3 /export/MDT0&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:54:22&amp;#93;&lt;/span&gt; Sven Trautmann: ok&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:54:24&amp;#93;&lt;/span&gt; James Beal: 162261.673514] LDISKFS-fs (loop3): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;162261.676015&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;162261.765128&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): barriers disabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;162261.766875&amp;#93;&lt;/span&gt; LDISKFS-fs (loop3): mounted filesystem with ordered data mode&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;162261.798124&amp;#93;&lt;/span&gt; Lustre: MGS: Logs for fs lus04 were removed by user request.  All servers must be restarted in order to regenerate the logs.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;162261.800164&amp;#93;&lt;/span&gt; Lustre: Enabling user_xattr&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;162261.801410&amp;#93;&lt;/span&gt; Lustre: lus04-MDT0000: Now serving lus04-MDT0000 on /dev/loop3 with recovery enabled&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:54:30&amp;#93;&lt;/span&gt; James Beal: Shall I try and mount a client ?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:55:18&amp;#93;&lt;/span&gt; Sven Trautmann: try a client, i&apos;m not very confident it will work&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:55:34&amp;#93;&lt;/span&gt; James Beal: I think it is as likely to work as with the OST&apos;s&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:55:43&amp;#93;&lt;/span&gt; Sven Trautmann: :x&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:56:35&amp;#93;&lt;/span&gt; James Beal: taking its time&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:57:22&amp;#93;&lt;/span&gt; James Beal: Oct 21 10:55:49 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927384.672219&amp;#93;&lt;/span&gt; Lustre: Acceptor stopping&lt;br/&gt;
Oct 21 10:55:51 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927386.672142&amp;#93;&lt;/span&gt; Lustre: Removed LNI 172.17.115.32@tcp&lt;br/&gt;
Oct 21 10:56:21 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927416.866605&amp;#93;&lt;/span&gt; Lustre: Build Version: v1_8_9_WC1sanger1--PRISTINE-2.6.32.59-sles-lustre-1.8.8wc1&lt;br/&gt;
Oct 21 10:56:21 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927416.937734&amp;#93;&lt;/span&gt; Lustre: Added LNI 172.17.115.32@tcp &lt;span class=&quot;error&quot;&gt;&amp;#91;8/256/0/180&amp;#93;&lt;/span&gt;&lt;br/&gt;
Oct 21 10:56:21 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927416.937915&amp;#93;&lt;/span&gt; Lustre: Accept secure, port 988&lt;br/&gt;
Oct 21 10:56:21 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927417.028382&amp;#93;&lt;/span&gt; Lustre: Lustre Client File System; &lt;a href=&quot;http://www.lustre.org/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://www.lustre.org/&lt;/a&gt;&lt;br/&gt;
Oct 21 10:56:21 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927417.048662&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 21 10:56:21 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927417.052298&amp;#93;&lt;/span&gt; LustreError: 11-0: an error occurred while communicating with 172.17.148.4@tcp. The mds_connect operation failed with -11&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:57:30&amp;#93;&lt;/span&gt; James Beal: Still trying&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:58:17&amp;#93;&lt;/span&gt; Sven Trautmann: resource temporarily unavailable, sounds like missing ost&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:58:26&amp;#93;&lt;/span&gt; James Beal: Oct 21 10:58:15 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17927530.816907&amp;#93;&lt;/span&gt; LustreError: 11-0: an error occurred while communicating with 172.17.148.4@tcp. The mds_connect operation failed with -11&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:58:31&amp;#93;&lt;/span&gt; James Beal: How about adding one OST&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:58:34&amp;#93;&lt;/span&gt; James Beal: vd00 ?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:58:46&amp;#93;&lt;/span&gt; James Beal: Or moving the MDS to .5 ?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:58:47&amp;#93;&lt;/span&gt; Sven Trautmann: you did the last one already?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:58:54&amp;#93;&lt;/span&gt; Sven Trautmann: 29&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:59:05&amp;#93;&lt;/span&gt; James Beal: Ok I will need to writeconf it again.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:59:10&amp;#93;&lt;/span&gt; Sven Trautmann: ok&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:59:25&amp;#93;&lt;/span&gt; James Beal: root@lus04-mds1:/# lctl dl&lt;br/&gt;
  0 UP mgs MGS MGS 13&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp 0654a817-b2ec-a591-d828-bb850199cfe1 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 10:59:57&amp;#93;&lt;/span&gt; James Beal: umounting all the OSS&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:00:16&amp;#93;&lt;/span&gt; Sven Trautmann: the ost where still mounted?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:00:22&amp;#93;&lt;/span&gt; James Beal: Yes.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:00:32&amp;#93;&lt;/span&gt; Sven Trautmann: ok&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:01:11&amp;#93;&lt;/span&gt; James Beal:  tunefs.lustre  --writeconf --erase-params  --servicenode=172.17.148.9 --servicenode=172.17.148.8  --mgsnode 172.17.148.4@tcp  --mgsnode 172.17.148.5@tcp  --fsname=lus04   /dev/mapper/vd29&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     lus04-OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1002&lt;br/&gt;
              (OST no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: failover.node=172.17.148.9@tcp failover.node=172.17.148.8@tcp mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     lus04-OST001d&lt;br/&gt;
Index:      29&lt;br/&gt;
Lustre FS:  lus04&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1142&lt;br/&gt;
              (OST update writeconf no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro,extents,mballoc&lt;br/&gt;
Parameters: failover.node=172.17.148.9@tcp failover.node=172.17.148.8@tcp mgsnode=172.17.148.4@tcp mgsnode=172.17.148.5@tcp&lt;/p&gt;

&lt;p&gt;Writing CONFIGS/mountdata&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:01:34&amp;#93;&lt;/span&gt; Sven Trautmann: ok&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:01:51&amp;#93;&lt;/span&gt; James Beal: oot@lus04-oss4:~# mount -t lustre /dev/mapper/vd29 /export/vd29&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:08&amp;#93;&lt;/span&gt; James Beal: Oct 21 11:01:44 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247251.590225&amp;#93;&lt;/span&gt; Lustre: lus04-OST001d: Now serving lus04-OST001d on /dev/mapper/vd29 with recovery enabled&lt;br/&gt;
Oct 21 11:01:44 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247251.590229&amp;#93;&lt;/span&gt; Lustre: lus04-OST001d: Will be in recovery for at least 5:00, or until 1 client reconnects&lt;br/&gt;
Oct 21 11:01:48 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247255.743705&amp;#93;&lt;/span&gt; LustreError: 20470:0:(ldlm_lib.c:887:target_handle_connect()) lus04-OST001d: NID 172.17.148.4@tcp (lus04-mdtlov_UUID) reconnected with 1 conn_cnt; cookies not random?&lt;br/&gt;
Oct 21 11:01:48 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247255.743908&amp;#93;&lt;/span&gt; LustreError: 20470:0:(ldlm_lib.c:887:target_handle_connect()) Skipped 7 previous similar messages&lt;br/&gt;
Oct 21 11:01:48 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247255.744049&amp;#93;&lt;/span&gt; LustreError: 20470:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (&lt;del&gt;114)  req@ffff88062b1e5c00 x1449498209419298/t0 o8&lt;/del&gt;&amp;gt;&amp;lt;?&amp;gt;@&amp;lt;?&amp;gt;:0/0 lens 368/264 e 0 to 0 dl 1382349808 ref 1 fl Interpret:/0/0 rc -114/0&lt;br/&gt;
Oct 21 11:01:48 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247255.744250&amp;#93;&lt;/span&gt; LustreError: 20470:0:(ldlm_lib.c:1921:target_send_reply_msg()) Skipped 7 previous similar messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:20&amp;#93;&lt;/span&gt; Sven Trautmann: good&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:27&amp;#93;&lt;/span&gt; James Beal: root@lus04-mds1:/# lctl dl&lt;br/&gt;
  0 UP mgs MGS MGS 9&lt;br/&gt;
  1 UP mgc MGC172.17.148.4@tcp 0654a817-b2ec-a591-d828-bb850199cfe1 5&lt;br/&gt;
  2 UP mdt MDS MDS_uuid 3&lt;br/&gt;
  3 UP lov lus04-mdtlov lus04-mdtlov_UUID 4&lt;br/&gt;
  4 UP mds lus04-MDT0000 lus04-MDT0000_UUID 3&lt;br/&gt;
  5 UP osc lus04-OST001d-osc lus04-mdtlov_UUID 5&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:32&amp;#93;&lt;/span&gt; Sven Trautmann: good&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:39&amp;#93;&lt;/span&gt; James Beal: Client still waiting&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:43&amp;#93;&lt;/span&gt; James Beal: Shall I interupt it&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:49&amp;#93;&lt;/span&gt; James Beal: And try again ?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:54&amp;#93;&lt;/span&gt; Sven Trautmann: no, the recovery needs to finish&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:02:58&amp;#93;&lt;/span&gt; James Beal: &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:03:09&amp;#93;&lt;/span&gt; Sven Trautmann: what is the mds saying about recovery?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:03:31&amp;#93;&lt;/span&gt; James Beal: Oct 21 11:02:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;162781.493095&amp;#93;&lt;/span&gt; Lustre: lus04-MDT0000: temporarily refusing client connection from 172.17.115.32@tcp&lt;br/&gt;
Oct 21 11:02:25 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;162781.493110&amp;#93;&lt;/span&gt; LustreError: 12347:0:(ldlm_lib.c:1921:target_send_reply_msg()) @@@ processing error (&lt;del&gt;11)  req@ffff88061f970800 x1449498384531486/t0 o38&lt;/del&gt;&amp;gt;&amp;lt;?&amp;gt;@&amp;lt;?&amp;gt;:0/0 lens 368/0 e 0 to 0 dl 1382349845 ref 1 fl Interpret:/0/0 rc -11/0&lt;br/&gt;
root@lus04-mds1:/#&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:03:42&amp;#93;&lt;/span&gt; James Beal: Which looks hopeful&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:04:17&amp;#93;&lt;/span&gt; Sven Trautmann: agree&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:04:23&amp;#93;&lt;/span&gt; James Beal: root@lus04-oss4:~# cat /proc/fs/lustre/obdfilter/lus04-OST001d/recovery_status&lt;br/&gt;
status: RECOVERING&lt;br/&gt;
recovery_start: 0&lt;br/&gt;
time_remaining: 0&lt;br/&gt;
connected_clients: 0/1&lt;br/&gt;
delayed_clients: 0/1&lt;br/&gt;
completed_clients: 0/1&lt;br/&gt;
replayed_requests: 0/??&lt;br/&gt;
queued_requests: 0&lt;br/&gt;
next_transno: 103079215105&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:04:39&amp;#93;&lt;/span&gt; James Beal: I would be happier if that was a bit different&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:06:15&amp;#93;&lt;/span&gt; James Beal: Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.171140&amp;#93;&lt;/span&gt; Lustre: lus04-OST001d: Recovery period over after 0:01, of 1 clients 1 recovered and 0 were evicted.&lt;br/&gt;
Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.171145&amp;#93;&lt;/span&gt; Lustre: Skipped 7 previous similar messages&lt;br/&gt;
Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.171148&amp;#93;&lt;/span&gt; Lustre: lus04-OST001d: sending delayed replies to recovered clients&lt;br/&gt;
Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.171150&amp;#93;&lt;/span&gt; Lustre: Skipped 7 previous similar messages&lt;br/&gt;
Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.171824&amp;#93;&lt;/span&gt; Lustre: lus04-OST001d: received MDS connection from 172.17.148.4@tcp&lt;br/&gt;
Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.171828&amp;#93;&lt;/span&gt; Lustre: Skipped 6 previous similar messages&lt;br/&gt;
Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.172491&amp;#93;&lt;/span&gt; Lustre: 20474:0:(filter.c:3129:filter_destroy_precreated()) lus04-OST001d: deleting orphan objects from 50974178 to 50974199, orphan objids won&apos;t be reused any more.&lt;br/&gt;
Oct 21 11:05:55 lus04-oss4 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;247502.172497&amp;#93;&lt;/span&gt; Lustre: 20474:0:(filter.c:3129:filter_destroy_precreated()) Skipped 6 previous similar messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:06:26&amp;#93;&lt;/span&gt; James Beal: cat /proc/fs/lustre/obdfilter/lus04-OST001d/recovery_status&lt;br/&gt;
status: COMPLETE&lt;br/&gt;
recovery_start: 1382349955&lt;br/&gt;
recovery_duration: 0&lt;br/&gt;
delayed_clients: 0/1&lt;br/&gt;
completed_clients: 1/1&lt;br/&gt;
replayed_requests: 0&lt;br/&gt;
last_transno: 103079215104&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:06:40&amp;#93;&lt;/span&gt; James Beal: Oct 21 11:05:55 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;162991.353819&amp;#93;&lt;/span&gt; Lustre: lus04-OST001d-osc: Connection restored to service lus04-OST001d using nid 172.17.148.9@tcp.&lt;br/&gt;
Oct 21 11:05:55 lus04-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;162991.354453&amp;#93;&lt;/span&gt; Lustre: MDS lus04-MDT0000: lus04-OST001d_UUID now active, resetting orphans&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:06:48&amp;#93;&lt;/span&gt; James Beal: ct 21 11:06:35 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17928030.831136&amp;#93;&lt;/span&gt; LustreError: 18685:0:(obd_mount.c:2067:lustre_fill_super()) Unable to mount  (-513)&lt;br/&gt;
Oct 21 11:06:35 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17928030.839158&amp;#93;&lt;/span&gt; Lustre: MGC172.17.148.4@tcp: Reactivating import&lt;br/&gt;
Oct 21 11:06:35 bc-11-2-07 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;17928030.844564&amp;#93;&lt;/span&gt; Lustre: Client lus04-client(ffff880260dc8c00) mount complete&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:07:07&amp;#93;&lt;/span&gt; Sven Trautmann: nice &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:07:28&amp;#93;&lt;/span&gt; James Beal: Lets get the rest of the OSS up.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:07:35&amp;#93;&lt;/span&gt; Sven Trautmann: &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/thumbs_up.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:07:42&amp;#93;&lt;/span&gt; Sven Trautmann: writeconf on all&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:07:47&amp;#93;&lt;/span&gt; James Beal: Yes&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:07:57&amp;#93;&lt;/span&gt; James Beal: A quick break &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:08:04&amp;#93;&lt;/span&gt; Sven Trautmann: sure&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:11:12&amp;#93;&lt;/span&gt; James Beal: Back.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:12:12&amp;#93;&lt;/span&gt; Sven Trautmann: ok, you do the tunefs on all OST?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:12:18&amp;#93;&lt;/span&gt; James Beal: I will&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:12:19&amp;#93;&lt;/span&gt; James Beal: root@lus04-oss3:/# for i in `seq -w 23 28 ` ; do echo tunefs.lustre  --writeconf --erase-params  --servicenode=172.17.148.9 --servicenode=172.17.148.8  --mgsnode 172.17.148.4@tcp  --mgsnode 172.17.148.5@tcp  --fsname=lus04 /dev/mapper/$i ; done&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.9 --servicenode=172.17.148.8 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/23&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.9 --servicenode=172.17.148.8 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/24&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.9 --servicenode=172.17.148.8 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/25&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.9 --servicenode=172.17.148.8 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/26&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.9 --servicenode=172.17.148.8 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/27&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.9 --servicenode=172.17.148.8 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/28&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:13:05&amp;#93;&lt;/span&gt; James Beal: for i in `seq -w 15 22 ` ; do echo tunefs.lustre  --writeconf --erase-params  --servicenode=172.17.148.8 --servicenode=172.17.148.9  --mgsnode 172.17.148.4@tcp  --mgsnode 172.17.148.5@tcp  --fsname=lus04 /dev/mapper/$i ; done&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/15&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/16&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/17&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/18&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/19&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/20&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/21&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.8 --servicenode=172.17.148.9 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/22&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:14:12&amp;#93;&lt;/span&gt; Sven Trautmann: looks ok&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:14:29&amp;#93;&lt;/span&gt; James Beal: root@lus04-oss2:~# for i in `seq -w 07 14 ` ; do echo tunefs.lustre  --writeconf --erase-params  --servicenode=172.17.148.6 --servicenode=172.17.148.7  --mgsnode 172.17.148.4@tcp  --mgsnode 172.17.148.5@tcp  --fsname=lus04 /dev/mapper/$i ; done&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/07&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/08&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/09&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/10&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/11&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/12&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/13&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.6 --servicenode=172.17.148.7 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/14&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:15:07&amp;#93;&lt;/span&gt; Sven Trautmann: i guess, i&apos;m not sure if the servicenode order has any influence at all&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:15:13&amp;#93;&lt;/span&gt; James Beal: root@lus04-oss1:~# for i in `seq -w 00 06 ` ; do echo tunefs.lustre  --writeconf --erase-params  --servicenode=172.17.148.7 --servicenode=172.17.148.6  --mgsnode 172.17.148.4@tcp  --mgsnode 172.17.148.5@tcp  --fsname=lus04 /dev/mapper/$i ; done&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.7 --servicenode=172.17.148.6 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/00&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.7 --servicenode=172.17.148.6 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/01&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.7 --servicenode=172.17.148.6 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/02&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.7 --servicenode=172.17.148.6 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/03&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.7 --servicenode=172.17.148.6 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/04&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.7 --servicenode=172.17.148.6 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/05&lt;br/&gt;
tunefs.lustre --writeconf --erase-params --servicenode=172.17.148.7 --servicenode=172.17.148.6 --mgsnode 172.17.148.4@tcp --mgsnode 172.17.148.5@tcp --fsname=lus04 /dev/mapper/06&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:15:21&amp;#93;&lt;/span&gt; James Beal: Nor do I but I am trying to be consistent&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:15:34&amp;#93;&lt;/span&gt; James Beal: Happy with those ?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:17:17&amp;#93;&lt;/span&gt; James Beal: ping&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:17:37&amp;#93;&lt;/span&gt; Sven Trautmann: as happy as can be&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:17:45&amp;#93;&lt;/span&gt; James Beal: ok going to do that&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:17:55&amp;#93;&lt;/span&gt; Sven Trautmann: &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/thumbs_up.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:18:16&amp;#93;&lt;/span&gt; James Beal: Was missing a vd...&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:19:19&amp;#93;&lt;/span&gt; James Beal: I will start the OST mounts&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:19:51&amp;#93;&lt;/span&gt; Sven Trautmann: ok&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:21:36&amp;#93;&lt;/span&gt; James Beal: for i in `seq -w 01 06 `;&lt;br/&gt;
do&lt;br/&gt;
mount -t lustre /dev/mapper/vd$i /export/vd$i&lt;br/&gt;
date&lt;br/&gt;
sleep 10&lt;br/&gt;
done&lt;br/&gt;
Mon Oct 21 11:21:11 BST 2013&lt;br/&gt;
Mon Oct 21 11:21:21 BST 2013&lt;br/&gt;
Mon Oct 21 11:21:31 BST 2013&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:22:21&amp;#93;&lt;/span&gt; James Beal: root@lus04-oss2:~# for i in `seq -w 01 06 `;&lt;br/&gt;
do&lt;br/&gt;
mount -t lustre /dev/mapper/vd$i /export/vd$i&lt;br/&gt;
date&lt;br/&gt;
sleep 10&lt;br/&gt;
done&lt;br/&gt;
Mon Oct 21 11:21:11 BST 2013&lt;br/&gt;
Mon Oct 21 11:21:21 BST 2013&lt;br/&gt;
Mon Oct 21 11:21:31 BST 2013&lt;br/&gt;
Mon Oct 21 11:21:42 BST 2013&lt;br/&gt;
Mon Oct 21 11:21:52 BST 2013&lt;br/&gt;
Mon Oct 21 11:22:02 BST 2013&lt;br/&gt;
root@lus04-oss2:~#&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:23:19&amp;#93;&lt;/span&gt; Sven Trautmann: lctl dl on the client?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:23:41&amp;#93;&lt;/span&gt; James Beal: root@bc-11-2-07:~# lctl dl&lt;br/&gt;
  0 UP mgc MGC172.17.148.4@tcp b2a666ed-eb18-98c4-5c4e-a98d7120a06b 5&lt;br/&gt;
  1 UP lov lus04-clilov-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 4&lt;br/&gt;
  2 UP mdc lus04-MDT0000-mdc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
  3 UP osc lus04-OST001d-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
  4 UP osc lus04-OST0000-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
  5 UP osc lus04-OST0001-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
  6 UP osc lus04-OST0002-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
  7 UP osc lus04-OST0003-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
  8 UP osc lus04-OST0004-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
  9 UP osc lus04-OST0005-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
 10 UP osc lus04-OST0006-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
 11 UP osc lus04-OST0007-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
 12 UP osc lus04-OST0008-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
 13 UP osc lus04-OST0009-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
 14 UP osc lus04-OST000a-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
 15 UP osc lus04-OST000b-osc-ffff880260dc8c00 ddad28f4-48ec-9bb6-043d-230b0b8696f1 5&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:23:42&amp;#93;&lt;/span&gt; James Beal: &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:24:04&amp;#93;&lt;/span&gt; Sven Trautmann: there are some missing?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:24:15&amp;#93;&lt;/span&gt; James Beal: I havn&apos;t finished the mounts&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:24:18&amp;#93;&lt;/span&gt; Sven Trautmann: ok&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:24:24&amp;#93;&lt;/span&gt; James Beal: root@lus04-oss1:~# for i in `seq -w 07 14 `;  do mount -t lustre /dev/mapper/vd$i /export/vd$i; date ; sleep 10; done&lt;br/&gt;
Mon Oct 21 11:22:46 BST 2013&lt;br/&gt;
Mon Oct 21 11:22:56 BST 2013&lt;br/&gt;
Mon Oct 21 11:23:06 BST 2013&lt;br/&gt;
Mon Oct 21 11:23:16 BST 2013&lt;br/&gt;
Mon Oct 21 11:23:27 BST 2013&lt;br/&gt;
Mon Oct 21 11:23:37 BST 2013&lt;br/&gt;
Mon Oct 21 11:23:47 BST 2013&lt;br/&gt;
Mon Oct 21 11:23:58 BST 2013&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:24:37&amp;#93;&lt;/span&gt; Sven Trautmann: right&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:26:32&amp;#93;&lt;/span&gt; James Beal: for i in `seq -w 15 22 `;  do mount -t lustre /dev/mapper/vd$i /export/vd$i; date ; sleep 10; done&lt;br/&gt;
Mon Oct 21 11:25:01 BST 2013&lt;br/&gt;
Mon Oct 21 11:25:11 BST 2013&lt;br/&gt;
Mon Oct 21 11:25:21 BST 2013&lt;br/&gt;
Mon Oct 21 11:25:32 BST 2013&lt;br/&gt;
Mon Oct 21 11:25:42 BST 2013&lt;br/&gt;
Mon Oct 21 11:25:52 BST 2013&lt;br/&gt;
Mon Oct 21 11:26:02 BST 2013&lt;br/&gt;
Mon Oct 21 11:26:13 BST 2013&lt;br/&gt;
root@lus04-oss3:/#&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:26:58&amp;#93;&lt;/span&gt; Sven Trautmann: recovery will take some time, i guess&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:27:04&amp;#93;&lt;/span&gt; James Beal: 5 minutes &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:27:28&amp;#93;&lt;/span&gt; Sven Trautmann: if everything goes as planned, yes&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;21/10/2013 11:27:49&amp;#93;&lt;/span&gt; James Beal: root@lus04-oss2:~# grep -i status /proc/fs/lustre/obdfilter/lus04-OST*/recovery_status&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0000/recovery_status:status: COMPLETE&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0001/recovery_status:status: COMPLETE&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0002/recovery_status:status: COMPLETE&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0003/recovery_status:status: COMPLETE&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0004/recovery_status:status: COMPLETE&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0005/recovery_status:status: COMPLETE&lt;br/&gt;
/proc/fs/lustre/obdfilter/lus04-OST0006/recovery_status:status: COMPLETE&lt;/p&gt;</comment>
                            <comment id="69560" author="svtr" created="Tue, 22 Oct 2013 18:04:09 +0000"  >&lt;p&gt;I had a similar problem today on another system.&lt;br/&gt;
For cosmetic reasons i had to change the hostnames and IPs of OSS and MDS servers.&lt;br/&gt;
I did the following:&lt;br/&gt;
1. unmounted Lustre&lt;br/&gt;
2. changed the IP addresses of the lnet interfaces (o2ib) and&lt;br/&gt;
3. did a tunefs on all targets.&lt;br/&gt;
4. mount mgs, mdt and osts&lt;/p&gt;

&lt;p&gt;After the first mount i got on the MDS:&lt;/p&gt;

&lt;p&gt;Oct 22 16:32:34 pfs2n12 kernel: : LustreError: 28201:0:(ldlm_lib.c:383:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
Oct 22 16:32:34 pfs2n12 kernel: : LustreError: 28201:0:(obd_config.c:565:class_setup()) setup pfs2dat2-OST000a-osc-MDT0000 failed (-2)&lt;br/&gt;
Oct 22 16:32:35 pfs2n12 kernel: : LustreError: 28201:0:(obd_config.c:1491:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
Oct 22 16:32:35 pfs2n12 kernel: : Lustre:    cmd=cf003 0:pfs2dat2-OST000a-osc-MDT0000  1:pfs2dat2-OST000a_UUID  2:0@&amp;lt;0:0&amp;gt;&lt;br/&gt;
Oct 22 16:32:50 pfs2n12 kernel: : Lustre: 28083:0:(ldlm_lib.c:952:target_handle_connect()) MGS: connection from bd8a639b-be45-6048-47b8-568404d7547d@172.26.8.19@o2ib t0 exp (null) cur 1382452370 last 0&lt;br/&gt;
Oct 22 16:32:50 pfs2n12 kernel: : Lustre: 28083:0:(ldlm_lib.c:952:target_handle_connect()) Skipped 2 previous similar messages&lt;/p&gt;

&lt;p&gt;the OSC device on the MDS was in state AT:&lt;/p&gt;

&lt;p&gt;  0 UP mgs MGS MGS 19&lt;br/&gt;
  1 UP mgc MGC172.26.8.12@o2ib c4ff7eb0-8c6f-9199-45f6-f75e490ac101 5&lt;br/&gt;
  2 UP lov pfs2dat2-MDT0000-mdtlov pfs2dat2-MDT0000-mdtlov_UUID 4&lt;br/&gt;
  3 UP mdt pfs2dat2-MDT0000 pfs2dat2-MDT0000_UUID 3&lt;br/&gt;
  4 UP mds mdd_obd-pfs2dat2-MDT0000 mdd_obd_uuid-pfs2dat2-MDT0000 3&lt;br/&gt;
  5 AT osc pfs2dat2-OST000a-osc-MDT0000 pfs2dat2-MDT0000-mdtlov_UUID 1&lt;/p&gt;

&lt;p&gt;just like in the problem described here.&lt;/p&gt;

&lt;p&gt;In my case the reason for the problem was that i forgot to unload the lnet kernel module after i change the IP of the Infiniband port. The lnet id did no longer match the IP of the underlying IB interface on the OSS.&lt;br/&gt;
After a clean lustre_rmmod the problem was gone and all OSTs could connect without problem.&lt;/p&gt;








</comment>
                            <comment id="97758" author="adilger" created="Tue, 28 Oct 2014 20:46:23 +0000"  >&lt;p&gt;Closing old bug.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzw647:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>11116</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>