<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:11:29 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-899] Client Connectivity Issues in Complex Lustre Environment</title>
                <link>https://jira.whamcloud.com/browse/LU-899</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Connectivity Issues:&lt;br/&gt;
Although the login nodes are able to mount both production systems, mounting of the second filesystem takes several minutes:&lt;/p&gt;

&lt;p&gt;client fe2:&lt;br/&gt;
Client fe2 - Mount Test:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# date&lt;br/&gt;
Mon Dec  5 17:31:17 UTC 2011&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# logger &quot;Start Testing&quot;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# date;mount /mnt/lustre1;date&lt;br/&gt;
Mon Dec  5 17:31:50 UTC 2011&lt;br/&gt;
Mon Dec  5 17:31:51 UTC 2011&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# date;mount /mnt/lustre2;date&lt;br/&gt;
Mon Dec  5 17:32:09 UTC 2011&lt;br/&gt;
Mon Dec  5 17:34:24 UTC 2011&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# logger &quot;End Testing&quot;&lt;br/&gt;
Log file attached - fe2.log&lt;/p&gt;

&lt;p&gt;Client fe2:&lt;br/&gt;
ib0: inet addr:10.174.0.38  Bcast:10.255.255.255  Mask:255.255.224.0&lt;br/&gt;
ib1: inet addr:10.175.0.38  Bcast:10.255.255.255  Mask:255.255.224.0&lt;br/&gt;
ib2: inet addr:10.174.81.11  Bcast:10.174.95.255  Mask:255.255.240.0&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib0(ib0), o2ib1(ib1), o2ib2(ib2)&quot;&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.0.38@o2ib&lt;br/&gt;
10.175.0.38@o2ib1&lt;br/&gt;
10.174.81.11@o2ib2&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# cat /etc/fstab | grep lustre&lt;br/&gt;
10.174.80.40@o2ib2:10.174.80.41@o2ib2:/scratch1	/mnt/lustre1	lustre	defaults,flock	0 0&lt;br/&gt;
10.174.80.42@o2ib2:10.174.80.43@o2ib2:/scratch2	/mnt/lustre2	lustre	defaults,flock	0 0&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# df -h | grep lustre&lt;br/&gt;
                      2.5P  4.7T  2.5P   1% /mnt/lustre1&lt;br/&gt;
                      3.1P  3.0T  3.1P   1% /mnt/lustre2&lt;/p&gt;

&lt;p&gt;The configuration of the data transfer nodes differs in that they only have 1 active ib port where the login nodes have 3.  Even so, they both use the same ib fabric to connect to the production filesystems.  The dtn nodes are able to mount the scratch2 filesystem without issue, but cannot mount the scratch1 filesystem.&lt;/p&gt;

&lt;p&gt;dtn1:&lt;br/&gt;
ib0: inet addr:10.174.81.1  Bcast:10.174.95.255  Mask:255.255.240.0&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib2(ib0)&quot;&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.81.1@o2ib2&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.80.40@o2ib2&lt;br/&gt;
12345-0@lo&lt;br/&gt;
12345-10.174.31.241@o2ib&lt;br/&gt;
12345-10.174.79.241@o2ib1&lt;br/&gt;
12345-10.174.80.40@o2ib2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.80.41@o2ib2&lt;br/&gt;
12345-0@lo&lt;br/&gt;
12345-10.174.31.251@o2ib&lt;br/&gt;
12345-10.174.79.251@o2ib1&lt;br/&gt;
12345-10.174.80.41@o2ib2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.80.42@o2ib2&lt;br/&gt;
12345-0@lo&lt;br/&gt;
12345-10.175.31.242@o2ib&lt;br/&gt;
12345-10.174.79.242@o2ib1&lt;br/&gt;
12345-10.174.80.42@o2ib2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.80.43@o2ib2&lt;br/&gt;
12345-0@lo&lt;br/&gt;
12345-10.175.31.252@o2ib&lt;br/&gt;
12345-10.174.79.252@o2ib1&lt;br/&gt;
12345-10.174.80.43@o2ib2&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# mount /mnt/lustre2&lt;br/&gt;
 &lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# df -h&lt;br/&gt;
Filesystem            Size  Used Avail Use% Mounted on&lt;br/&gt;
/dev/mapper/vg_dtn1-lv_root&lt;br/&gt;
                       50G  9.4G   38G  20% /&lt;br/&gt;
tmpfs                  24G   88K   24G   1% /dev/shm&lt;br/&gt;
/dev/sda1             485M   52M  408M  12% /boot&lt;br/&gt;
10.181.1.2:/contrib   132G  2.9G  129G   3% /contrib&lt;br/&gt;
10.181.1.2:/apps/v1   482G   38G  444G   8% /apps&lt;br/&gt;
10.181.1.2:/home      4.1T  404G  3.7T  10% /home&lt;br/&gt;
10.174.80.42@o2ib2:10.174.80.43@o2ib2:/scratch2&lt;br/&gt;
                      3.1P  3.0T  3.1P   1% /mnt/lustre2&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# mount /mnt/lustre1&lt;br/&gt;
mount.lustre: mount 10.174.80.40@o2ib2:10.174.80.41@o2ib2:/scratch1 at /mnt/lustre1 failed: No such file or directory&lt;br/&gt;
Is the MGS specification correct?&lt;br/&gt;
Is the filesystem name correct?&lt;br/&gt;
If upgrading, is the copied client log valid? (see upgrade docs)&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# cat /etc/fstab | grep lustre&lt;br/&gt;
10.174.80.40@o2ib2:10.174.80.41@o2ib2:/scratch1 /mnt/lustre1 lustre defaults,flock 0 0&lt;br/&gt;
10.174.80.42@o2ib2:10.174.80.43@o2ib2:/scratch2 /mnt/lustre2 lustre defaults,flock 0 0&lt;/p&gt;

&lt;p&gt;Finally, the TDS compute nodes cannot access the production filesystems.  They have the TDS filesystems mounted (lustre1 and lustre2).&lt;br/&gt;
This may be a simple networking issue.  Still investigating.&lt;/p&gt;

</description>
                <environment>The cluster configuration is as follows:&lt;br/&gt;
&lt;br/&gt;
scratch1 Lustre filesystem - 2 MDS, 16 OSS, 4 DDN SFA 10K arrays&lt;br/&gt;
scratch2 Lustre filesystem - 2 MDS, 20 OSS, 5 DDN SFA 10K arrays&lt;br/&gt;
&lt;br/&gt;
The scratch1 and scratch2 servers each have 4 IB ports.  The ports are used for client connectivity as follows:&lt;br/&gt;
&lt;br/&gt;
Production compute clients access scratch1 via the ib0 port.&lt;br/&gt;
Production compute clients access scratch2 via the ib1 port.&lt;br/&gt;
Test and Development systems (TDS) access both filesystems through the ib2 port.&lt;br/&gt;
Login nodes and data transfer nodes (DTN) access both filesystems through the ib3 port.&lt;br/&gt;
&lt;br/&gt;
The servers are running CentOS 5.5 (2.6.18-238.12.1.el5)&lt;br/&gt;
Lustre 1.8.6 with the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-530&quot; title=&quot;group qoutas not enforced&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-530&quot;&gt;&lt;strike&gt;LU-530&lt;/strike&gt;&lt;/a&gt; patch installed&lt;br/&gt;
The clients are currently running RHEL 6.0.&lt;br/&gt;
&lt;br/&gt;
Server Configuration:&lt;br/&gt;
lfs-mds-1-1&lt;br/&gt;
ib0 inet addr:10.174.31.241 Bcast:10.174.31.255 Mask:255.255.224.0&lt;br/&gt;
ib1 inet addr:10.175.31.241 Bcast:10.175.31.255 Mask:255.255.224.0&lt;br/&gt;
ib2 inet addr:10.174.79.241 Bcast:10.174.79.255 Mask:255.255.240.0&lt;br/&gt;
ib3 inet addr:10.174.80.40 Bcast:10.174.111.255 Mask:255.255.240.0&lt;br/&gt;
&lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-1-1&apos;&gt;root@lfs-mds-1-1&lt;/a&gt; ~]# cat /etc/modprobe.d/lustre.conf&lt;br/&gt;
options lnet networks=&amp;quot;o2ib0(ib0), o2ib1(ib2), o2ib2(ib3)&amp;quot;&lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-1-1&apos;&gt;root@lfs-mds-1-1&lt;/a&gt; config]# lctl list_nids&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.31.241@o2ib&apos;&gt;10.174.31.241@o2ib&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.79.241@o2ib1&apos;&gt;10.174.79.241@o2ib1&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.80.40@o2ib2&apos;&gt;10.174.80.40@o2ib2&lt;/a&gt;&lt;br/&gt;
lfs-mds-1-2:&lt;br/&gt;
ib0 inet addr:10.174.31.251 Bcast:10.174.31.255 Mask:255.255.224.0&lt;br/&gt;
ib1 inet addr:10.175.31.251 Bcast:10.175.31.255 Mask:255.255.224.0&lt;br/&gt;
ib2 inet addr:10.174.79.251 Bcast:10.174.79.255 Mask:255.255.240.0&lt;br/&gt;
ib3 inet addr:10.174.80.41 Bcast:10.174.111.255 Mask:255.255.240.0&lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-1-2&apos;&gt;root@lfs-mds-1-2&lt;/a&gt; ~]# cat /etc/modprobe.d/lustre.conf&lt;br/&gt;
options lnet networks=&amp;quot;o2ib0(ib0), o2ib1(ib2), o2ib2(ib3)&amp;quot; &lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-1-2&apos;&gt;root@lfs-mds-1-2&lt;/a&gt; ~]# lctl list_nids&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.31.251@o2ib&apos;&gt;10.174.31.251@o2ib&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.79.251@o2ib1&apos;&gt;10.174.79.251@o2ib1&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.80.41@o2ib2&apos;&gt;10.174.80.41@o2ib2&lt;/a&gt;&lt;br/&gt;
lfs-mds-2-1:&lt;br/&gt;
ib0 inet addr:10.174.31.242 Bcast:10.174.31.255 Mask:255.255.224.0&lt;br/&gt;
ib1 inet addr:10.175.31.242 Bcast:10.175.31.255 Mask:255.255.224.0&lt;br/&gt;
ib2 inet addr:10.174.79.242 Bcast:10.174.79.255 Mask:255.255.240.0&lt;br/&gt;
ib3 inet addr:10.174.80.42 Bcast:10.174.111.255 Mask:255.255.240.0&lt;br/&gt;
&lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-2-1&apos;&gt;root@lfs-mds-2-1&lt;/a&gt; ~]# cat /etc/modprobe.d/lustre.conf&lt;br/&gt;
options lnet networks=&amp;quot;o2ib0(ib0), o2ib1(ib2), o2ib2(ib3)&amp;quot; &lt;br/&gt;
&lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-2-1&apos;&gt;root@lfs-mds-2-1&lt;/a&gt; ~]# lctl list_nids&lt;br/&gt;
&lt;a href=&apos;mailto:10.175.31.242@o2ib&apos;&gt;10.175.31.242@o2ib&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.79.242@o2ib1&apos;&gt;10.174.79.242@o2ib1&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.80.42@o2ib2&apos;&gt;10.174.80.42@o2ib2&lt;/a&gt;&lt;br/&gt;
&lt;br/&gt;
lfs-mds-2-2:&lt;br/&gt;
ib0 inet addr:10.174.31.252 Bcast:10.174.31.255 Mask:255.255.224.0&lt;br/&gt;
ib1 inet addr:10.175.31.252 Bcast:10.175.31.255 Mask:255.255.224.0&lt;br/&gt;
ib2 inet addr:10.174.79.252 Bcast:10.174.79.255 Mask:255.255.240.0&lt;br/&gt;
ib3 inet addr:10.174.80.43 Bcast:10.174.111.255 Mask:255.255.240.0&lt;br/&gt;
&lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-2-2&apos;&gt;root@lfs-mds-2-2&lt;/a&gt; ~]# cat /etc/modprobe.d/lustre.conf&lt;br/&gt;
options lnet networks=&amp;quot;o2ib0(ib0), o2ib1(ib2), o2ib2(ib3)&amp;quot; &lt;br/&gt;
&lt;br/&gt;
[&lt;a href=&apos;mailto:root@lfs-mds-2-2&apos;&gt;root@lfs-mds-2-2&lt;/a&gt; ~]# lctl list_nids&lt;br/&gt;
&lt;a href=&apos;mailto:10.175.31.252@o2ib&apos;&gt;10.175.31.252@o2ib&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.79.252@o2ib1&apos;&gt;10.174.79.252@o2ib1&lt;/a&gt;&lt;br/&gt;
&lt;a href=&apos;mailto:10.174.80.43@o2ib2&apos;&gt;10.174.80.43@o2ib2&lt;/a&gt;&lt;br/&gt;
</environment>
        <key id="12598">LU-899</key>
            <summary>Client Connectivity Issues in Complex Lustre Environment</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="cliffw">Cliff White</assignee>
                                    <reporter username="dnelson@ddn.com">Dennis Nelson</reporter>
                        <labels>
                    </labels>
                <created>Mon, 5 Dec 2011 16:28:01 +0000</created>
                <updated>Wed, 14 Dec 2011 15:07:34 +0000</updated>
                            <resolved>Wed, 14 Dec 2011 14:47:51 +0000</resolved>
                                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="23690" author="cliffw" created="Mon, 5 Dec 2011 17:37:47 +0000"  >&lt;p&gt;I do not see any information about you MGS, are you running the MGS co-located with the MDS? It might be better for this configuration to have one separate MGS for all the filesystems.  If lctl ping works, it is odd that the mount would fail, it may be indeed a network issue. Also, you might check the MDS disk (tunefs --print) to see if the failover NIDS are correct on the disk.&lt;/p&gt;</comment>
                            <comment id="23691" author="dnelson@ddn.com" created="Mon, 5 Dec 2011 18:52:25 +0000"  >&lt;p&gt;The MGS is co-located with the MDS.  I did neglect to include the MDT nid information:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# tunefs.lustre --dryrun /dev/vg_scratch1/mdt &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     scratch1-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch1&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 mgsnode=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     scratch1-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch1&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 mgsnode=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;

&lt;p&gt;The MGT does not have any NID information.  It is my understanding that the client mount command specifies the nids of the systems where the MGT will be mounted.&lt;/p&gt;</comment>
                            <comment id="23692" author="dnelson@ddn.com" created="Mon, 5 Dec 2011 19:01:30 +0000"  >&lt;p&gt;The other thing I would point out is that the dtn nodes are able to mount the scratch2 filesystem.  They are attempting to mount the scratch1 filesystem over the same ib subnet (10.174.80.xx).  Additionally, the login nodes are able to mount both filesystems using that same subnet.  I don&apos;t understand how the cause could be a network issue when both the client and the server seem to be able to communicate over the subnet without issues.&lt;/p&gt;</comment>
                            <comment id="23694" author="cliffw" created="Mon, 5 Dec 2011 19:16:18 +0000"  >&lt;p&gt;You didn&apos;t include the scratch2 info, but on scratch1 I notice you have everything listed twice:&lt;/p&gt;

&lt;p&gt;Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 mgsnode=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;/p&gt;

&lt;p&gt;I think what you want would be everything listed once:&lt;br/&gt;
Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;br/&gt;
You do not need the double listing. &lt;/p&gt;</comment>
                            <comment id="23695" author="dnelson@ddn.com" created="Mon, 5 Dec 2011 19:21:58 +0000"  >&lt;p&gt;I do not know what you mean.  I have mgsnode listed twice, once for one mds, listing the 3 nids that are used for scratch1, and another time for the other mds, again, listing the 3 nids that are used for scratch1.  None of the nids are repeated.&lt;/p&gt;</comment>
                            <comment id="23748" author="cliffw" created="Tue, 6 Dec 2011 12:27:15 +0000"  >&lt;p&gt;You have the same NIDs listed as &apos;mgsnode&apos; and as &apos;failnode&apos; mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 ... failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 ... mdt.quota_type=ug&lt;/p&gt;

&lt;p&gt;That is not necessary, and may have something to do with the delay in mounting. A NID should be listed either as mgsnode, or failnode, not both as you have here. A NID only needs to be listed once, as mgsnode and failnode are both used when finding a server. The NIDs in the &apos;failover.node&apos; list do NOT have to be in the &apos;mgsnode&apos; list and should not be duplicated in this fashion. &lt;/p&gt;</comment>
                            <comment id="23753" author="dnelson@ddn.com" created="Tue, 6 Dec 2011 13:12:54 +0000"  >&lt;p&gt;So, would you suggest using mgsnode or failover.node?  Are they identical?  It appears that the DDN tools add both of these.&lt;/p&gt;</comment>
                            <comment id="23754" author="apittman" created="Tue, 6 Dec 2011 13:26:14 +0000"  >&lt;p&gt;This is the output of tunefs.lustre --print for the MDT on scratch2, it&apos;s from a snapshot taken last week so may not be up to date, Dennis, can you check this and update if it&apos;s wrong.&lt;/p&gt;

&lt;p&gt;checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     scratch2-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch2&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 mgsnode=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 failover.node=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 failover.node=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 mdt.quota_type=ug&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     scratch2-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch2&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 mgsnode=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 failover.node=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 failover.node=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 mdt.quota_type=ug&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;</comment>
                            <comment id="23755" author="dnelson@ddn.com" created="Tue, 6 Dec 2011 13:34:34 +0000"  >&lt;p&gt;Yes, it is the same.  It has not been modified since it was initially installed.&lt;/p&gt;</comment>
                            <comment id="23756" author="apittman" created="Tue, 6 Dec 2011 13:47:12 +0000"  >&lt;p&gt;To be clear, when Dennis says &quot;The MGS is co-located with the MDS&quot; what we mean is that it&apos;s running from a different partition on the same hardware.  The NIDS used to access it should be the same but it is a different partition and external to the MDT.&lt;/p&gt;</comment>
                            <comment id="23760" author="cliffw" created="Tue, 6 Dec 2011 14:13:38 +0000"  >&lt;p&gt;The primary NIDS for the node do not need to be listed as &apos;failnode&apos; When the MDT is registered with the MGS (which should always be done from the primary)&lt;br/&gt;
the primary NIDS will be added to the config log automatically. Afaik, all the duplication does is increase the length of the list searched when the primary isn&apos;t reachable. &lt;/p&gt;</comment>
                            <comment id="23761" author="cliffw" created="Tue, 6 Dec 2011 14:28:33 +0000"  >&lt;p&gt;You have multiple issues going on here. First, can you attach syslogs from a mount attempt from dtn1?&lt;br/&gt;
Second, the fe2 log would seem to indicate some network issues, I notice this sequence:&lt;br/&gt;
-------&lt;/p&gt;

&lt;p&gt;Dec  5 17:33:15 fe2 kernel: Lustre: scratch1-MDT0000-mdc-ffff8817f3d6e400: Connection to service scratch1-MDT0000 via nid 10.174.31.241@o2ib was lost; in progress operations using this service will wait for recovery to complete.&lt;br/&gt;
...&lt;/p&gt;

&lt;p&gt;Dec  5 17:33:29 fe2 kernel: Lustre: 20703:0:(import.c:517:import_select_connection()) scratch1-MDT0000-mdc-ffff8817f3d6e400: tried all connections, increasing latency to 2s&lt;br/&gt;
...&lt;/p&gt;

&lt;p&gt;Dec  5 17:34:24 fe2 kernel: Lustre: scratch1-MDT0000-mdc-ffff8817f3d6e400: Connection restored to service scratch1-MDT0000 using nid 10.174.31.241@o2ib.&lt;br/&gt;
--------&lt;br/&gt;
Which looks like what we see with network issues. &lt;/p&gt;</comment>
                            <comment id="23763" author="dnelson@ddn.com" created="Tue, 6 Dec 2011 14:35:02 +0000"  >&lt;p&gt;Yes, I believe there are multiple issues.  Different systems have different symptoms even across the same subnet.  I have not seen any indication of a network issue, although, I certainly will not rule out network issues.&lt;/p&gt;</comment>
                            <comment id="23799" author="dnelson@ddn.com" created="Wed, 7 Dec 2011 11:06:53 +0000"  >&lt;p&gt;Sorry, I sent the trace for the other case but did not send the trace for dtn1. Here it is.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lustre_rmmod&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# modprobe lustre&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl set_param debug=+trace lnet.debug=+trace&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl get_param debug&lt;br/&gt;
lnet.debug=trace ioctl neterror warning error emerg ha config console&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# mount /mnt/lustre1&lt;br/&gt;
mount.lustre: mount 10.174.80.40@o2ib2:10.174.80.41@o2ib2:/scratch1 at /mnt/lustre1 failed: No such file or directory&lt;br/&gt;
Is the MGS specification correct?&lt;br/&gt;
Is the filesystem name correct?&lt;br/&gt;
If upgrading, is the copied client log valid? (see upgrade docs)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lctl dk &amp;gt; /tmp/lustre-scratch1&lt;/p&gt;</comment>
                            <comment id="23803" author="cliffw" created="Wed, 7 Dec 2011 13:16:49 +0000"  >&lt;p&gt;Ah, the lustre dumps were requested on bug 890 - I need the &lt;em&gt;syslog&lt;/em&gt; (/var/log/messages or dmesg) from the dtn1 mount attempt. Thanks again&lt;/p&gt;</comment>
                            <comment id="23804" author="cliffw" created="Wed, 7 Dec 2011 13:38:23 +0000"  >&lt;p&gt;Hmm. I am starting to suspect there may be config log issues, I see this:&lt;br/&gt;
You attempt to mount using the ib3 address (.40), but when the mount starts to reach the MDT, it only tries the&lt;br/&gt;
primary address:&lt;br/&gt;
....&lt;br/&gt;
00000020:00000080:13:1323273617.875570:0:5786:0:(obd_config.c:857:class_process_config()) marker 4 (0x1) scratch1-MDT0000 add mdc&lt;br/&gt;
....&lt;br/&gt;
00000020:00000080:13:1323273617.875579:0:5786:0:(obd_config.c:799:class_process_config()) adding mapping from uuid 10.174.31.241@o2ib to nid 0x500000aae1ff1 (10.174.31.241@o2ib)&lt;br/&gt;
...&lt;br/&gt;
00000100:00000100:13:1323273617.875666:0:5786:0:(client.c:69:ptlrpc_uuid_to_connection()) cannot find peer 10.174.31.241@o2ib!&lt;br/&gt;
00010000:00080000:13:1323273617.875667:0:5786:0:(ldlm_lib.c:71:import_set_conn()) can&apos;t find connection 10.174.31.241@o2ib&lt;br/&gt;
00010000:00000001:13:1323273617.875667:0:5786:0:(ldlm_lib.c:72:import_set_conn()) Process leaving (rc=18446744073709551614 : -2 : fffffffffffffffe)&lt;br/&gt;
00010000:00020000:13:1323273617.875668:0:5786:0:(ldlm_lib.c:331:client_obd_setup()) can&apos;t add initial connection&lt;/p&gt;

&lt;p&gt;and i don&apos;t see it trying any of the alternate addresses. &lt;br/&gt;
So, it is possible the on-disk (config log) configuration has an issue. &lt;/p&gt;</comment>
                            <comment id="23831" author="dnelson@ddn.com" created="Wed, 7 Dec 2011 14:33:15 +0000"  >&lt;p&gt;So, would a --writeconf for all devices be in order.  I was thinking of doing that but I have not had any downtime to do it.  The filesystem is not in production but others are using it to prepare for acceptance.  I can certainly schedule time if you suggest that is the right plan of action.  That might also fix the issue that I am seeing in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-890&quot; title=&quot;MDS Failover Issue - Clients not reconnecting after MGT/MDT fail over to other MDS.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-890&quot;&gt;&lt;del&gt;LU-890&lt;/del&gt;&lt;/a&gt; (clients do not reconect after MDS failover).&lt;/p&gt;</comment>
                            <comment id="23842" author="dnelson@ddn.com" created="Wed, 7 Dec 2011 14:53:59 +0000"  >&lt;p&gt;Sorry, trying to do many things at once.  Here are the syslogs from dtn1:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# umount -at lustre&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# lustre_rmmod&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# logger &quot;Start Test&quot;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# mount -at lustre&lt;br/&gt;
mount.lustre: mount 10.174.80.40@o2ib2:10.174.80.41@o2ib2:/scratch1 at /mnt/lustre1 failed: No such file or directory&lt;br/&gt;
Is the MGS specification correct?&lt;br/&gt;
Is the filesystem name correct?&lt;br/&gt;
If upgrading, is the copied client log valid? (see upgrade docs)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@dtn1 ~&amp;#93;&lt;/span&gt;# logger &quot;stop Test&quot;&lt;/p&gt;



&lt;p&gt;Dec  7 14:51:49 dtn1 root: Start Test&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: OBD class driver, &lt;a href=&quot;http://wiki.whamcloud.com/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://wiki.whamcloud.com/&lt;/a&gt;&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre:     Lustre Version: 1.8.6&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre:     Build Version: jenkins-wc1-ga73a0cf-PRISTINE-2.6.32-71.el6.x86_64&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: Listener bound to ib0:10.174.81.1:987:mlx4_0&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: Register global MR array, MR size: 0xffffffffffffffff, array size: 1&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: Added LNI 10.174.81.1@o2ib2 &lt;span class=&quot;error&quot;&gt;&amp;#91;8/64/0/180&amp;#93;&lt;/span&gt;&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: Lustre Client File System; &lt;a href=&quot;http://www.lustre.org/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://www.lustre.org/&lt;/a&gt;&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: MGC10.174.80.40@o2ib2: Reactivating import&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7420:0:(ldlm_lib.c:331:client_obd_setup()) can&apos;t add initial connection&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7420:0:(obd_config.c:372:class_setup()) setup scratch1-MDT0000-mdc-ffff88060fc9f800 failed (-2)&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7420:0:(obd_config.c:1199:class_config_llog_handler()) Err -2 on cfg command:&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre:    cmd=cf003 0:scratch1-MDT0000-mdc  1:scratch1-MDT0000_UUID  2:10.174.31.241@o2ib  &lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 15c-8: MGC10.174.80.40@o2ib2: The configuration from log &apos;scratch1-client&apos; failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7369:0:(llite_lib.c:1095:ll_fill_super()) Unable to process log: -2&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7369:0:(obd_config.c:443:class_cleanup()) Device 2 not setup&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7369:0:(ldlm_request.c:1039:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7369:0:(ldlm_request.c:1597:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: client ffff88060fc9f800 umount complete&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: LustreError: 7369:0:(obd_mount.c:2065:lustre_fill_super()) Unable to mount  (-2)&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: MGC10.174.80.42@o2ib2: Reactivating import&lt;br/&gt;
Dec  7 14:51:56 dtn1 kernel: Lustre: Client scratch2-client has started&lt;br/&gt;
Dec  7 14:52:09 dtn1 root: stop Test&lt;/p&gt;</comment>
                            <comment id="23880" author="cliffw" created="Wed, 7 Dec 2011 15:57:10 +0000"  >&lt;p&gt;We&apos;d really like to know what is wrong before doing the writeconf. Would be good to know what the current config is. &lt;br/&gt;
Please do this for both scratch1 and 2, my example uses &apos;test2&apos; where you would use &apos;scratch1&apos; and &apos;scratch2&apos;, and replace /dev/hdb with your MGT partition.&lt;br/&gt;
Using the MGS partition (MGT)&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;umount /mnt/mgs&lt;/li&gt;
	&lt;li&gt;mount -t ldiskfs /dev/hdb /mnt/mgs&lt;/li&gt;
	&lt;li&gt;/usr/sbin/llog_reader /mnt/mgs/CONFIGS/test2-client | grep uuid&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;The result should be something like this:&lt;/p&gt;

&lt;p&gt;Target uuid : config_uuid &lt;br/&gt;
                uuid=test2-clilov_UUID  stripe:cnt=1 size=1048576 offset=18446744073709551615 pattern=0x1&lt;br/&gt;
                uuid=test2-clilmv_UUID  stripe:cnt=0 size=0 offset=0 pattern=0&lt;br/&gt;
#10 (080)add_uuid  nid=10.67.73.72@tcp(0x200000a434948)  0:  1:10.67.73.72@tcp  &lt;br/&gt;
#19 (080)add_uuid  nid=10.67.73.71@tcp(0x200000a434947)  0:  1:10.67.73.71@tcp  &lt;br/&gt;
#25 (080)add_uuid  nid=10.67.73.71@tcp(0x200000a434947)  0:  1:10.67.73.71@tcp  &lt;br/&gt;
#31 (080)add_uuid  nid=10.67.73.82@tcp(0x200000a434952)  0:  1:10.67.73.82@tcp &lt;/p&gt;

&lt;p&gt;In my config, 10.67.73.82 is the MDS failover NID, the other NIDS are OSS/MDS primary NID. Check your results and verify the proper NIDS are being given to the clients. If all the NIDS aren&apos;t in the config log, a writeconf is needed. If the clients are getting a correct NID list from these config logs, then the issue is most likely networking. &lt;/p&gt;</comment>
                            <comment id="23911" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 10:44:43 +0000"  >&lt;p&gt;Sorry for the delay.  I captured the info last night and before I could upload the files my laptop died.  I had to reschedule time to get the system again.&lt;/p&gt;</comment>
                            <comment id="23915" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 11:22:08 +0000"  >&lt;p&gt;So, it is my understanding that I should find the client nids in the output.  I cannot find the dtn1 nid in either although dtn1 has scratch2 mounted.  In fact, I cannot find any of the 10.174.81.xx nids in the lustre1.  I assume that is part of our problem but what would cause that?&lt;/p&gt;</comment>
                            <comment id="23922" author="cliffw" created="Thu, 8 Dec 2011 12:00:31 +0000"  >&lt;p&gt;No, there are no client NIDs in the config log. As I mentioned previous, there are only server NIDS, we wanted to see if all the server NIDS were being given to the clients. The error from the dtn1 mount would appear to indicate a possible corrupt config log.&lt;/p&gt;</comment>
                            <comment id="23926" author="cliffw" created="Thu, 8 Dec 2011 12:24:42 +0000"  >&lt;p&gt;this is being escalated - please attach the full config logs from both systems. Do the same thing as previous, but instead of the &apos;grep uuid&apos; just &amp;gt; the whole thing to a file and attach.&lt;/p&gt;</comment>
                            <comment id="23931" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 14:35:57 +0000"  >&lt;p&gt;OK, here they are.&lt;/p&gt;</comment>
                            <comment id="23940" author="johann" created="Thu, 8 Dec 2011 17:56:52 +0000"  >&lt;p&gt;It seems that the config logs of scratch1 do not properly set up the 3 nids of lfs-mds-1-1. Only 10.174.31.241@o2ib is added to the niduuid before attach/setup. Could you please run the following command on a login node which has the 2 filesystems mounted:&lt;br/&gt;
$ lctl get_param mdc.*.import&lt;/p&gt;

&lt;p&gt;The only way to fix this would be to regenerate the config logs with writeconf.&lt;/p&gt;</comment>
                            <comment id="23943" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 19:11:00 +0000"  >&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# lctl get_param mdc.*.import&lt;br/&gt;
mdc.scratch1-MDT0000-mdc-ffff88109cc8a000.import=&lt;br/&gt;
import:&lt;br/&gt;
    name: scratch1-MDT0000-mdc-ffff88109cc8a000&lt;br/&gt;
    target: scratch1-MDT0000_UUID&lt;br/&gt;
    state: FULL&lt;br/&gt;
    connect_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;version, inode_bit_locks, join_file, getattr_by_fid, no_oh_for_devices, early_lock_cancel, adaptive_timeouts, version_recovery, pools&amp;#93;&lt;/span&gt;&lt;br/&gt;
    import_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;replayable, pingable&amp;#93;&lt;/span&gt;&lt;br/&gt;
    connection:&lt;br/&gt;
       failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;10.174.31.241@o2ib, 10.174.80.41@o2ib&amp;#93;&lt;/span&gt;&lt;br/&gt;
       current_connection: 10.174.31.241@o2ib&lt;br/&gt;
       connection_attempts: 10&lt;br/&gt;
       generation: 1&lt;br/&gt;
       in-progress_invalidations: 0&lt;br/&gt;
    rpcs:&lt;br/&gt;
       inflight: 0&lt;br/&gt;
       unregistering: 0&lt;br/&gt;
       timeouts: 9&lt;br/&gt;
       avg_waittime: 1402 usec&lt;br/&gt;
    service_estimates:&lt;br/&gt;
       services: 1 sec&lt;br/&gt;
       network: 1 sec&lt;br/&gt;
    transactions:&lt;br/&gt;
       last_replay: 0&lt;br/&gt;
       peer_committed: 0&lt;br/&gt;
       last_checked: 0&lt;br/&gt;
mdc.scratch2-MDT0000-mdc-ffff8817b83d0800.import=&lt;br/&gt;
import:&lt;br/&gt;
    name: scratch2-MDT0000-mdc-ffff8817b83d0800&lt;br/&gt;
    target: scratch2-MDT0000_UUID&lt;br/&gt;
    state: FULL&lt;br/&gt;
    connect_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;version, inode_bit_locks, join_file, getattr_by_fid, no_oh_for_devices, early_lock_cancel, adaptive_timeouts, version_recovery, pools&amp;#93;&lt;/span&gt;&lt;br/&gt;
    import_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;replayable, pingable&amp;#93;&lt;/span&gt;&lt;br/&gt;
    connection:&lt;br/&gt;
       failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;10.175.31.242@o2ib, 10.174.63.242@o2ib, 10.174.63.252@o2ib, 10.175.31.252@o2ib&amp;#93;&lt;/span&gt;&lt;br/&gt;
       current_connection: 10.175.31.242@o2ib&lt;br/&gt;
       connection_attempts: 16&lt;br/&gt;
       generation: 1&lt;br/&gt;
       in-progress_invalidations: 0&lt;br/&gt;
    rpcs:&lt;br/&gt;
       inflight: 0&lt;br/&gt;
       unregistering: 0&lt;br/&gt;
       timeouts: 15&lt;br/&gt;
       avg_waittime: 1643 usec&lt;br/&gt;
    service_estimates:&lt;br/&gt;
       services: 1 sec&lt;br/&gt;
       network: 1 sec&lt;br/&gt;
    transactions:&lt;br/&gt;
       last_replay: 0&lt;br/&gt;
       peer_committed: 0&lt;br/&gt;
       last_checked: 0&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe2 ~&amp;#93;&lt;/span&gt;# &lt;/p&gt;

&lt;p&gt;Any ideas why scratch2 would not show the 10.174.80.&lt;span class=&quot;error&quot;&gt;&amp;#91;42,43&amp;#93;&lt;/span&gt; addresses listed?  The way this was designed by the customer was that the login nodes would use the .80 subnet.  The login nodes really only should have the single nid I added the others as a workaround.&lt;/p&gt;</comment>
                            <comment id="23945" author="cliffw" created="Thu, 8 Dec 2011 19:33:48 +0000"  >&lt;p&gt;writeconf should fix the issue&lt;/p&gt;</comment>
                            <comment id="23946" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 19:53:59 +0000"  >&lt;p&gt;OK, so I understand we are going to need to do a writeconf.  I actually asked about doing that already.  My customer is going to be asking a lot of questions tomorrow morning, so let me ask a few now.&lt;/p&gt;

&lt;p&gt;1.  Any idea why this did not work in the first place?&lt;br/&gt;
2.  Is there a limit to the numbers of nids that a server has?  Are we reaching some limit in LNET?&lt;br/&gt;
3.  What makes us think it will work after doing a writeconf?&lt;br/&gt;
4.  Is there any point of trying to troubleshoot any longer before we do the writeconf or should I go ahead and do it tomorrow?&lt;br/&gt;
5.  The DDN tools set --mgsnode and --servernode.  Based on previous info from Cliff, he suggested that we do not need both.  So, should I just use the --mgsnode flags and drop the --servernode flags?&lt;/p&gt;

&lt;p&gt;Thanks,&lt;/p&gt;</comment>
                            <comment id="23984" author="johann" created="Fri, 9 Dec 2011 02:44:07 +0000"  >&lt;p&gt;&amp;gt; Any ideas why scratch2 would not show the 10.174.80.&lt;span class=&quot;error&quot;&gt;&amp;#91;42,43&amp;#93;&lt;/span&gt; addresses listed?&lt;/p&gt;

&lt;p&gt;The first time a target is mounted, it registers all its configured nids to the MGS.&lt;br/&gt;
Then this list of nids is passed to the client which selects the most suitable one (it discards the others) to be used according to its local network configuration. Failover will only be done across nids registered with the --failnode option.&lt;/p&gt;

&lt;p&gt;&amp;gt; 1. Any idea why this did not work in the first place?&lt;/p&gt;

&lt;p&gt;Could you please tell us the exact commands you ran when you formatted the the MDTs?&lt;br/&gt;
Could you please also run tunefs.lustre --print against the MDT devices?&lt;/p&gt;

&lt;p&gt;&amp;gt; 2. Is there a limit to the numbers of nids that a server has? Are we reaching some limit in LNET?&lt;/p&gt;

&lt;p&gt;Not as far as i know. However, configuring multiple failover nids can increase the recovery time significantly.&lt;/p&gt;

&lt;p&gt;&amp;gt; 3. What makes us think it will work after doing a writeconf?&lt;br/&gt;
&amp;gt; 4. Is there any point of trying to troubleshoot any longer before we do the writeconf or should I go ahead and do it tomorrow?&lt;br/&gt;
&amp;gt; 5. The DDN tools set --mgsnode and --servernode. Based on previous info from Cliff, he suggested that we do not need both. So, should I just use the --mgsnode flags and drop the --servernode flags?&lt;/p&gt;

&lt;p&gt;Let&apos;s first check how you configured the filesystem in the first place before going down this path.&lt;/p&gt;</comment>
                            <comment id="23993" author="dnelson@ddn.com" created="Fri, 9 Dec 2011 09:07:00 +0000"  >&lt;p&gt;I used the DDN tools to format the MDTs.  It is done in two steps.  First, the formatting command uses generic placeholders for the NIDs.  Then, a tunefs step is performed:&lt;/p&gt;

&lt;p&gt;Step 1:&lt;br/&gt;
mkfs.lustre --mgs /dev/vg_scratch1/mgs&lt;br/&gt;
tune2fs -O MMP /dev/vg_scratch1/mgs&lt;br/&gt;
tune2fs -p 5 /dev/vg_scratch1/mgs&lt;br/&gt;
mkfs.lustre --mgsnode=127.0.0.2@tcp --failnode=127.0.0.2@tcp --fsname=scratch1 --mdt /dev/vg_scratch1/mdt&lt;br/&gt;
tune2fs -p 5 /dev/vg_scratch1/mdt&lt;/p&gt;

&lt;p&gt;Step 2:&lt;br/&gt;
tunefs.lustre --erase-params /dev/vg_scratch1/mgs&lt;br/&gt;
tunefs.lustre --erase-params --mgsnode=10.174.31.241@o2ib0,10.174.79.241@o2ib1,10.174.80.40@o2ib2 --mgsnode=10.174.31.251@o2ib0,10.174.79.251@o2ib1,10.174.80.41@o2ib2 --servicenode=10.174.31.241@o2ib0,10.174.79.241@o2ib1,10.174.80.40@o2ib2 --servicenode=10.174.31.251@o2ib0,10.174.79.251@o2ib1,10.174.80.41@o2ib2 --param mdt.quota_type=ug /dev/vg_scratch1/mdt&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 scratch1&amp;#93;&lt;/span&gt;# tunefs.lustre --print /dev/vg_scratch1/mdt &lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     scratch1-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch1&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 mgsnode=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     scratch1-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch1&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 mgsnode=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;


</comment>
                            <comment id="24012" author="johann" created="Fri, 9 Dec 2011 10:07:54 +0000"  >&lt;p&gt;Could you please also run tunefs.lustre on the MDT of scracth2?&lt;/p&gt;</comment>
                            <comment id="24013" author="apittman" created="Fri, 9 Dec 2011 10:12:06 +0000"  >&lt;p&gt;What options do you recommend for the writeconf?  As well as the configuration data I assume they options --erase-params and --writeconf should both be set?&lt;/p&gt;</comment>
                            <comment id="24014" author="dnelson@ddn.com" created="Fri, 9 Dec 2011 10:14:20 +0000"  >&lt;p&gt;Commands used:&lt;/p&gt;

&lt;p&gt;mkfs.lustre --mgsnode=127.0.0.2@tcp --failnode=127.0.0.2@tcp --fsname=scratch2 --mdt /dev/vg_scratch2/mdt&lt;br/&gt;
tune2fs -p 5 /dev/vg_scratch2/mdt&lt;/p&gt;

&lt;p&gt;tunefs.lustre --erase-params /dev/vg_scratch2/mgs&lt;br/&gt;
tunefs.lustre --erase-params --mgsnode=10.175.31.242@o2ib0,10.174.79.242@o2ib1,10.174.80.42@o2ib2 --mgsnode=10.175.31.252@o2ib0,10.174.79.252@o2ib1,10.174.80.43@o2ib2 --servicenode=10.175.31.242@o2ib0,10.174.79.242@o2ib1,10.174.80.42@o2ib2 --servicenode=10.175.31.252@o2ib0,10.174.79.252@o2ib1,10.174.80.43@o2ib2 --param mdt.quota_type=ug /dev/vg_scratch2/mdt&lt;/p&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-2-1 ~&amp;#93;&lt;/span&gt;# tunefs.lustre --print /dev/vg_scratch2/mdt&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     scratch2-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch2&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 mgsnode=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 failover.node=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 failover.node=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 mdt.quota_type=ug&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     scratch2-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch2&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 mgsnode=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 failover.node=10.175.31.242@o2ib,10.174.79.242@o2ib1,10.174.80.42@o2ib2 failover.node=10.175.31.252@o2ib,10.174.79.252@o2ib1,10.174.80.43@o2ib2 mdt.quota_type=ug&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;
</comment>
                            <comment id="24018" author="johann" created="Fri, 9 Dec 2011 10:49:53 +0000"  >&lt;p&gt;&amp;gt; What options do you recommend for the writeconf?&lt;/p&gt;

&lt;p&gt;writeconf can now be passed as a mount option. So i would unmount all clients, MDT and OSTs of scratch1 (not the shared mgs). Then mount the MDT again with -o writeconf and then the OSTs with the same mount option. Then once all targets are up and running, you can mount clients again.&lt;br/&gt;
Before proceeding with writeconf, i would advise to back up the CONFIGS directory of the MGS.&lt;/p&gt;

&lt;p&gt;I also looked at the commands you used to format the MDS and everything looks good. It is still unclear why the resulting logs are bogus.&lt;/p&gt;</comment>
                            <comment id="24021" author="dnelson@ddn.com" created="Fri, 9 Dec 2011 10:58:02 +0000"  >&lt;p&gt;Let me clarify something.  The customer specified that each filesystem needed to be fully independent,  Given that, there is not a common MDS.  There is an MDS/MGT for each filesystem.  The MGS services runs co-located on the MDS system.  The MGT is a separate LVM from the MDT although they both reside on a single volume group.&lt;/p&gt;</comment>
                            <comment id="24023" author="dnelson@ddn.com" created="Fri, 9 Dec 2011 11:09:05 +0000"  >&lt;p&gt;Another question, after looking at the tunefs.lustre output, Cliff suggested that the mgsnode and failover.node definitions were duplicates of each other and did not both need to be there.  Now, you are saying that the commands are correct.  We use the --servicenode syntax in our tunefs commands yet lustre.tunefs displays failover.node.  Just to confirm, the mgsnode and failover.node (or servicenode) options must both be present, and the servicenode/failover.node entries are different ways of setting the same parameter?  In our case where the mgs and the mds are always on the same server, mgsnode and servicenodes are identical but that would not always be the case.&lt;/p&gt;</comment>
                            <comment id="24029" author="johann" created="Fri, 9 Dec 2011 12:29:29 +0000"  >&lt;p&gt;&amp;gt; Let me clarify something. The customer specified that each filesystem needed to be fully independent ...&lt;/p&gt;

&lt;p&gt;Understood. This should not change the procedure in any case.&lt;/p&gt;

&lt;p&gt;&amp;gt; Just to confirm, the mgsnode and failover.node (or servicenode) options must both be present&lt;/p&gt;

&lt;p&gt;Yes. The mgsnode(s) is the nid(s) that will be used by the MDT to connect to the MGS, while failover.node is what will be registered by the MDS to to the MDS and used by client nodes to reach the MDS.&lt;/p&gt;

&lt;p&gt;&amp;gt; the servicenode/failover.node entries are different ways of setting the same parameter?&lt;/p&gt;

&lt;p&gt;Correct.&lt;/p&gt;

&lt;p&gt;&amp;gt; In our case where the mgs and the mds are always on the same server, mgsnode and servicenodes are identical but that would not always be the case&lt;/p&gt;

&lt;p&gt;Exactly. It can (must) only be skipped if you run a combo MGT/MDT on the same device.&lt;/p&gt;</comment>
                            <comment id="24033" author="cliffw" created="Fri, 9 Dec 2011 12:47:21 +0000"  >&lt;p&gt;I was unaware you were using --servicenode instead of --failnode, that explains the discrepancy. &lt;/p&gt;</comment>
                            <comment id="24034" author="dnelson@ddn.com" created="Fri, 9 Dec 2011 13:03:29 +0000"  >&lt;p&gt;Thanks.  I need to give an update to my customer.  At what point are we going to decide to do the writeconf and how confident are we that it will work when we do it?&lt;/p&gt;</comment>
                            <comment id="24036" author="cliffw" created="Fri, 9 Dec 2011 15:27:05 +0000"  >&lt;p&gt;Per Johann, go ahead and do the writeconf. We will attempt to replicate the issue in our lab with multiple nids. &lt;/p&gt;</comment>
                            <comment id="24041" author="johann" created="Fri, 9 Dec 2011 17:47:42 +0000"  >&lt;p&gt;I think it would be interesting to collect debug logs of the MGS during the re-registration.&lt;br/&gt;
Before remounting the MDS, please run the following comment on the MGS node:&lt;/p&gt;

&lt;p&gt;$ lctl set_param subsystem_debug=mgs # only collect debug messages from the MGS&lt;br/&gt;
$ lctl set_param debug=+config+info  # enable config &amp;amp; info in the debug mask&lt;br/&gt;
$ lctl set_param debug_mb=500        #&#160;set debug buffer to 500MB&lt;/p&gt;

&lt;p&gt;And then run lctl dk &amp;gt; /tmp/logs once the MDS has been successfully mounted.&lt;br/&gt;
Please also collect the new output of llog_reader so that we can compare.&lt;/p&gt;

&lt;p&gt;We will try to reproduce on our side too.&lt;/p&gt;</comment>
                            <comment id="24042" author="dnelson@ddn.com" created="Fri, 9 Dec 2011 17:57:32 +0000"  >&lt;p&gt;OK, one more question.  You mentioned that the writeconf could be done as a mount option.  Should I do it that way or should I use the script previously provided?  Does it make any difference?  Do you have a preference of how we do it?  If we do it as a mount option, I think I would need more info on how that works.  How do the mgsnode and servicenode options get added doing it with as a mount option?&lt;/p&gt;</comment>
                            <comment id="24046" author="johann" created="Fri, 9 Dec 2011 19:26:13 +0000"  >&lt;p&gt;&amp;gt; OK, one more question. You mentioned that the writeconf could be done as a mount option.&lt;/p&gt;

&lt;p&gt;Right.&lt;/p&gt;

&lt;p&gt;&amp;gt; Should I do it that way or should I use the script previously provided?&lt;/p&gt;

&lt;p&gt;It is really as you prefer. If your tool supports writeconf, then you can just use it. On my side, I just find it very convenient to pass -o writeconf as a mount option.&lt;/p&gt;

&lt;p&gt;&amp;gt; Does it make any difference?&lt;/p&gt;

&lt;p&gt;With tunefs.lustre, you can also erase the paramaters &amp;amp; restore them. However i don&apos;t think we need to do this here since we don&apos;t intend to change anything (like a nid).&lt;/p&gt;

&lt;p&gt;&amp;gt; If we do it as a mount option, I think I would need more info on how that works.&lt;/p&gt;

&lt;p&gt;It really works as i mentioned in my comment earlier, unmount everything and then mount the MDT with -o writeconf and then the OSTs with the same mount option.&lt;/p&gt;

&lt;p&gt;&amp;gt; How do the mgsnode and servicenode options get added doing it with as a mount option?&lt;/p&gt;

&lt;p&gt;Those parameters are removed if you use the --erase-params (e.g. when you want to change one of the parameters). In this case, i don&apos;t think we need to do this and we just want to run a simple writeconf.&lt;/p&gt;

&lt;p&gt;BTW, if you use OST pools, be aware that running writeconf erases all pools information on the MGS (as well as any other parameters set via lctl conf_param).&lt;/p&gt;</comment>
                            <comment id="24066" author="dnelson@ddn.com" created="Sat, 10 Dec 2011 18:02:27 +0000"  >&lt;p&gt;OK, Finally got the time scheduled to do the writeconf:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# pdsh -a modprobe lustre&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl set_param subsystem_debug=mgslnet.subsystem_debug=mgs&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl set_param debug=+config+info&lt;br/&gt;
lnet.debug=+config+info&lt;br/&gt;
error: set_param: writing to file /proc/sys/lnet/debug: Invalid argument&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl set_param debug=+config&lt;br/&gt;
lnet.debug=+config&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl set_param debug=+info&lt;br/&gt;
lnet.debug=+info&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl get_param debug&lt;br/&gt;
lnet.debug=info ioctl neterror warning error emerg ha config console&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl set_param debug_mb=500&lt;br/&gt;
lnet.debug_mb=500&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# vgchange -ay vg_scratch1&lt;br/&gt;
  2 logical volume(s) in volume group &quot;vg_scratch1&quot; now active&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# mount -t lustre -o writeconf /dev/vg_scratch1/mgs /lustre/mgs&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# mount -t lustre -o writeconf /dev/vg_scratch1/mdt /lustre/scratch1/mdt&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl dk &amp;gt; /tmp/log1&lt;br/&gt;
Ran script to mount all 176 OSTs with -o writeconf&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# lctl dk &amp;gt; /tmp/log2&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# pdsh -a umount -at lustre&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# df&lt;br/&gt;
Filesystem           1K-blocks      Used Available Use% Mounted on&lt;br/&gt;
/dev/sda1             16246428   3437164  11970676  23% /&lt;br/&gt;
/dev/mapper/VolGroup00-LogVol05&lt;br/&gt;
                     1857406056    200344 1761333100   1% /scratch&lt;br/&gt;
/dev/mapper/VolGroup00-LogVol04&lt;br/&gt;
                       4062912    139396   3713804   4% /home&lt;br/&gt;
/dev/mapper/VolGroup00-LogVol03&lt;br/&gt;
                       4062912    196308   3656892   6% /tmp&lt;br/&gt;
/dev/mapper/VolGroup00-LogVol02&lt;br/&gt;
                       8125880   1155932   6550520  15% /var&lt;br/&gt;
tmpfs                 37020324         0  37020324   0% /dev/shm&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# mount -t ldiskfs /dev/vg_scratch1/mgs /lustre/mgs&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# llog_reader /lustre/mgs/CONFIGS/scratch1-client &amp;gt; log.client&lt;/p&gt;

&lt;p&gt;After the writeconf, failover works &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-890&quot; title=&quot;MDS Failover Issue - Clients not reconnecting after MGT/MDT fail over to other MDS.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-890&quot;&gt;&lt;del&gt;LU-890&lt;/del&gt;&lt;/a&gt;, but I still cannot mount the fe2 client over ib2.  I am working on getting the log files.  Transferring files through the customer&apos;s bastion host is difficult.&lt;/p&gt;
</comment>
                            <comment id="24067" author="dnelson@ddn.com" created="Sat, 10 Dec 2011 18:11:29 +0000"  >&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib0(ib2)&quot;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# modprobe lustre&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.80.40@o2ib2&lt;br/&gt;
failed to ping 10.174.80.40@o2ib2: Input/output error&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# ping 10.174.80.40&lt;br/&gt;
PING 10.174.80.40 (10.174.80.40) 56(84) bytes of data.&lt;br/&gt;
64 bytes from 10.174.80.40: icmp_seq=1 ttl=64 time=3.53 ms&lt;br/&gt;
64 bytes from 10.174.80.40: icmp_seq=2 ttl=64 time=0.162 ms&lt;br/&gt;
^C
	&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
		&lt;li&gt;
		&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
			&lt;li&gt;10.174.80.40 ping statistics &amp;#8212;&lt;br/&gt;
2 packets transmitted, 2 received, 0% packet loss, time 1766ms&lt;br/&gt;
rtt min/avg/max/mdev = 0.162/1.850/3.538/1.688 ms&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# vi /etc/modprobe.d/lustre.conf &lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# lustre_rmmod&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/li&gt;
		&lt;/ul&gt;
		&lt;/li&gt;
	&lt;/ul&gt;
	&lt;/li&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib0(ib0)&quot;&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# modprobe lustre&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# mount -t lustre 10.174.31.241@o2ib0:10.174.31.251@o2ib0:/scratch1 /mnt/lustre1&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# ifconfig ib2&lt;br/&gt;
Ifconfig uses the ioctl access method to get the full address information, which limits hardware addresses to 8 bytes.&lt;br/&gt;
Because Infiniband address has 20 bytes, only the first 8 bytes are displayed correctly.&lt;br/&gt;
Ifconfig is obsolete! For replacement check ip.&lt;br/&gt;
ib2       Link encap:InfiniBand  HWaddr 80:00:00:48:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00  &lt;br/&gt;
          inet addr:10.174.81.10  Bcast:10.174.95.255  Mask:255.255.240.0&lt;br/&gt;
          inet6 addr: fe80::202:c903:f:aa67/64 Scope:Link&lt;br/&gt;
          UP BROADCAST RUNNING MULTICAST  MTU:65520  Metric:1&lt;br/&gt;
          RX packets:502712 errors:0 dropped:0 overruns:0 frame:0&lt;br/&gt;
          TX packets:7348 errors:0 dropped:0 overruns:0 carrier:0&lt;br/&gt;
          collisions:0 txqueuelen:256 &lt;br/&gt;
          RX bytes:41049681 (39.1 MiB)  TX bytes:1246560 (1.1 MiB)&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@fe1 ~&amp;#93;&lt;/span&gt;# df&lt;br/&gt;
Filesystem           1K-blocks      Used Available Use% Mounted on&lt;br/&gt;
/dev/sda6            972826992 401657664 522399748  44% /&lt;br/&gt;
tmpfs                 49532484         0  49532484   0% /dev/shm&lt;br/&gt;
/dev/sda5               142335     45735     89251  34% /boot&lt;br/&gt;
10.181.1.2:/apps/v1  504889344  40024672 464864672   8% /apps&lt;br/&gt;
10.181.1.2:/contrib  137625600   2996448 134629152   3% /contrib&lt;br/&gt;
10.181.1.2:/home     4315676672 458082464 3857594208  11% /home&lt;br/&gt;
10.174.31.241@o2ib0:10.174.31.251@o2ib0:/scratch1&lt;br/&gt;
                     2688660012544 14582672096 2647162648864   1% /mnt/lustre1&lt;/p&gt;</comment>
                            <comment id="24092" author="dnelson@ddn.com" created="Mon, 12 Dec 2011 11:34:40 +0000"  >&lt;p&gt;After some more testing, I think we have also resolved the issue mounting the filesystems on the login clients.&lt;/p&gt;

&lt;p&gt;I had tried with the following in /etc/modprobe.d/lustre.conf:&lt;br/&gt;
options lnet networks=&quot;o2ib0(ib2)&quot;&lt;/p&gt;

&lt;p&gt;What I have found is that if I set it to this:&lt;br/&gt;
options lnet networks=&quot;o2ib2(ib2)&quot;&lt;/p&gt;

&lt;p&gt;The mounts succeed.&lt;/p&gt;

&lt;p&gt;It appears that I have 1 problem remaining:&lt;/p&gt;

&lt;p&gt;I cannot mount the production filesystems on the TDS (Test and Development System) clients:&lt;/p&gt;

&lt;p&gt;The TDS clients mount the TDS Lustre filesystems over the client ib1 ports.  They are supposed to mount the production filesystems over ib0.  The ib2 ports of the production lustre servers are connected into the ib0 fabric of the TDS cluster.&lt;/p&gt;

&lt;p&gt;TDS Client (r1i3n15):&lt;br/&gt;
ib0          inet addr:10.174.64.65  Bcast:10.174.79.255  Mask:255.255.240.0&lt;br/&gt;
ib1          inet addr:10.174.96.65  Bcast:10.174.111.255  Mask:255.255.240.0&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib0(ib1), o2ib1(ib0)&quot;&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.96.65@o2ib&lt;br/&gt;
10.174.64.65@o2ib1&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.79.241@o2ib1&lt;br/&gt;
12345-0@lo&lt;br/&gt;
12345-10.174.31.241@o2ib&lt;br/&gt;
12345-10.174.79.241@o2ib1&lt;br/&gt;
12345-10.174.80.40@o2ib2&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/fstab&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;&amp;lt;file system&amp;gt; &amp;lt;mount point&amp;gt;   &amp;lt;type&amp;gt;  &amp;lt;options&amp;gt;       &amp;lt;dump&amp;gt;  &amp;lt;pass&amp;gt;&lt;br/&gt;
...&lt;br/&gt;
10.174.96.138@o2ib:/lustre1	/mnt/tds_lustre1	lustre	defaults,flock 0 0&lt;br/&gt;
10.174.96.138@o2ib:/lustre2	/mnt/tds_lustre2	lustre	defaults,flock 0 0&lt;br/&gt;
10.174.79.241@o2ib1:10.174.79.251@o2ib1:/scratch1 /mnt/lsc_lustre1 lustre defaults,flock 0 0&lt;br/&gt;
10.174.79.241@o2ib1:10.174.79.251@o2ib1:/scratch2 /mnt/lsc_lustre2 lustre defaults,flock 0 0&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;When I attempt to mount, the mount command simply hangs with with very little in the client log file:&lt;/p&gt;

&lt;p&gt;Dec 12 16:26:28 r1i3n15 kernel: Lustre: MGC10.174.79.241@o2ib1: Reactivating import&lt;br/&gt;
Dec 12 16:29:35 r1i3n15 kernel: Lustre: 4374:0:(import.c:517:import_select_connection()) scratch1-MDT0000-mdc-ffff880338c74c00: tried all connections, increasing latency to 7s&lt;br/&gt;
Dec 12 16:29:35 r1i3n15 kernel: Lustre: 4374:0:(import.c:517:import_select_connection()) Skipped 7 previous similar messages&lt;/p&gt;

&lt;p&gt;It appears to be communicating with the server because I initially, inadvertently used the wrong filesystem name and got this message:&lt;/p&gt;

&lt;p&gt;Dec 12 15:52:43 r1i3n15 kernel: Lustre: MGC10.174.79.241@o2ib1: Reactivating import&lt;br/&gt;
Dec 12 15:52:43 r1i3n15 kernel: LustreError: 156-2: The client profile &apos;lustre1-client&apos; could not be read from the MGS.  Does that filesystem exist?&lt;/p&gt;

&lt;p&gt;It correctly responded that the lustre1-client profile does not exist.&lt;/p&gt;

</comment>
                            <comment id="24104" author="cliffw" created="Mon, 12 Dec 2011 13:04:57 +0000"  >&lt;p&gt;since you have resolved the initial issue and this new problem is on a different set of servers, please close this bug and open up a new bug for the new issue. &lt;br/&gt;
We&apos;ll need to see the lnet config for the servers on the lustre1 and lustre2 filesystems. &lt;/p&gt;</comment>
                            <comment id="24110" author="dnelson@ddn.com" created="Mon, 12 Dec 2011 13:11:20 +0000"  >&lt;p&gt;I&apos;ll be glad to open a new ticket for this but it is the same set of servers and it was referenced in my initial post.&lt;/p&gt;</comment>
                            <comment id="24121" author="cliffw" created="Mon, 12 Dec 2011 13:32:36 +0000"  >&lt;p&gt;I am sorry I am a bit confused as to which servers are which.&lt;br/&gt;
These messages are telling you two different things:&lt;br/&gt;
---------&lt;br/&gt;
Dec 12 16:26:28 r1i3n15 kernel: Lustre: MGC10.174.79.241@o2ib1: Reactivating import&lt;br/&gt;
Dec 12 16:29:35 r1i3n15 kernel: Lustre: 4374:0:(import.c:517:import_select_connection()) scratch1-MDT0000-mdc-ffff880338c74c00: tried all connections, increasing latency to 7s&lt;br/&gt;
Dec 12 16:29:35 r1i3n15 kernel: Lustre: 4374:0:(import.c:517:import_select_connection()) Skipped 7 previous similar messages&lt;br/&gt;
------------&lt;br/&gt;
This is telling you the MDS cannot be located &lt;/p&gt;

&lt;p&gt;Dec 12 15:52:43 r1i3n15 kernel: Lustre: MGC10.174.79.241@o2ib1: Reactivating import&lt;br/&gt;
Dec 12 15:52:43 r1i3n15 kernel: LustreError: 156-2: The client profile &apos;lustre1-client&apos; could not be read from the MGS. &lt;br/&gt;
This is referring to the MGS. Since the lustre1-config is missing, this client isn&apos;t trying to find the MDS. Same server, but not the same service.&lt;br/&gt;
-------------&lt;br/&gt;
Can we get a new tunefs.lustre --print for scratch1 MGT and MDT, now that you&apos;ve done the writeconf?&lt;/p&gt;</comment>
                            <comment id="24218" author="dnelson@ddn.com" created="Mon, 12 Dec 2011 14:05:49 +0000"  >&lt;p&gt;I understand about being confused.  As I said, it is a very complex configuration.  I think it is going to be a nightmare to support.&lt;/p&gt;

&lt;p&gt;I threw in the error message of the lustre1-client simply because I had made a mistake and put lustre1 instead of scratch1.  To me, this indicates that it is communicating with the MGS since it was able to tell me that that the lustre1-client profile does not exist.  When I use the right filesystem name, the mount just hangs,&lt;/p&gt;

&lt;p&gt;Here is the tunefs.lustre --print information after the writeconf.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-2 ~&amp;#93;&lt;/span&gt;# tunefs.lustre --print /dev/vg_scratch1/mdt&lt;br/&gt;
checking for existing Lustre data: found CONFIGS/mountdata&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/p&gt;

&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     scratch1-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch1&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 mgsnode=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     scratch1-MDT0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  scratch1&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1401&lt;br/&gt;
              (MDT no_primnode )&lt;br/&gt;
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro&lt;br/&gt;
Parameters: mgsnode=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 mgsnode=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 failover.node=10.174.31.241@o2ib,10.174.79.241@o2ib1,10.174.80.40@o2ib2 failover.node=10.174.31.251@o2ib,10.174.79.251@o2ib1,10.174.80.41@o2ib2 mdt.quota_type=ug&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;</comment>
                            <comment id="24407" author="johann" created="Mon, 12 Dec 2011 17:16:18 +0000"  >&lt;p&gt;I looked at the new configuration log (i.e. file log.client) and the nid setup now looks fine:&lt;/p&gt;

&lt;p&gt;#06 (088)add_uuid  nid=10.174.31.241@o2ib(0x500000aae1ff1)  0:  1:10.174.31.241@o2ib  &lt;br/&gt;
#07 (088)add_uuid  nid=10.174.79.241@o2ib1(0x500010aae4ff1)  0:  1:10.174.31.241@o2ib  &lt;br/&gt;
#08 (088)add_uuid  nid=10.174.80.40@o2ib2(0x500020aae5028)  0:  1:10.174.31.241@o2ib  &lt;br/&gt;
#09 (136)attach    0:scratch1-MDT0000-mdc  1:mdc  2:scratch1-MDT0000-mdc_UUID  &lt;br/&gt;
#10 (144)setup     0:scratch1-MDT0000-mdc  1:scratch1-MDT0000_UUID  2:10.174.31.241@o2ib &lt;/p&gt;

&lt;p&gt;while in the previous file:&lt;/p&gt;

&lt;p&gt;#06 (088)add_uuid  nid=10.174.31.241@o2ib(0x500000aae1ff1)  0:  1:10.174.31.241@o2ib   &lt;br/&gt;
#07 (136)attach    0:scratch1-MDT0000-mdc  1:mdc  2:scratch1-MDT0000-mdc_UUID  &lt;br/&gt;
#08 (144)setup     0:scratch1-MDT0000-mdc  1:scratch1-MDT0000_UUID  2:10.174.31.241@o2ib  &lt;br/&gt;
#09 (088)add_uuid  nid=10.174.31.241@o2ib(0x500000aae1ff1)  0:  1:10.174.31.241@o2ib  &lt;br/&gt;
#10 (088)add_uuid  nid=10.174.79.241@o2ib(0x500000aae4ff1)  0:  1:10.174.31.241@o2ib  &lt;br/&gt;
#11 (088)add_uuid  nid=10.174.80.40@o2ib(0x500000aae5028)  0:  1:10.174.31.241@o2ib&lt;/p&gt;

&lt;p&gt;While the mount hangs, could you please try to collect import information (lctl get_param mdc.*.import) to check what nid we try to access?&lt;/p&gt;</comment>
                            <comment id="24446" author="dnelson@ddn.com" created="Mon, 12 Dec 2011 18:33:24 +0000"  >&lt;p&gt;Note; The filesystems listed as lustre1 and lustre2 are the TDS lustre filesystems, not production.  They use different servers.  The problem filesystems are scratch1 and scratch2.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl get_param mdc.*.import&lt;br/&gt;
mdc.lustre1-MDT0000-mdc-ffff88033cea4000.import=&lt;br/&gt;
import:&lt;br/&gt;
    name: lustre1-MDT0000-mdc-ffff88033cea4000&lt;br/&gt;
    target: lustre1-MDT0000_UUID&lt;br/&gt;
    state: FULL&lt;br/&gt;
    connect_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;version, inode_bit_locks, join_file, getattr_by_fid, no_oh_for_devices, early_lock_cancel, adaptive_timeouts, version_recovery, pools&amp;#93;&lt;/span&gt;&lt;br/&gt;
    import_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;replayable, pingable&amp;#93;&lt;/span&gt;&lt;br/&gt;
    connection:&lt;br/&gt;
       failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;10.174.96.138@o2ib&amp;#93;&lt;/span&gt;&lt;br/&gt;
       current_connection: 10.174.96.138@o2ib&lt;br/&gt;
       connection_attempts: 1&lt;br/&gt;
       generation: 1&lt;br/&gt;
       in-progress_invalidations: 0&lt;br/&gt;
    rpcs:&lt;br/&gt;
       inflight: 0&lt;br/&gt;
       unregistering: 0&lt;br/&gt;
       timeouts: 0&lt;br/&gt;
       avg_waittime: 235 usec&lt;br/&gt;
    service_estimates:&lt;br/&gt;
       services: 1 sec&lt;br/&gt;
       network: 1 sec&lt;br/&gt;
    transactions:&lt;br/&gt;
       last_replay: 0&lt;br/&gt;
       peer_committed: 0&lt;br/&gt;
       last_checked: 0&lt;br/&gt;
mdc.lustre2-MDT0000-mdc-ffff8803390ee800.import=&lt;br/&gt;
import:&lt;br/&gt;
    name: lustre2-MDT0000-mdc-ffff8803390ee800&lt;br/&gt;
    target: lustre2-MDT0000_UUID&lt;br/&gt;
    state: FULL&lt;br/&gt;
    connect_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;version, inode_bit_locks, join_file, getattr_by_fid, no_oh_for_devices, early_lock_cancel, adaptive_timeouts, version_recovery, pools&amp;#93;&lt;/span&gt;&lt;br/&gt;
    import_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;replayable, pingable&amp;#93;&lt;/span&gt;&lt;br/&gt;
    connection:&lt;br/&gt;
       failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;10.174.96.138@o2ib&amp;#93;&lt;/span&gt;&lt;br/&gt;
       current_connection: 10.174.96.138@o2ib&lt;br/&gt;
       connection_attempts: 1&lt;br/&gt;
       generation: 1&lt;br/&gt;
       in-progress_invalidations: 0&lt;br/&gt;
    rpcs:&lt;br/&gt;
       inflight: 0&lt;br/&gt;
       unregistering: 0&lt;br/&gt;
       timeouts: 0&lt;br/&gt;
       avg_waittime: 220 usec&lt;br/&gt;
    service_estimates:&lt;br/&gt;
       services: 1 sec&lt;br/&gt;
       network: 1 sec&lt;br/&gt;
    transactions:&lt;br/&gt;
       last_replay: 0&lt;br/&gt;
       peer_committed: 0&lt;br/&gt;
       last_checked: 0&lt;br/&gt;
mdc.scratch1-MDT0000-mdc-ffff88033b4f1400.import=&lt;br/&gt;
import:&lt;br/&gt;
    name: scratch1-MDT0000-mdc-ffff88033b4f1400&lt;br/&gt;
    target: scratch1-MDT0000_UUID&lt;br/&gt;
    state: DISCONN&lt;br/&gt;
    connect_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;version, acl, inode_bit_locks, join_file, getattr_by_fid, no_oh_for_devices, early_lock_cancel, adaptive_timeouts, lru_resize, fid_is_enabled, version_recovery, pools, 64bithash&amp;#93;&lt;/span&gt;&lt;br/&gt;
    import_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;replayable&amp;#93;&lt;/span&gt;&lt;br/&gt;
    connection:&lt;br/&gt;
       failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;10.174.31.241@o2ib, 10.174.31.251@o2ib&amp;#93;&lt;/span&gt;&lt;br/&gt;
       current_connection: 10.174.31.241@o2ib&lt;br/&gt;
       connection_attempts: 1&lt;br/&gt;
       generation: 1&lt;br/&gt;
       in-progress_invalidations: 0&lt;br/&gt;
    rpcs:&lt;br/&gt;
       inflight: 1&lt;br/&gt;
       unregistering: 0&lt;br/&gt;
       timeouts: 1&lt;br/&gt;
       avg_waittime: 0 usec&lt;br/&gt;
    service_estimates:&lt;br/&gt;
       services: 5 sec&lt;br/&gt;
       network: 0 sec&lt;br/&gt;
    transactions:&lt;br/&gt;
       last_replay: 0&lt;br/&gt;
       peer_committed: 0&lt;br/&gt;
       last_checked: 0&lt;/p&gt;

&lt;p&gt;I notice that it says that it is using 10.174.31.241@o2ib.  Currently, the MGS and the MDT are mounted on the other node (10.174.31..251).  Also, it just says o2ib not o2ib1?  Previously, I would not have worried about that but it seemed to make a difference on the production login clients (On the production clients, I tried defining the nid as o2ib0 and they would not mount, yet they did mount when the nid was defined as o2ib2).&lt;/p&gt;</comment>
                            <comment id="24494" author="johann" created="Mon, 12 Dec 2011 18:53:40 +0000"  >&lt;p&gt;So the failover_nids list looks good. The client tries to reach the MDS via 10.174.31.241@o2ib and it should then try through 10.174.31.251@o2ib.&lt;br/&gt;
As for o2ib/o2ib2, i think it only matters in the lnet module configuration, but i will have to ask Liang to confirm.&lt;/p&gt;

&lt;p&gt;Can you successfully ping 10.174.31.251@o2ib from r1i3n15?&lt;/p&gt;</comment>
                            <comment id="24555" author="dnelson@ddn.com" created="Mon, 12 Dec 2011 20:53:53 +0000"  >&lt;p&gt;Sorry, I made a mistake.  That is the problem.  It is trying to connect through the wrong nid.  Copying the original data from above:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.96.65@o2ib&lt;br/&gt;
10.174.64.65@o2ib1&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.79.241@o2ib1&lt;br/&gt;
12345-0@lo&lt;br/&gt;
12345-10.174.31.241@o2ib&lt;br/&gt;
12345-10.174.79.241@o2ib1&lt;br/&gt;
12345-10.174.80.40@o2ib2&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/fstab&lt;/p&gt;

&lt;p&gt;    &amp;lt;file system&amp;gt; &amp;lt;mount point&amp;gt; &amp;lt;type&amp;gt; &amp;lt;options&amp;gt; &amp;lt;dump&amp;gt; &amp;lt;pass&amp;gt;&lt;br/&gt;
    ...&lt;br/&gt;
    10.174.96.138@o2ib:/lustre1 /mnt/tds_lustre1 lustre defaults,flock 0 0&lt;br/&gt;
    10.174.96.138@o2ib:/lustre2 /mnt/tds_lustre2 lustre defaults,flock 0 0&lt;br/&gt;
    10.174.79.241@o2ib1:10.174.79.251@o2ib1:/scratch1 /mnt/lsc_lustre1 lustre defaults,flock 0 0&lt;br/&gt;
    10.174.79.241@o2ib1:10.174.79.251@o2ib1:/scratch2 /mnt/lsc_lustre2 lustre defaults,flock 0 0&lt;/p&gt;

&lt;p&gt;The only path to these servers from this client is through the 10.174.79.xxx addresses.&lt;/p&gt;

&lt;p&gt;Why is it trying 10.174.31.xxx?  There is no route for that subnet on these clients:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# netstat -rn&lt;br/&gt;
Kernel IP routing table&lt;br/&gt;
Destination     Gateway         Genmask         Flags   MSS Window  irtt Iface&lt;br/&gt;
10.181.1.0      10.174.64.67    255.255.255.0   UG        0 0          0 ib0&lt;br/&gt;
192.168.159.0   0.0.0.0         255.255.255.0   U         0 0          0 eth0&lt;br/&gt;
10.174.64.0     0.0.0.0         255.255.240.0   U         0 0          0 ib0&lt;br/&gt;
10.174.96.0     0.0.0.0         255.255.240.0   U         0 0          0 ib1&lt;br/&gt;
169.254.0.0     0.0.0.0         255.255.0.0     U         0 0          0 ib1&lt;br/&gt;
169.254.0.0     0.0.0.0         255.255.0.0     U         0 0          0 eth0&lt;/p&gt;</comment>
                            <comment id="24573" author="johann" created="Tue, 13 Dec 2011 04:34:32 +0000"  >&lt;p&gt;I have no idea why lnet selected 10.174.31.241@o2ib/10.174.31.251@o2ib instead of 10.174.79.241@o2ib/10.174.79.251@o2ib.&lt;br/&gt;
Liang, any idea?&lt;/p&gt;</comment>
                            <comment id="24653" author="liang" created="Tue, 13 Dec 2011 22:51:38 +0000"  >&lt;p&gt;Sorry I&apos;m a little confused about this setting, and have a few questions:&lt;/p&gt;

&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;r1i3n15 has two IB networks: 10.174.96.65@o2ib and 10.174.64.65@o2ib1, I think 10.174.96.65@o2ib is supposed to connect to production filesystem and 10.174.64.65@o2ib1 is supposed to connect to TDS filesystem, right?&lt;/li&gt;
&lt;/ul&gt;


&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;12345-10.174.31.241@o2ib and 12345-10.174.79.241@o2ib1 are NIDs of TDS MGS right?&lt;/li&gt;
&lt;/ul&gt;


&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;I know r1i3n15 can lctl ping 10.174.79.241@o2ib1 (second NID of TDS MGS), my question is: are you able to lctl ping 10.174.31.241@o2ib (first NID of TDS MGS), also, can you use regular ping to reach 10.174.31.241 from rli3n15?&lt;/li&gt;
&lt;/ul&gt;


&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;can you mount TDS filesystem if you disable this NID (10.174.96.65@o2ib) on rli3n15?&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Thanks&lt;br/&gt;
Liang&lt;/p&gt;</comment>
                            <comment id="24657" author="dnelson@ddn.com" created="Wed, 14 Dec 2011 06:04:32 +0000"  >&lt;p&gt;No, these clients cannot lctl ping, or ping, the 10.174.31.241 address.  That bid exists on the servers to support the scratch1 filesystems from the production clients.&lt;/p&gt;

&lt;p&gt;Yes, r1i3n15 can mount the TDS filesystem.&lt;/p&gt;</comment>
                            <comment id="24659" author="dnelson@ddn.com" created="Wed, 14 Dec 2011 07:34:11 +0000"  >&lt;p&gt;I realized that I did not answer one question.  There is only one MDS on the TDS filesystem and it has only one nid:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@mds01 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.96.138@o2ib&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# netstat -rn&lt;br/&gt;
Kernel IP routing table&lt;br/&gt;
Destination     Gateway         Genmask         Flags   MSS Window  irtt Iface&lt;br/&gt;
10.181.1.0      10.174.64.67    255.255.255.0   UG        0 0          0 ib0&lt;br/&gt;
192.168.159.0   0.0.0.0         255.255.255.0   U         0 0          0 eth0&lt;br/&gt;
10.174.64.0     0.0.0.0         255.255.240.0   U         0 0          0 ib0&lt;br/&gt;
10.174.96.0     0.0.0.0         255.255.240.0   U         0 0          0 ib1&lt;br/&gt;
169.254.0.0     0.0.0.0         255.255.0.0     U         0 0          0 ib1&lt;br/&gt;
169.254.0.0     0.0.0.0         255.255.0.0     U         0 0          0 eth0&lt;/p&gt;

&lt;p&gt;As you can see, there is no route to 10.174.31.241.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# ping 10.174.31.241&lt;br/&gt;
connect: Network is unreachable&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib(ib1)&quot;&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# df&lt;br/&gt;
Filesystem           1K-blocks      Used Available Use% Mounted on&lt;br/&gt;
tmpfs                   153600        60    153540   1% /tmp&lt;br/&gt;
...&lt;br/&gt;
10.174.96.138@o2ib:/lustre1&lt;br/&gt;
                     30523086656 2401014176 27816223672   8% /mnt/tds_lustre1&lt;br/&gt;
10.174.96.138@o2ib:/lustre2&lt;br/&gt;
                     30523086656 268760500 29948475720   1% /mnt/tds_lustre2&lt;/p&gt;

&lt;p&gt;If I unmount the TDS filesystems and change the modprobe.d/lustre.conf file to only include the ib0 port:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib(ib0)&quot;&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;I cannot communicate with the MDS.  I get this error:&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.64.65@o2ib&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/fstab&lt;br/&gt;
...&lt;br/&gt;
10.174.79.241@o2ib:10.174.79.251@o2ib:/scratch1 /mnt/lsc_lustre1 lustre defaults,flock 0 0&lt;br/&gt;
10.174.79.242@o2ib:10.174.79.252@o2ib:/scratch2 /mnt/lsc_lustre2 lustre defaults,flock 0 0&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# mount -at lustre&lt;br/&gt;
mount.lustre: mount 10.174.79.241@o2ib:10.174.79.251@o2ib:/scratch1 at /mnt/lsc_lustre1 failed: Cannot send after transport endpoint shutdown&lt;br/&gt;
mount.lustre: mount 10.174.79.242@o2ib:10.174.79.252@o2ib:/scratch2 at /mnt/lsc_lustre2 failed: Cannot send after transport endpoint shutdown&lt;/p&gt;

&lt;p&gt;Dec 14 12:17:23 r1i3n15 kernel: Lustre: 27084:0:(client.c:1487:ptlrpc_expire_one_request()) @@@ Request x1388172991791113 sent from MGC10.174.79.241@o2ib to NID 10.174.79.241@o2ib 0s ago has failed due to network error (5s prior to deadline).&lt;br/&gt;
Dec 14 12:17:23 r1i3n15 kernel:  req@ffff880639bf3400 x1388172991791113/t0 o250-&amp;gt;MGS@MGC10.174.79.241@o2ib_0:26/25 lens 368/584 e 0 to 1 dl 1323865048 ref 1 fl Rpc:N/0/0 rc 0/0&lt;br/&gt;
Dec 14 12:17:23 r1i3n15 kernel: LustreError: 1280:0:(o2iblnd_cb.c:2532:kiblnd_rejected()) 10.174.79.241@o2ib rejected: o2iblnd fatal error&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel: Lustre: 27084:0:(client.c:1487:ptlrpc_expire_one_request()) @@@ Request x1388172991791115 sent from MGC10.174.79.241@o2ib to NID 10.174.79.251@o2ib 0s ago has failed due to network error (5s prior to deadline).&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel:  req@ffff880639347c00 x1388172991791115/t0 o250-&amp;gt;MGS@MGC10.174.79.241@o2ib_1:26/25 lens 368/584 e 0 to 1 dl 1323865073 ref 1 fl Rpc:N/0/0 rc 0/0&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel: LustreError: 27292:0:(client.c:858:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff880639bf3000 x1388172991791116/t0 o501-&amp;gt;MGS@MGC10.174.79.241@o2ib_1:26/25 lens 264/432 e 0 to 1 dl 0 ref 1 fl Rpc:/0/0 rc 0/0&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel: LustreError: 1280:0:(o2iblnd_cb.c:2532:kiblnd_rejected()) 10.174.79.251@o2ib rejected: o2iblnd fatal error&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel: LustreError: 15c-8: MGC10.174.79.241@o2ib: The configuration from log &apos;scratch1-client&apos; failed (-108). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel: LustreError: 27292:0:(llite_lib.c:1095:ll_fill_super()) Unable to process log: -108&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel: Lustre: client ffff8803386abc00 umount complete&lt;br/&gt;
Dec 14 12:17:48 r1i3n15 kernel: LustreError: 27292:0:(obd_mount.c:2065:lustre_fill_super()) Unable to mount  (-108)&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: Lustre: 27084:0:(client.c:1487:ptlrpc_expire_one_request()) @@@ Request x1388172991791119 sent from MGC10.174.79.242@o2ib to NID 10.174.79.252@o2ib 0s ago has failed due to network error (5s prior to deadline).&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel:  req@ffff88063be32400 x1388172991791119/t0 o250-&amp;gt;MGS@MGC10.174.79.242@o2ib_1:26/25 lens 368/584 e 0 to 1 dl 1323865098 ref 1 fl Rpc:N/0/0 rc 0/0&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: Lustre: 27084:0:(client.c:1487:ptlrpc_expire_one_request()) Skipped 1 previous similar message&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: LustreError: 27343:0:(client.c:858:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff8803247b2000 x1388172991791120/t0 o501-&amp;gt;MGS@MGC10.174.79.242@o2ib_1:26/25 lens 264/432 e 0 to 1 dl 0 ref 1 fl Rpc:/0/0 rc 0/0&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: LustreError: 15c-8: MGC10.174.79.242@o2ib: The configuration from log &apos;scratch2-client&apos; failed (-108). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: LustreError: 27343:0:(llite_lib.c:1095:ll_fill_super()) Unable to process log: -108&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: LustreError: 1293:0:(o2iblnd_cb.c:2532:kiblnd_rejected()) 10.174.79.252@o2ib rejected: o2iblnd fatal error&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: Lustre: client ffff880322a0c400 umount complete&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: LustreError: 27343:0:(obd_mount.c:2065:lustre_fill_super()) Unable to mount  (-108)&lt;br/&gt;
Dec 14 12:18:13 r1i3n15 kernel: LustreError: 1293:0:(o2iblnd_cb.c:2532:kiblnd_rejected()) Skipped 1 previous similar message&lt;/p&gt;

&lt;p&gt;From what I can see, there is no indication of a network problem:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# ibstat&lt;br/&gt;
CA &apos;mlx4_0&apos;&lt;br/&gt;
	CA type: MT26428&lt;br/&gt;
	Number of ports: 2&lt;br/&gt;
	Firmware version: 2.9.1000&lt;br/&gt;
	Hardware version: b0&lt;br/&gt;
	Node GUID: 0x0002c9030010c5f4&lt;br/&gt;
	System image GUID: 0x0002c9030010c5f7&lt;br/&gt;
	Port 1:&lt;br/&gt;
		State: Active&lt;br/&gt;
		Physical state: LinkUp&lt;br/&gt;
		Rate: 40&lt;br/&gt;
		Base lid: 152&lt;br/&gt;
		LMC: 0&lt;br/&gt;
		SM lid: 1&lt;br/&gt;
		Capability mask: 0x02510868&lt;br/&gt;
		Port GUID: 0x0002c9030010c5f5&lt;br/&gt;
		Link layer: IB&lt;br/&gt;
	Port 2:&lt;br/&gt;
		State: Active&lt;br/&gt;
		Physical state: LinkUp&lt;br/&gt;
		Rate: 40&lt;br/&gt;
		Base lid: 5232&lt;br/&gt;
		LMC: 0&lt;br/&gt;
		SM lid: 2&lt;br/&gt;
		Capability mask: 0x02510868&lt;br/&gt;
		Port GUID: 0x0002c9030010c5f6&lt;br/&gt;
		Link layer: IB&lt;br/&gt;
CA &apos;mlx4_1&apos;&lt;br/&gt;
	CA type: MT26428&lt;br/&gt;
	Number of ports: 2&lt;br/&gt;
	Firmware version: 2.9.1000&lt;br/&gt;
	Hardware version: b0&lt;br/&gt;
	Node GUID: 0x0002c9030010c6b0&lt;br/&gt;
	System image GUID: 0x0002c9030010c6b3&lt;br/&gt;
	Port 1:&lt;br/&gt;
		State: Active&lt;br/&gt;
		Physical state: LinkUp&lt;br/&gt;
		Rate: 40&lt;br/&gt;
		Base lid: 104&lt;br/&gt;
		LMC: 0&lt;br/&gt;
		SM lid: 106&lt;br/&gt;
		Capability mask: 0x02510868&lt;br/&gt;
		Port GUID: 0x0002c9030010c6b1&lt;br/&gt;
		Link layer: IB&lt;br/&gt;
	Port 2:&lt;br/&gt;
		State: Active&lt;br/&gt;
		Physical state: LinkUp&lt;br/&gt;
		Rate: 40&lt;br/&gt;
		Base lid: 64&lt;br/&gt;
		LMC: 0&lt;br/&gt;
		SM lid: 1&lt;br/&gt;
		Capability mask: 0x02510868&lt;br/&gt;
		Port GUID: 0x0002c9030010c6b2&lt;br/&gt;
		Link layer: IB&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@lfs-mds-1-1 ~&amp;#93;&lt;/span&gt;# ibping -C mlx4_1 -P 1 -S&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# ibping -G 0x0002c9030010c6b1&lt;br/&gt;
Pong from lfs-mds-1-1.(none) (Lid 104): time 0.156 ms&lt;br/&gt;
Pong from lfs-mds-1-1.(none) (Lid 104): time 0.154 ms&lt;br/&gt;
Pong from lfs-mds-1-1.(none) (Lid 104): time 0.137 ms&lt;br/&gt;
Pong from lfs-mds-1-1.(none) (Lid 104): time 0.134 ms&lt;br/&gt;
Pong from lfs-mds-1-1.(none) (Lid 104): time 0.056 ms&lt;br/&gt;
Pong from lfs-mds-1-1.(none) (Lid 104): time 0.131 ms&lt;br/&gt;
^C&lt;br/&gt;
&amp;#8212; lfs-mds-1-1.(none) (Lid 104) ibping statistics &amp;#8212;&lt;br/&gt;
6 packets transmitted, 6 received, 0% packet loss, time 5123 ms&lt;br/&gt;
rtt min/avg/max = 0.056/0.128/0.156 ms&lt;/p&gt;

&lt;p&gt;Yet, lctl ping fails:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl ping 10.174.79.241@o2ib&lt;br/&gt;
failed to ping 10.174.79.241@o2ib: Input/output error&lt;/p&gt;

&lt;p&gt;If I go back to the original configuration:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib(ib1), o2ib1(ib0)&quot;&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.96.65@o2ib&lt;br/&gt;
10.174.64.65@o2ib1&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/fstab&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;&amp;lt;file system&amp;gt; &amp;lt;mount point&amp;gt;   &amp;lt;type&amp;gt;  &amp;lt;options&amp;gt;       &amp;lt;dump&amp;gt;  &amp;lt;pass&amp;gt;&lt;br/&gt;
...&lt;br/&gt;
10.174.96.138@o2ib:/lustre1	/mnt/tds_lustre1	lustre	defaults,flock 0 0&lt;br/&gt;
10.174.96.138@o2ib:/lustre2	/mnt/tds_lustre2	lustre	defaults,flock 0 0&lt;br/&gt;
10.174.79.241@o2ib1:10.174.79.251@o2ib1:/scratch1 /mnt/lsc_lustre1 lustre defaults,flock 0 0&lt;br/&gt;
10.174.79.242@o2ib1:10.174.79.252@o2ib1:/scratch2 /mnt/lsc_lustre2 lustre defaults,flock 0 0&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;The TDS filesystems mount (lustre1, lustre2) and the production filesystems (scratch1, scratch2) just hang while performing the mount.&lt;/p&gt;
</comment>
                            <comment id="24690" author="liang" created="Wed, 14 Dec 2011 08:38:25 +0000"  >&lt;p&gt;Here is my undestanding about your setting, please correct me if I was wrong:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt; 
client                      TDS MDS                    Production MDS
---------                   ---------                  -------
rli3n15                     mds01                      lfs-mds-1-1 (scratch1)
10.174.96.64@o2ib0(ib1)     10.174.96.138@o2ib0 [y]    10.174.31.241@o2ib0 [n]
10.174.64.65@o2ib1(ib0)                                10.174.79.241@o2ib1 [y]

[y] == [yes], means we can reach that NID via &quot;lctl ping&quot; from rli3n15
[n] == [no],  means we can not reach that NID via &quot;lctl ping&quot; from rli3n15

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; 
&lt;p&gt;So between rli3n15 and lfs-mds-1-1:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;10.174.64.65@o2ib1(ib0) and 10.174.79.241@o2ib1 are on the same LNet network,and they are physically reachable to each other&lt;/li&gt;
	&lt;li&gt;10.174.96.64@o2ib0(ib1) and 10.174.31.241@o2ib0 are on the same LNet network,&lt;br/&gt;
but they are physically unreachable to each other&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;I think if you try to mount scratch1 from rli3n15, it will firstly look at all N&lt;br/&gt;
IDs of lfs-mds-1-1, and it found both itself and lfs-mds-1-1 have two local NIDs on o2ib0 and o2ib1 (although they can&apos;t reach eath other on o2ib0), and LNet hop of these two NIDs are same and both interfaces are healthy, so ptlrpc will choose the first NID of lfs-mds-1-1 10.174.31.241@o2ib0, which is actually unreachable to rli3n15.&lt;/p&gt;

&lt;p&gt;I would suggest to try with this one rli3n15:&lt;br/&gt;
options lnet networks=&quot;o2ib1(ib0)&quot;&lt;/p&gt;

&lt;p&gt;and try to mount scratch1,2, if it can work, I would suggest to use configuratio&lt;br/&gt;
n like this:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt; 
client                      TDS MDS                    Production MDS
---------                   ---------                  -------
rli3n15                     mds01                      lfs-mds-1-1 (scratch1)
10.174.96.64@o2ib3(ib1)     10.174.96.138@o2ib3 [y]    
10.174.64.65@o2ib1(ib0)                                10.174.79.241@o2ib1 [y]
                                                       10.174.31.241@o2ib0 [y]

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; 
&lt;p&gt;The only change we made here is:&lt;br/&gt;
o2ib0 on rli3n15 and mds01 is replaced by o2ib3, of course, if it can work you will have to change all nodes on TDS to o2ib3...&lt;/p&gt;</comment>
                            <comment id="24700" author="dnelson@ddn.com" created="Wed, 14 Dec 2011 08:54:48 +0000"  >&lt;p&gt;OK, I tried the following:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/modprobe.d/lustre.conf &lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;Lustre module configuration file&lt;br/&gt;
options lnet networks=&quot;o2ib3(ib1), o2ib1(ib0)&quot;&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# lctl list_nids&lt;br/&gt;
10.174.96.65@o2ib3&lt;br/&gt;
10.174.64.65@o2ib1&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# cat /etc/fstab&lt;br/&gt;
...&lt;br/&gt;
10.174.96.138@o2ib3:/lustre1	/mnt/tds_lustre1	lustre	defaults,flock 0 0&lt;br/&gt;
10.174.96.138@o2ib3:/lustre2	/mnt/tds_lustre2	lustre	defaults,flock 0 0&lt;br/&gt;
10.174.79.241@o2ib1:10.174.79.251@o2ib1:/scratch1 /mnt/lsc_lustre1 lustre defaults,flock 0 0&lt;br/&gt;
10.174.79.242@o2ib1:10.174.79.252@o2ib1:/scratch2 /mnt/lsc_lustre2 lustre defaults,flock 0 0&lt;/p&gt;

&lt;p&gt;Now, the production filesystems (scratch1, scratch2) mount and the TDS filesystems fail to mount.&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# mount -at lustre&lt;br/&gt;
mount.lustre: mount 10.174.96.138@o2ib3:/lustre1 at /mnt/tds_lustre1 failed: Cannot send after transport endpoint shutdown&lt;br/&gt;
mount.lustre: mount 10.174.96.138@o2ib3:/lustre2 at /mnt/tds_lustre2 failed: File exists&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;root@r1i3n15 ~&amp;#93;&lt;/span&gt;# df&lt;br/&gt;
Filesystem           1K-blocks      Used Available Use% Mounted on&lt;br/&gt;
tmpfs                   153600      1708    151892   2% /tmp&lt;br/&gt;
10.181.1.2:/contrib  137625600   3002528 134623072   3% /contrib&lt;br/&gt;
10.181.1.2:/testapps/v1&lt;br/&gt;
                      45875200  35991488   9883712  79% /apps&lt;br/&gt;
10.181.1.2:/testhome 550764544 166799968 383964576  31% /home&lt;br/&gt;
10.174.79.241@o2ib1:10.174.79.251@o2ib1:/scratch1&lt;br/&gt;
                     2688660012544 29627611556 2632114228424   2% /mnt/lsc_lustre1&lt;br/&gt;
10.174.79.242@o2ib1:10.174.79.252@o2ib1:/scratch2&lt;br/&gt;
                     3360825015680 785492156 3326396150596   1% /mnt/lsc_lustre2&lt;/p&gt;</comment>
                            <comment id="24707" author="liang" created="Wed, 14 Dec 2011 09:44:33 +0000"  >&lt;p&gt;have you also changed MDS/MGS and other servers in TDS filesystem to o2ib3 as well (i.e: mds01)? Because you are using o2ib3 as TDS network number, so all clients and servers on TDS network should use that network number (o2ib3).&lt;br/&gt;
Also, try &quot;lctl ping&quot; to verify network is reachable is always a good idea, &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/wink.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;</comment>
                            <comment id="24708" author="dnelson@ddn.com" created="Wed, 14 Dec 2011 09:47:25 +0000"  >&lt;p&gt;Ah, no.  I will have to schedule some time with the customer to do that.  I have one node that is not currently in the job queue that I can use for testing.  To take the whole filesystem down, I will have to schedule it.&lt;/p&gt;

&lt;p&gt;I will get that scheduled today.&lt;/p&gt;</comment>
                            <comment id="24747" author="dnelson@ddn.com" created="Wed, 14 Dec 2011 14:38:14 +0000"  >&lt;p&gt;I made the change on the TDS servers and had to perform a writeconf in order to get it mounted up again.  Everything seems to be working now.&lt;/p&gt;

&lt;p&gt;Thank you very much for all of your help!&lt;/p&gt;</comment>
                            <comment id="24748" author="pjones" created="Wed, 14 Dec 2011 14:47:51 +0000"  >&lt;p&gt;Dennis&lt;/p&gt;

&lt;p&gt;Thanks for the update. So can we close both this ticket and LU890?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="24750" author="dnelson@ddn.com" created="Wed, 14 Dec 2011 14:51:46 +0000"  >&lt;p&gt;Yes.  I already suggested that &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-890&quot; title=&quot;MDS Failover Issue - Clients not reconnecting after MGT/MDT fail over to other MDS.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-890&quot;&gt;&lt;del&gt;LU-890&lt;/del&gt;&lt;/a&gt; be closed and it was closed by Cliff.  This one can be also.&lt;/p&gt;</comment>
                            <comment id="24752" author="pjones" created="Wed, 14 Dec 2011 15:07:34 +0000"  >&lt;p&gt;Great - thanks Dennis!&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="10650" name="fe2.log" size="9012" author="dnelson@ddn.com" created="Mon, 5 Dec 2011 16:28:01 +0000"/>
                            <attachment id="10668" name="log.client" size="248568" author="dnelson@ddn.com" created="Sat, 10 Dec 2011 18:05:25 +0000"/>
                            <attachment id="10666" name="log1" size="89921" author="dnelson@ddn.com" created="Sat, 10 Dec 2011 18:05:25 +0000"/>
                            <attachment id="10667" name="log2" size="6031085" author="dnelson@ddn.com" created="Sat, 10 Dec 2011 18:05:25 +0000"/>
                            <attachment id="10654" name="lustre-scratch1" size="845463" author="dnelson@ddn.com" created="Wed, 7 Dec 2011 11:06:53 +0000"/>
                            <attachment id="10658" name="lustre1_uuids.txt" size="142241" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 10:44:43 +0000"/>
                            <attachment id="10659" name="lustre2_uuids.txt" size="355813" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 10:44:43 +0000"/>
                            <attachment id="10662" name="scratch1.log" size="248385" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 14:35:57 +0000"/>
                            <attachment id="10663" name="scratch2.log" size="626592" author="dnelson@ddn.com" created="Thu, 8 Dec 2011 14:35:57 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvhnb:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>6508</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>