<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:45:31 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4749] ZFS-backed OST  mkfs.lustre --servicenode does not correctly add failover_nids</title>
                <link>https://jira.whamcloud.com/browse/LU-4749</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;When creating ZFS-backed OSTs using the --servicenode syntax, only one failover nids is stored.&lt;/p&gt;

&lt;p&gt;mkfs.lustre --ost --index=1 --fsname=saturn --backfstype=zfs --mgsnode=192.168.122.73@tcp --servicenode=192.168.122.76@tcp --servicenode=192.168.122.78@tcp lsrv3/saturn-ost1&lt;/p&gt;

&lt;ol&gt;
	&lt;li&gt;tunefs.lustre --print lsrv3/saturn-ost1&lt;br/&gt;
checking for existing Lustre data: found&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     saturn-OST0001&lt;br/&gt;
Index:      1&lt;br/&gt;
Lustre FS:  saturn&lt;br/&gt;
Mount type: zfs&lt;br/&gt;
Flags:      0x1002&lt;br/&gt;
              (OST no_primnode )&lt;br/&gt;
Persistent mount opts: &lt;br/&gt;
Parameters: failover.node=192.168.122.78@tcp mgsnode=192.168.122.73@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     saturn-OST0001&lt;br/&gt;
Index:      1&lt;br/&gt;
Lustre FS:  saturn&lt;br/&gt;
Mount type: zfs&lt;br/&gt;
Flags:      0x1002&lt;br/&gt;
              (OST no_primnode )&lt;br/&gt;
Persistent mount opts: &lt;br/&gt;
Parameters: failover.node=192.168.122.78@tcp mgsnode=192.168.122.73@tcp&lt;/p&gt;


&lt;p&gt;On MGS:&lt;/p&gt;

&lt;ol&gt;
	&lt;li&gt;lctl get_param  osp.saturn-OST0001-osc-MDT0000.import&lt;br/&gt;
osp.saturn-OST0001-osc-MDT0000.import=&lt;br/&gt;
import:&lt;br/&gt;
    name: saturn-OST0001-osc-MDT0000&lt;br/&gt;
    target: saturn-OST0001_UUID&lt;br/&gt;
    state: FULL&lt;br/&gt;
    instance: 1&lt;br/&gt;
    connect_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;lov_index, unused, version, request_portal, adaptive_timeouts, lru_resize, fid_is_enabled, skip_orphan, full20, lvb_type&amp;#93;&lt;/span&gt;&lt;br/&gt;
    import_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;replayable, pingable&amp;#93;&lt;/span&gt;&lt;br/&gt;
    connection:&lt;br/&gt;
       failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;192.168.122.78@tcp&amp;#93;&lt;/span&gt;&lt;br/&gt;
       current_connection: 192.168.122.78@tcp&lt;/li&gt;
&lt;/ol&gt;



&lt;p&gt;For a ldiskfs-backed OST, two NIDs are stored:&lt;/p&gt;
&lt;ol&gt;
	&lt;li&gt;mkfs.lustre --ost --index=0 --fsname=saturn --servicenode=192.168.122.76@tcp --mgsnode=192.168.122.73@tcp --reformat /dev/vdb&lt;/li&gt;
&lt;/ol&gt;


&lt;ol&gt;
	&lt;li&gt;tunefs.lustre --print /dev/vdb&lt;br/&gt;
checking for existing Lustre data: found&lt;br/&gt;
Reading CONFIGS/mountdata&lt;/li&gt;
&lt;/ol&gt;


&lt;p&gt;   Read previous values:&lt;br/&gt;
Target:     saturn-OST0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  saturn&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1002&lt;br/&gt;
              (OST no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro&lt;br/&gt;
Parameters: failover.node=192.168.122.76@tcp failover.node=192.168.122.78@tcp mgsnode=192.168.122.73@tcp&lt;/p&gt;


&lt;p&gt;   Permanent disk data:&lt;br/&gt;
Target:     saturn-OST0000&lt;br/&gt;
Index:      0&lt;br/&gt;
Lustre FS:  saturn&lt;br/&gt;
Mount type: ldiskfs&lt;br/&gt;
Flags:      0x1002&lt;br/&gt;
              (OST no_primnode )&lt;br/&gt;
Persistent mount opts: errors=remount-ro&lt;br/&gt;
Parameters: failover.node=192.168.122.76@tcp failover.node=192.168.122.78@tcp mgsnode=192.168.122.73@tcp&lt;/p&gt;

&lt;p&gt;exiting before disk write.&lt;/p&gt;

&lt;ol&gt;
	&lt;li&gt;lctl get_param  osp.saturn-OST0000-osc-MDT0000.import&lt;br/&gt;
osp.saturn-OST0000-osc-MDT0000.import=&lt;br/&gt;
import:&lt;br/&gt;
    name: saturn-OST0000-osc-MDT0000&lt;br/&gt;
    target: saturn-OST0000_UUID&lt;br/&gt;
    state: FULL&lt;br/&gt;
    instance: 2&lt;br/&gt;
    connect_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;lov_index, unused, version, request_portal, adaptive_timeouts, lru_resize, fid_is_enabled, skip_orphan, full20, lvb_type&amp;#93;&lt;/span&gt;&lt;br/&gt;
    import_flags: &lt;span class=&quot;error&quot;&gt;&amp;#91;replayable, pingable&amp;#93;&lt;/span&gt;&lt;br/&gt;
    connection:&lt;br/&gt;
       failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;192.168.122.76@tcp, 192.168.122.78@tcp&amp;#93;&lt;/span&gt;&lt;br/&gt;
       current_connection: 192.168.122.76@tcp&lt;/li&gt;
&lt;/ol&gt;
</description>
                <environment>CentOS 6.4,  ZFS  0.6.2</environment>
        <key id="23591">LU-4749</key>
            <summary>ZFS-backed OST  mkfs.lustre --servicenode does not correctly add failover_nids</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="liwei">Li Wei</assignee>
                                    <reporter username="aalba6675">Anthony Alba</reporter>
                        <labels>
                            <label>prz</label>
                            <label>zfs</label>
                    </labels>
                <created>Tue, 11 Mar 2014 14:03:20 +0000</created>
                <updated>Mon, 27 Apr 2015 20:17:51 +0000</updated>
                            <resolved>Thu, 9 Oct 2014 16:24:43 +0000</resolved>
                                    <version>Lustre 2.4.2</version>
                    <version>Lustre 2.7.0</version>
                                    <fixVersion>Lustre 2.7.0</fixVersion>
                    <fixVersion>Lustre 2.5.4</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="78992" author="aalba6675" created="Tue, 11 Mar 2014 14:05:28 +0000"  >&lt;p&gt;For the ldiskfs I omitted adding the failover NID:&lt;/p&gt;

&lt;ol&gt;
	&lt;li&gt;tunefs.lustre  --erase-params --servicenode=192.168.122.76@tcp --servicenode=192.168.122.78@tcp --mgsnode=192.168.122.73@tcp /dev/vdb&lt;/li&gt;
&lt;/ol&gt;
</comment>
                            <comment id="78993" author="aalba6675" created="Tue, 11 Mar 2014 14:08:23 +0000"  >&lt;p&gt;1. A second oddity: I think --mgsnode=ABCD --mgsnode=XYZW also doesn&apos;t work on 2.4.2/ZFS-backed OSTs. The 2nd mgsnode overrides the first. For LDISKFS-backed OSTs it seems to work.&lt;/p&gt;

&lt;p&gt;2. Does the syntax --mgsnode=Pri_NID:Sec_NID work for mkfs.lustre or should one be using&lt;br/&gt;
 --mgsnode=Pri_NID --mgsnode=Sec_NID&lt;/p&gt;
</comment>
                            <comment id="92029" author="jlevi" created="Wed, 20 Aug 2014 12:40:08 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/11161&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/11161&lt;/a&gt; will fix this issue&lt;/p&gt;</comment>
                            <comment id="92638" author="isaac" created="Wed, 27 Aug 2014 18:22:59 +0000"  >&lt;p&gt;This looks like a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4334&quot; title=&quot;With ZFS can only declare a single mgsnode for MDT or OST&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4334&quot;&gt;&lt;del&gt;LU-4334&lt;/del&gt;&lt;/a&gt;. Both tunefs.lustre and mkfs.lustre are built from the same sources: mkfs_lustre.c mount_utils.c mount_utils.h.&lt;/p&gt;</comment>
                            <comment id="94216" author="liwei" created="Wed, 17 Sep 2014 05:59:44 +0000"  >&lt;p&gt;I was waiting for the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4334&quot; title=&quot;With ZFS can only declare a single mgsnode for MDT or OST&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4334&quot;&gt;&lt;del&gt;LU-4334&lt;/del&gt;&lt;/a&gt; patch to fix this, but that patch turned out to insufficient.  Current mgs cannot handle colons in failover.node values.  I&apos;ll submit my patch soon.&lt;/p&gt;</comment>
                            <comment id="94222" author="liwei" created="Wed, 17 Sep 2014 07:39:04 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/11956&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/11956&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="94841" author="jlevi" created="Wed, 24 Sep 2014 15:48:22 +0000"  >&lt;p&gt;Patch landed to Master&lt;/p&gt;</comment>
                            <comment id="95724" author="bogl" created="Mon, 6 Oct 2014 17:41:27 +0000"  >&lt;p&gt;backport to b2_5:&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/12196&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/12196&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="95740" author="green" created="Mon, 6 Oct 2014 19:26:35 +0000"  >&lt;p&gt;I think this patch causes failures in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5706&quot; title=&quot;conf-sanity test_57a: @@@@@@ FAIL: OST registration from failnode should fail &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5706&quot;&gt;&lt;del&gt;LU-5706&lt;/del&gt;&lt;/a&gt;, e.g.: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/6c3961a2-4aa8-11e4-95b1-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/6c3961a2-4aa8-11e4-95b1-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Once this was cherrypicked to b2_5 as a separate patch, it started to hit.&lt;br/&gt;
Might be it was a fluke, but reverting it made the problem disappear.&lt;br/&gt;
&lt;a href=&quot;http://review.whamcloud.com/#/c/12166/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/12166/&lt;/a&gt; vs &lt;a href=&quot;http://review.whamcloud.com/#/c/12183/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/12183/&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="95741" author="green" created="Mon, 6 Oct 2014 19:31:15 +0000"  >&lt;p&gt;Also looking in maloo results, it&apos;s really end of Sptember where these problems started to appear, and before that all failures were in 2013.&lt;/p&gt;

&lt;p&gt;So i think chances are high this is the culprit.&lt;/p&gt;</comment>
                            <comment id="95771" author="adilger" created="Tue, 7 Oct 2014 07:08:45 +0000"  >&lt;p&gt;Reopen due to potential problems with the patch.&lt;/p&gt;</comment>
                            <comment id="95772" author="adilger" created="Tue, 7 Oct 2014 07:26:13 +0000"  >&lt;p&gt;It looks like there is some garbage being written into the ZFS properties.  From the test log output of &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a33da7e2-4a9b-11e4-adcb-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a33da7e2-4a9b-11e4-adcb-5254006e85c2&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;   Permanent disk data:
Target:     lustre-OST0000
Index:      0
Lustre FS:  lustre
Mount type: zfs
Flags:      0x42
              (OST update )
Persistent mount opts: 
Parameters: sys.timeout=20 mgsnode=10.1.5.243@tcp failover.node=��&#65533;&#65533;6

Writing lustre-ost1/ost1 properties
  lustre:version=1
  lustre:flags=66
  lustre:index=0
  lustre:fsname=lustre
  lustre:svname=lustre-OST0000
  lustre:sys.timeout=20
  lustre:mgsnode=10.1.5.243@tcp
  lustre:failover.node=��&#65533;&#65533;6
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="95787" author="liwei" created="Tue, 7 Oct 2014 12:44:07 +0000"  >&lt;p&gt;This is not the cause of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5706&quot; title=&quot;conf-sanity test_57a: @@@@@@ FAIL: OST registration from failnode should fail &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5706&quot;&gt;&lt;del&gt;LU-5706&lt;/del&gt;&lt;/a&gt;; please see my comments there.&lt;/p&gt;</comment>
                            <comment id="95981" author="liwei" created="Thu, 9 Oct 2014 00:30:39 +0000"  >&lt;p&gt;I think this should be either closed or left open for Bob&apos;s b2_5 port.  Removed the link to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-5706&quot; title=&quot;conf-sanity test_57a: @@@@@@ FAIL: OST registration from failnode should fail &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-5706&quot;&gt;&lt;del&gt;LU-5706&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="22291">LU-4334</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                    <customfield id="customfield_10030" key="com.atlassian.jira.plugin.system.customfieldtypes:labels">
                        <customfieldname>Epic/Theme</customfieldname>
                        <customfieldvalues>
                                        <label>ZFS</label>
    
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwhfr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>13075</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>