<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:13:13 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7938] Recovery on secondary OSS node stalled</title>
                <link>https://jira.whamcloud.com/browse/LU-7938</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Description&lt;/p&gt;

&lt;p&gt;Error happens during soak testing of build &apos;20160324&apos; (see &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160324&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160324&lt;/a&gt;). DNE is enabled. MDTs had been formatted with ldiskfs, OSTs using zfs. MDS and OSS nodes are configured in HA active-active failover configuration. MDS nodes operated wiht &lt;b&gt;1&lt;/b&gt; MDT per MDS, while OSSes  running 4 OST per node.&lt;br/&gt;
Nodes lola-4 and lola-5 form a HA cluster.&lt;/p&gt;

&lt;p&gt;Event history&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-03-29 07:48:04,307:fsmgmt.fsmgmt:INFO     triggering fault oss_failover of node &lt;tt&gt;lola-4&lt;/tt&gt;&lt;br/&gt;
powercyle node&lt;/li&gt;
	&lt;li&gt;2016-03-29 07:52:27,876:fsmgmt.fsmgmt:INFO     lola-4 is up&lt;/li&gt;
	&lt;li&gt;2016-03-29 07:53:09,557:fsmgmt.fsmgmt:INFO     zpool import and mount of &lt;tt&gt;lola-4&lt;/tt&gt;&apos;s OSTs complete&lt;/li&gt;
	&lt;li&gt;Recovery don&apos;t complete after recovery_time is zero&lt;/li&gt;
	&lt;li&gt;2016-03-29 08:03 (approximately) Aborted recovery manually (&lt;tt&gt;lctl --device ... abort_recovery&lt;/tt&gt;)&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Attached files:&lt;br/&gt;
messages, console and debug log before (lustre-log-20160329-0759-recovery-stalled) and after recovery was aborted (lustre-log-20160329-0803-recovery-aborted)&lt;/p&gt;</description>
                <environment>lola&lt;br/&gt;
build: 2.8 GA + patches</environment>
        <key id="35679">LU-7938</key>
            <summary>Recovery on secondary OSS node stalled</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="6">Not a Bug</resolution>
                                        <assignee username="yong.fan">nasf</assignee>
                                    <reporter username="heckes">Frank Heckes</reporter>
                        <labels>
                            <label>soak</label>
                    </labels>
                <created>Tue, 29 Mar 2016 15:53:41 +0000</created>
                <updated>Wed, 27 Jul 2016 13:12:38 +0000</updated>
                            <resolved>Wed, 27 Jul 2016 13:12:38 +0000</resolved>
                                    <version>Lustre 2.8.0</version>
                    <version>Lustre 2.9.0</version>
                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="147227" author="green" created="Tue, 29 Mar 2016 17:42:54 +0000"  >&lt;p&gt;lustre-log-20160329-0759-recovery-stalled is referenced, but not really provided?&lt;/p&gt;</comment>
                            <comment id="147320" author="heckes" created="Wed, 30 Mar 2016 08:50:45 +0000"  >&lt;p&gt;Sorry, it took many attempts to upload the file.&lt;/p&gt;</comment>
                            <comment id="147321" author="heckes" created="Wed, 30 Mar 2016 08:52:57 +0000"  >&lt;p&gt;The error can be reproduced easily. Please let me know if debug logs with mask different than default are needed.&lt;/p&gt;</comment>
                            <comment id="149067" author="heckes" created="Fri, 15 Apr 2016 12:02:53 +0000"  >&lt;p&gt;The error occurred also in soak test for build &apos;20160413&apos; (see &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160413&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160413&lt;/a&gt;). Test set-up is the same as the one described above, besides&lt;br/&gt;
that the Lustre FS has been extended by 2 OSS nodes hosting 4 OSTs each.&lt;/p&gt;

&lt;p&gt;The recovery is stalled, but can be manually aborted, too. &lt;br/&gt;
Sequence of events:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-04-15 03:57:14,905:fsmgmt.fsmgmt:INFO     triggering fault oss_failover     (&lt;tt&gt;lola-5&lt;/tt&gt; OSTs ----&amp;gt; &lt;tt&gt;lola-4&lt;/tt&gt;)&lt;/li&gt;
	&lt;li&gt;2016-04-15 04:02:13,325:fsmgmt.fsmgmt:INFO     Wait for recovery to complete  (--&amp;gt; all OSTs mounted on secondary node &lt;tt&gt;lola-4&lt;/tt&gt;)&lt;/li&gt;
	&lt;li&gt;2016-04-15 04:20       created debug log&lt;/li&gt;
	&lt;li&gt;2016-04-15 04:28:29  Recovery aborted for ost-
{3,9,15,21}
&lt;p&gt; (decimal index)&lt;/p&gt;&lt;/li&gt;
	&lt;li&gt;2016-04-15 04:30       created debug log&lt;br/&gt;
Attached files:&lt;br/&gt;
console, messages files of lola-4 and lola-4-lustre-log-20160415-0420 (before abort), lola-4-lustre-log-20160415-0430 (after abort)&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="154807" author="cliffw" created="Mon, 6 Jun 2016 19:29:40 +0000"  >&lt;p&gt;The error is occurring again on the tip of master plus the two patches.&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;# cat /proc/fs/lustre/mdt/soaked-MDT0002/recovery_status 
status: RECOVERING
recovery_start: 1465232959
time_remaining: 0
connected_clients: 20/20
req_replay_clients: 0
lock_repay_clients: 1
completed_clients: 19
evicted_clients: 0
replayed_requests: 0
queued_requests: 0
next_transno: 442381632838
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="154809" author="cliffw" created="Mon, 6 Jun 2016 19:31:16 +0000"  >&lt;p&gt;Log errors&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Lustre: Skipped 299 previous similar messages
Lustre: soaked-MDT0002: Recovery already passed deadline 137:09, It is most likely due to DNE recovery is failed or stuck, please wait a few more minutes or abort the recovery.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="154810" author="cliffw" created="Mon, 6 Jun 2016 19:36:08 +0000"  >&lt;p&gt;Does not appear possible to abort_recovery, results in timeouts.&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;INFO: task lctl:41183 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.
      Tainted: P           -- ------------    2.6.32-573.26.1.el6_lustre.x86_64 #1
&lt;span class=&quot;code-quote&quot;&gt;&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot;&lt;/span&gt; disables &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; message.
lctl          D 0000000000000019     0 41183  41083 0x00000080
 ffff8806bdda7b68 0000000000000086 0000000000000000 ffff88082fe86ab0
 ffff88082fe86ae8 ffff880439371150 0000ece58069441d 0000000000000000
 ffff8806bdd8eae8 000000010f83d3bd ffff8806bdd8f068 ffff8806bdda7fd8
Call Trace:
 [&amp;lt;ffffffff8153aa35&amp;gt;] schedule_timeout+0x215/0x2e0
 [&amp;lt;ffffffff8153a693&amp;gt;] wait_for_common+0x123/0x180
 [&amp;lt;ffffffff81067650&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffff8153a7cd&amp;gt;] wait_for_completion+0x1d/0x20
 [&amp;lt;ffffffffa0992e80&amp;gt;] target_stop_recovery_thread+0x50/0xe0 [ptlrpc]
 [&amp;lt;ffffffffa112ac02&amp;gt;] mdt_iocontrol+0x6b2/0x920 [mdt]
 [&amp;lt;ffffffff812337cf&amp;gt;] ? security_inode_permission+0x1f/0x30
 [&amp;lt;ffffffffa0783edc&amp;gt;] class_handle_ioctl+0x15fc/0x20d0 [obdclass]
 [&amp;lt;ffffffff811a5998&amp;gt;] ? do_filp_open+0x798/0xd20
 [&amp;lt;ffffffffa07682ab&amp;gt;] obd_class_ioctl+0x4b/0x190 [obdclass]
 [&amp;lt;ffffffff811a7f82&amp;gt;] vfs_ioctl+0x22/0xa0
 [&amp;lt;ffffffff811a8124&amp;gt;] do_vfs_ioctl+0x84/0x580
 [&amp;lt;ffffffff811a03f6&amp;gt;] ? final_putname+0x26/0x50
 [&amp;lt;ffffffff811a86a1&amp;gt;] sys_ioctl+0x81/0xa0
 [&amp;lt;ffffffff810e905e&amp;gt;] ? __audit_syscall_exit+0x25e/0x290
 [&amp;lt;ffffffff8100b0d2&amp;gt;] system_call_fastpath+0x16/0x1b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="159572" author="heckes" created="Fri, 22 Jul 2016 12:02:55 +0000"  >&lt;p&gt;Comments from Cliff above seems to be related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8089&quot; title=&quot;MGT/MDT mount fails on secondary HA node&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8089&quot;&gt;&lt;del&gt;LU-8089&lt;/del&gt;&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;The error for for OST reported in the description above, also occurred for build &lt;a href=&quot;https://build.hpdd.intel.com/job/lustre-master/3406&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://build.hpdd.intel.com/job/lustre-master/3406&lt;/a&gt; (see &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160713&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160713&lt;/a&gt;)&lt;br/&gt;
and is reproducible for any failover initiated.&lt;br/&gt;
Test environment configuration is&lt;br/&gt;
4 MDSes with 1 MDT each formatted with &lt;em&gt;ldiskfs&lt;/em&gt; in active-active failover confguration &lt;br/&gt;
6 OSSes with 4 OSTs each formatted with &lt;em&gt;zfs&lt;/em&gt; in active-active failover configuration&lt;br/&gt;
16 clients&lt;br/&gt;
The stalled recovery process can be interrupted (aborted) at any time.&lt;/p&gt;

&lt;p&gt;I attached new message, console and debug log files (take during recovery, after recovery abort) of secondary node &lt;tt&gt;lola-5&lt;/tt&gt;. &lt;br/&gt;
Debug files of all other nodes taken at the same time, are available and can be uploaded on demand.&lt;/p&gt;

&lt;p&gt;Sequence of events:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-07-22 04:21:54,665:fsmgmt.fsmgmt:INFO     triggering fault oss_failover lola-4&lt;/li&gt;
	&lt;li&gt;2016-07-22 04:26:37,971:fsmgmt.fsmgmt:INFO     lola-4 is up&lt;/li&gt;
	&lt;li&gt;2016-07-22 04:27:01,278:fsmgmt.fsmgmt:INFO     ... soaked-OST0008 mounted successfully on lola-5  (NOTE: decimal id for OST)&lt;/li&gt;
	&lt;li&gt;2016-07-22 04:27:12,348:fsmgmt.fsmgmt:INFO     ... soaked-OST0020 mounted successfully on lola-5&lt;/li&gt;
	&lt;li&gt;2016-07-22 04:27:23,515:fsmgmt.fsmgmt:INFO     ... soaked-OST0002 mounted successfully on lola-5&lt;/li&gt;
	&lt;li&gt;2016-07-22 04:27:36,047:fsmgmt.fsmgmt:INFO     ... soaked-OST0014 mounted successfully on lola-5&lt;/li&gt;
	&lt;li&gt;2016-07-22 04:27:36                                              Recovery started&lt;/li&gt;
	&lt;li&gt;2016-07-22 04:37                                                   Recovery aborted&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Status of recovery on OSS node &lt;tt&gt;lola-5&lt;/tt&gt; and client status were:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@lola-5 ~]# lctl get_param *.*.recovery_status
obdfilter.soaked-OST0002.recovery_status=
status: RECOVERING
recovery_start: 0
time_remaining: 0
connected_clients: 0/23
req_replay_clients: 0
lock_repay_clients: 0
completed_clients: 0
evicted_clients: 0
replayed_requests: 0
queued_requests: 0
next_transno: 68719725397
obdfilter.soaked-OST0003.recovery_status=
status: COMPLETE 
recovery_start: 1468949759
recovery_duration: 30
completed_clients: 23/23
replayed_requests: 0
last_transno: 34368384861
VBR: DISABLED
IR: ENABLED
obdfilter.soaked-OST0008.recovery_status=
status: RECOVERING
recovery_start: 0
time_remaining: 0
connected_clients: 0/23
req_replay_clients: 0
lock_repay_clients: 0
completed_clients: 0
evicted_clients: 0
replayed_requests: 0
queued_requests: 0
next_transno: 68719723367
obdfilter.soaked-OST0009.recovery_status=
status: COMPLETE
recovery_start: 1468949751
recovery_duration: 38
completed_clients: 23/23
replayed_requests: 0
last_transno: 34368183131
VBR: DISABLED
IR: ENABLED
obdfilter.soaked-OST000e.recovery_status=
status: RECOVERING
recovery_start: 0
time_remaining: 0
connected_clients: 0/23
req_replay_clients: 0
lock_repay_clients: 0
completed_clients: 0
evicted_clients: 0
replayed_requests: 0
queued_requests: 0
next_transno: 68719751551
obdfilter.soaked-OST000f.recovery_status=
status: COMPLETE 
recovery_start: 1468949763
recovery_duration: 26
completed_clients: 23/23
replayed_requests: 0
last_transno: 34368269722
VBR: DISABLED
IR: ENABLED
obdfilter.soaked-OST0014.recovery_status=
status: RECOVERING
recovery_start: 0
time_remaining: 0
connected_clients: 0/23
req_replay_clients: 0
lock_repay_clients: 0
completed_clients: 0
evicted_clients: 0
replayed_requests: 0
queued_requests: 0
next_transno: 68719756594
obdfilter.soaked-OST0015.recovery_status=
status: COMPLETE 
recovery_start: 1468949751
recovery_duration: 38
completed_clients: 23/23
replayed_requests: 0
last_transno: 34368229218
VBR: DISABLED
IR: ENABLED
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The recovery stalled in a previous session running during night time over several hours without completion. That&apos;s why I aborted the recovery after ten minutes; I wasn&apos;t sure debug logs would contain relevant information, but confident that the recovery would never end.&lt;/p&gt;</comment>
                            <comment id="159678" author="yong.fan" created="Sun, 24 Jul 2016 13:39:26 +0000"  >&lt;p&gt;According to the recovery status, no clients connected to the new mounted OSTs on the lola-5, and there are no client connection records in the lola-5 debug logs (I am sure that it is NOT because of logs overwritten). So it seems that the clients were not aware of the OST load switching from lola-4 to lola-5, as to no connection to the lola-5, then no recovery. We need the full debug logs on the MGS and the client (one active client is enough). Thanks!&lt;/p&gt;</comment>
                            <comment id="159757" author="heckes" created="Mon, 25 Jul 2016 16:12:04 +0000"  >&lt;p&gt;Resources of &lt;tt&gt;lola-2&lt;/tt&gt; (&lt;tt&gt;OST0012 OST0000 OST0006 OST000c&lt;/tt&gt;) failed over to &lt;tt&gt;lola-3&lt;/tt&gt;&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-07-25 08:49:12,572:fsmgmt.fsmgmt:INFO     Wait for recovery to complete&lt;/li&gt;
	&lt;li&gt;2016-07-25 08:55                                                   created debug logs&lt;/li&gt;
	&lt;li&gt;2016-07-25 08:57:51                                              aborted recovery&lt;/li&gt;
	&lt;li&gt;2016-07-25 08:59                                                   created debug logs&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;Upload of debug logs has been started, but it might take some time.&lt;/p&gt;</comment>
                            <comment id="159846" author="heckes" created="Tue, 26 Jul 2016 05:40:48 +0000"  >&lt;p&gt;debug files have been uploaded. Roles:&lt;br/&gt;
lola-8: MGS&lt;br/&gt;
lola-3: secondary OSS&lt;br/&gt;
lola-30: Lustre client&lt;/p&gt;</comment>
                            <comment id="159854" author="yong.fan" created="Tue, 26 Jul 2016 07:37:18 +0000"  >&lt;p&gt;Thanks Frank for the detailed logs. According to the client side log, it always tried to connect lola2 for the failed OSTs (OST0012 OST0000 OST0006 OST000c):&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;00000100:00000001:30.0:1469462043.991030:0:125723:0:(import.c:508:import_select_connection()) Process entered&lt;br/&gt;
00000100:00080000:30.0:1469462043.991033:0:125723:0:(import.c:523:import_select_connection()) soaked-OST0000-osc-ffff8810717d2000: connect to NID 192.168.1.102@o2ib10 last attempt 5189761512&lt;br/&gt;
00000100:00080000:30.0:1469462043.991036:0:125723:0:(import.c:567:import_select_connection()) soaked-OST0000-osc-ffff8810717d2000: tried all connections, increasing latency to 50s&lt;br/&gt;
00000100:00080000:30.0:1469462043.991099:0:125723:0:(import.c:601:import_select_connection()) soaked-OST0000-osc-ffff8810717d2000: import ffff880ff6706000 using connection 192.168.1.102@o2ib10/192.168.1.102@o2ib10&lt;br/&gt;
00000100:00000001:30.0:1469462043.991104:0:125723:0:(import.c:605:import_select_connection()) Process leaving (rc=0 : 0 : 0)&lt;br/&gt;
...&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;That means there is only one candidate can be selected to connect the OST0000, but according to your description, there should be another candidate 192.168.1.103@o2ib for connection. So I am wondering how did you config the lola2 and lola3 as the failover pairs?&lt;/p&gt;</comment>
                            <comment id="159863" author="heckes" created="Tue, 26 Jul 2016 11:58:22 +0000"  >&lt;p&gt;Indeed it&apos;s a configuration error. Looks like NID of the secondary node (&lt;tt&gt;192.168.1.103@o2ib&lt;/tt&gt;) weren&apos;t written during FS format. &lt;br/&gt;
I set them explicitly again using &lt;tt&gt;zfs set lustre:failover.node&lt;/tt&gt;, and check whether this resolves the issue.&lt;/p&gt;</comment>
                            <comment id="159905" author="heckes" created="Tue, 26 Jul 2016 15:37:48 +0000"  >&lt;p&gt;I set the parameter &apos;&lt;tt&gt;lustre:failover.node&lt;/tt&gt;&apos; for on all OSTs to the appropriate values for the failover nodes (see attached file ost-failover-setttings-20160726_0702).&lt;br/&gt;
Actually this didn&apos;t solve the problem. Recovery is stalled still on failover node.&lt;/p&gt;

&lt;p&gt;I also reformatted the soak FS. The parameter setting will be filled (e.g.) with the following content:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;192.168.1.102: mkfs_cmd = zfs create -o canmount=off -o xattr=sa soaked-ost18/ost18
192.168.1.102: Writing soaked-ost18/ost18 properties
192.168.1.102:   lustre:version=1
192.168.1.102:   lustre:flags=4194
192.168.1.102:   lustre:index=18
192.168.1.102:   lustre:fsname=soaked
192.168.1.102:   lustre:svname=soaked:OST0012
192.168.1.102:   lustre:mgsnode=192.168.1.108@o2ib10:192.168.1.109@o2ib10
192.168.1.102:   lustre:failover.node=192.168.1.102@o2ib10:192.168.1.103@o2ib10
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Additionally I used option &apos;&lt;tt&gt;--force-nohostid&lt;/tt&gt;&apos; for mkfs.lustre - command. I think the patch enabling hostid haven&apos;t landed, yet and spl fails during reformat if /etc/hostid isn&apos;t available.&lt;/p&gt;</comment>
                            <comment id="159908" author="heckes" created="Tue, 26 Jul 2016 15:39:29 +0000"  >&lt;p&gt;reformat is onging. I&apos;m going to start soak again, once it has completed.&lt;/p&gt;</comment>
                            <comment id="159911" author="heckes" created="Tue, 26 Jul 2016 15:41:39 +0000"  >&lt;p&gt;failover.node is specified for mkfs.lustre command  via option &lt;tt&gt;--servicenode&lt;/tt&gt;&lt;/p&gt;</comment>
                            <comment id="160027" author="heckes" created="Wed, 27 Jul 2016 08:50:02 +0000"  >&lt;p&gt;Failover works after reformat of Lustre FS. Ticket can be closed.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                                        </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="22314" name="console-lola-5.log-20160722.bz2" size="60576" author="heckes" created="Fri, 22 Jul 2016 12:24:00 +0000"/>
                            <attachment id="20933" name="console-lola-5.log.bz2" size="41428" author="heckes" created="Tue, 29 Mar 2016 15:59:32 +0000"/>
                            <attachment id="21149" name="lola-4-console-20160415.bz2" size="97506" author="heckes" created="Fri, 15 Apr 2016 12:21:53 +0000"/>
                            <attachment id="21152" name="lola-4-lustre-log-20160415-0420.bz2" size="259" author="heckes" created="Fri, 15 Apr 2016 12:30:08 +0000"/>
                            <attachment id="21151" name="lola-4-lustre-log-20160415-0430.bz2" size="958037" author="heckes" created="Fri, 15 Apr 2016 12:23:02 +0000"/>
                            <attachment id="21150" name="lola-4-messages-20160415.bz2" size="219963" author="heckes" created="Fri, 15 Apr 2016 12:21:53 +0000"/>
                            <attachment id="20959" name="lustre-log-20160329-0759-recovery-stalled.bz2" size="279" author="heckes" created="Wed, 30 Mar 2016 08:49:48 +0000"/>
                            <attachment id="20934" name="lustre-log-20160329-0803-recovery-aborted.bz2" size="8315" author="heckes" created="Tue, 29 Mar 2016 15:59:46 +0000"/>
                            <attachment id="22344" name="lustre-log-lola-3-2016-07-25_0855-ost-recovery-stalled.bz2" size="305" author="heckes" created="Mon, 25 Jul 2016 18:38:11 +0000"/>
                            <attachment id="22352" name="lustre-log-lola-3-2016-07-25_0859-after-ost-recovery-aborted.bz2" size="317" author="heckes" created="Tue, 26 Jul 2016 05:39:45 +0000"/>
                            <attachment id="22353" name="lustre-log-lola-30-2016-07-25_0855-ost-recovery-stalled.bz2" size="308" author="heckes" created="Tue, 26 Jul 2016 05:39:45 +0000"/>
                            <attachment id="22354" name="lustre-log-lola-30-2016-07-25_0859-after-ost-recovery-aborted.bz2" size="319" author="heckes" created="Tue, 26 Jul 2016 05:39:45 +0000"/>
                            <attachment id="22315" name="lustre-log-lola-5-20160722_0434_ost_recovery_stalled.bz2" size="301" author="heckes" created="Fri, 22 Jul 2016 12:24:00 +0000"/>
                            <attachment id="22316" name="lustre-log-lola-5-20160722_0438_after_ost_recovery_aborted.bz2" size="179793" author="heckes" created="Fri, 22 Jul 2016 12:24:00 +0000"/>
                            <attachment id="22342" name="lustre-log-lola-8-2016-07-25_0855-ost-recovery-stalled.bz2" size="305" author="heckes" created="Mon, 25 Jul 2016 18:38:11 +0000"/>
                            <attachment id="22343" name="lustre-log-lola-8-2016-07-25_0859-after-ost-recovery-aborted.bz2" size="317" author="heckes" created="Mon, 25 Jul 2016 18:38:11 +0000"/>
                            <attachment id="22313" name="messages-lola-5.log-20160722.bz2" size="225780" author="heckes" created="Fri, 22 Jul 2016 12:24:00 +0000"/>
                            <attachment id="20932" name="messages-lola-5.log.bz2" size="49989" author="heckes" created="Tue, 29 Mar 2016 15:59:16 +0000"/>
                            <attachment id="22357" name="ost-failover-setttings-20160726_0702" size="16340" author="heckes" created="Tue, 26 Jul 2016 15:38:10 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzy62v:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>