<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:15:55 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-8250] MDT recovery stalled on secondary node</title>
                <link>https://jira.whamcloud.com/browse/LU-8250</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;The error happened during soak testing of build &apos;20160601&apos; (see: &lt;a href=&quot;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160601&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://wiki.hpdd.intel.com/display/Releases/Soak+Testing+on+Lola#SoakTestingonLola-20160601&lt;/a&gt;). DNE is enabled. MDT have been formatted using &lt;em&gt;ldiskfs&lt;/em&gt;, OSTs using &lt;em&gt;zfs&lt;/em&gt;. MDSes host one MDT per node. MDSes are configured in active-active failover configuration.&lt;/p&gt;

&lt;p&gt;The issue might be related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7848&quot; title=&quot;Recovery process on MDS stalled&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7848&quot;&gt;&lt;del&gt;LU-7848&lt;/del&gt;&lt;/a&gt; although the change is part of build under test.&lt;br/&gt;
The error results in the start of the oom-killer, which is documented in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7836&quot; title=&quot;MDSes crashed with oom-killer&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7836&quot;&gt;&lt;del&gt;LU-7836&lt;/del&gt;&lt;/a&gt;. This ticket might be a duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7836&quot; title=&quot;MDSes crashed with oom-killer&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7836&quot;&gt;&lt;del&gt;LU-7836&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Events:&lt;br/&gt;
1st Event:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-06-03 11:31:10   - failover resource of lola-10 (MDT-2) --&amp;gt; lola-11&lt;/li&gt;
	&lt;li&gt;2016-06-03 11:36:37   -   ... soaked-MDT0002 mounted successfully on lola-11&lt;/li&gt;
	&lt;li&gt;till 2016-06-04-00:44   - soaked-MDT0002 in status &apos;RECOVERING&apos;.&lt;/li&gt;
	&lt;li&gt;2016-06-04-00:44:52  - lola-11 crash with oom-killer&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;2nd Event:&lt;/p&gt;
&lt;ul&gt;
	&lt;li&gt;2016-06-07 08:34:06,621 triggering fault mds_failover lola-10 (MDT-2) --&amp;gt; lola-11&lt;/li&gt;
	&lt;li&gt;2016-06-07 08:38:42  - Mounting soaked-MDT0002 on lola-11&lt;/li&gt;
	&lt;li&gt;since 2016-06-07 08:39:32,155 Wait for recovery to complete&lt;/li&gt;
	&lt;li&gt;memory resources are nearly exhausted:
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[root@lola-11 ~]# date
Wed Jun  8 07:59:49 PDT 2016
[root@lola-11 ~]# collectl -sm --verbose
waiting for 1 second sample...

# MEMORY SUMMARY
#&amp;lt;-------------------------------Physical Memory--------------------------------------&amp;gt;&amp;lt;-----------Swap------------&amp;gt;&amp;lt;-------Paging------&amp;gt;
#   Total    Used    Free    Buff  Cached    Slab  Mapped    Anon  Commit  Locked Inact Total  Used  Free   In  Out Fault MajFt   In  Out
   32006M  30564M   1441M 127144K 676256K  28701M  16196K  69072K 201740K   5008K  509M   15G     0   15G    0    0    28     0    0    8
   32006M  30565M   1441M 127144K 676256K  28701M  16196K  69072K 201740K   5008K  509M   15G     0   15G    0    0    63     0    0    4
   32006M  30565M   1441M 127144K 676256K  28701M  16196K  69072K 201740K   5008K  509M   15G     0   15G    0    0     1     0    0    0
   32006M  30564M   1441M 127144K 676256K  28701M  16196K  69072K 201740K   5008K  509M   15G     0   15G    0    0    17     0    0    0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Attached files:&lt;br/&gt;
*1st event only: Saved crash dump file to &lt;tt&gt;lhn.hpdd.intel.com:/var/crashdumps/lu-7836/lola-11/127.0.0.1-2016-06-04-00:44:52&lt;/tt&gt;&lt;/p&gt;&lt;/li&gt;
	&lt;li&gt;2nd event only: kernel debug log of lola-11, dmesg&lt;/li&gt;
	&lt;li&gt;Both event: messages, console logs&lt;/li&gt;
&lt;/ul&gt;
</description>
                <environment>lola&lt;br/&gt;
build: commit aa84fbc8165f526dae4bd824a48c186c3ac2f639 + patches</environment>
        <key id="37472">LU-8250</key>
            <summary>MDT recovery stalled on secondary node</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="tappro">Mikhail Pershin</assignee>
                                    <reporter username="heckes">Frank Heckes</reporter>
                        <labels>
                            <label>soak</label>
                    </labels>
                <created>Wed, 8 Jun 2016 15:54:57 +0000</created>
                <updated>Wed, 26 Oct 2016 23:24:03 +0000</updated>
                            <resolved>Wed, 26 Oct 2016 23:24:03 +0000</resolved>
                                    <version>Lustre 2.9.0</version>
                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>13</watches>
                                                                            <comments>
                            <comment id="155097" author="heckes" created="Wed, 8 Jun 2016 16:22:45 +0000"  >&lt;p&gt;Triggering an abort of the recovery hangs:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LustreError: 139015:0:(mdt_handler.c:5896:mdt_iocontrol()) soaked-MDT0002: Aborting recovery for device
LustreError: 139015:0:(ldlm_lib.c:2577:target_stop_recovery_thread()) soaked-MDT0002: Aborting recovery
INFO: task lctl:139015 blocked for more than 120 seconds.
      Tainted: P           -- ------------    2.6.32-573.26.1.el6_lustre.x86_64 #1
&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.
lctl          D 0000000000000013     0 139015 129872 0x00000080
 ffff8804187cfb68 0000000000000082 0000000000000000 ffff880431d7eab0
 ffff880431d7eae8 ffff880431d7f150 00ff8804187cfb18 0000000000000000
 ffff8803fc2c0078 ffff880038775a00 ffff8803fc2c05f8 ffff8804187cffd8
Call Trace:
 [&amp;lt;ffffffff81539b0e&amp;gt;] ? thread_return+0x4e/0x7d0
 [&amp;lt;ffffffff8153aa35&amp;gt;] schedule_timeout+0x215/0x2e0
 [&amp;lt;ffffffff8153a693&amp;gt;] wait_for_common+0x123/0x180
 [&amp;lt;ffffffff81067650&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffff8153a7cd&amp;gt;] wait_for_completion+0x1d/0x20
 [&amp;lt;ffffffffa0ae3e80&amp;gt;] target_stop_recovery_thread+0x50/0xe0 [ptlrpc]
 [&amp;lt;ffffffffa1223c02&amp;gt;] mdt_iocontrol+0x6b2/0x920 [mdt]
 [&amp;lt;ffffffff812337cf&amp;gt;] ? security_inode_permission+0x1f/0x30
 [&amp;lt;ffffffffa08d4edc&amp;gt;] class_handle_ioctl+0x15fc/0x20d0 [obdclass]
 [&amp;lt;ffffffff811a5998&amp;gt;] ? do_filp_open+0x798/0xd20
 [&amp;lt;ffffffffa08b92ab&amp;gt;] obd_class_ioctl+0x4b/0x190 [obdclass]
 [&amp;lt;ffffffff811a7f82&amp;gt;] vfs_ioctl+0x22/0xa0
 [&amp;lt;ffffffff811a8124&amp;gt;] do_vfs_ioctl+0x84/0x580
 [&amp;lt;ffffffff811a03f6&amp;gt;] ? final_putname+0x26/0x50
 [&amp;lt;ffffffff811a86a1&amp;gt;] sys_ioctl+0x81/0xa0
 [&amp;lt;ffffffff810e905e&amp;gt;] ? __audit_syscall_exit+0x25e/0x290
 [&amp;lt;ffffffff8100b0d2&amp;gt;] system_call_fastpath+0x16/0x1b
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="162419" author="cliffw" created="Thu, 18 Aug 2016 18:38:08 +0000"  >&lt;p&gt;We seem to have hit this again.&lt;br/&gt;
Build lustre-master 3424.&lt;br/&gt;
MDS crashes:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Kernel panic - not syncing: Attempted to kill init!^M
Pid: 1, comm: init Tainted: P           -- ------------    2.6.32-573.26.1.el6_lustre.x86_64 #1^M
Call Trace:^M
 [&amp;lt;ffffffff81539407&amp;gt;] ? panic+0xa7/0x16f^M
 [&amp;lt;ffffffff8107d0a7&amp;gt;] ? do_exit+0x867/0x870^M
 [&amp;lt;ffffffff8107d1b7&amp;gt;] ? sys_exit+0x17/0x20^M
 [&amp;lt;ffffffff8100b0d2&amp;gt;] ? system_call_fastpath+0x16/0x1b^M
^[[2J^[[1;1HCopyright(c) 2009 - 2012 Intel Corporation.All rights reserved. ^
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Upon restart, MDS fails to complete recovery, times out, During this time the soak controller halted lola-11 for failover. Attempting to abort recovery results in timeouts&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;LustreError: 7532:0:(ldlm_lib.c:2576:target_stop_recovery_thread()) soaked-MDT0003: Aborting recovery^M
Lustre: 6032:0:(ldlm_lib.c:2025:target_recovery_overseer()) recovery is aborted, evict exports in recovery^M
Lustre: 4563:0:(client.c:2113:ptlrpc_expire_one_request()) @@@ Request sent has failed due to network error: [sent 1471532976/real 1471532976]  req@ffff8804070af9c0 x1543013122722384/t0(0) o38-&amp;gt;soaked-MDT0002-osp-MDT0003@192.168.1.111@o2ib10:24/4 lens 520/544 e 0 to 1 dl 1471533032 ref 1 fl Rpc:eXN/0/ffffffff rc 0/-1^M
Lustre: 4563:0:(client.c:2113:ptlrpc_expire_one_request()) Skipped 5 previous similar messages^M
Aug 18 08:10:01 lola-10 TIME: Time stamp &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; console
INFO: task lctl:7532 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.^M
      Tainted: P           -- ------------    2.6.32-573.26.1.el6_lustre.x86_64 #1^M
&lt;span class=&quot;code-quote&quot;&gt;&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot;&lt;/span&gt; disables &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; message.^M
lctl          D 0000000000000011     0  7532   7440 0x00000080^M
 ffff8804069dbb68 0000000000000086 0000000000000000 ffff880835ef3520^M
 ffff880835ef3558 ffff880835ef3bc0 000000d247916847 0000000000000000^M
 ffff880437394078 0000000100092ec2 ffff8804373945f8 ffff8804069dbfd8^M
Call Trace:^M
 [&amp;lt;ffffffff8153aa35&amp;gt;] schedule_timeout+0x215/0x2e0^M
 [&amp;lt;ffffffff8153a693&amp;gt;] wait_for_common+0x123/0x180^M
 [&amp;lt;ffffffff81067650&amp;gt;] ? default_wake_function+0x0/0x20^M
 [&amp;lt;ffffffff8153a7cd&amp;gt;] wait_for_completion+0x1d/0x20^M
 [&amp;lt;ffffffffa0970f80&amp;gt;] target_stop_recovery_thread+0x50/0xe0 [ptlrpc]^M
 [&amp;lt;ffffffffa10ebf22&amp;gt;] mdt_iocontrol+0x6a2/0xa80 [mdt]^M
 [&amp;lt;ffffffff812337cf&amp;gt;] ? security_inode_permission+0x1f/0x30^M
 [&amp;lt;ffffffffa07843ec&amp;gt;] class_handle_ioctl+0x15fc/0x20d0 [obdclass]^M
 [&amp;lt;ffffffff811a5998&amp;gt;] ? do_filp_open+0x798/0xd20^M
 [&amp;lt;ffffffffa07682ab&amp;gt;] obd_class_ioctl+0x4b/0x190 [obdclass]^M
 [&amp;lt;ffffffff811a7f82&amp;gt;] vfs_ioctl+0x22/0xa0^M
 [&amp;lt;ffffffff811a8124&amp;gt;] do_vfs_ioctl+0x84/0x580^M
 [&amp;lt;ffffffff811a03f6&amp;gt;] ? final_putname+0x26/0x50^M
 [&amp;lt;ffffffff811a86a1&amp;gt;] sys_ioctl+0x81/0xa0^M
 [&amp;lt;ffffffff810e905e&amp;gt;] ? __audit_syscall_exit+0x25e/0x290^M
 [&amp;lt;ffffffff8100b0d2&amp;gt;] system_call_fastpath+0x16/0x1b^M
LustreError: 137-5: soaked-MDT0002_UUID: not available &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; connect from 192.168.1.131@o2ib100 (no target). If you are running an HA pair check that the target is mounted on the other server.^M
LustreError: Skipped 351 previous similar messages^M
INFO: task lctl:7532 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.^M
      Tainted: P           -- ------------    2.6.32-573.26.1.el6_lustre.x86_64 #1^M
&lt;span class=&quot;code-quote&quot;&gt;&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot;&lt;/span&gt; disables &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; message.^M
lctl          D 0000000000000011     0  7532   7440 0x00000080^M
 ffff8804069dbb68 0000000000000086 0000000000000000 ffff880835ef3520^M
 ffff880835ef3558 ffff880835ef3bc0 000000d247916847 0000000000000000^M
 ffff880437394078 0000000100092ec2 ffff8804373945f8 ffff8804069dbfd8^M
Call Trace:^M
 [&amp;lt;ffffffff8153aa35&amp;gt;] schedule_timeout+0x215/0x2e0^M
 [&amp;lt;ffffffff8153a693&amp;gt;] wait_for_common+0x123/0x180^M
 [&amp;lt;ffffffff81067650&amp;gt;] ? default_wake_function+0x0/0x20^M
 [&amp;lt;ffffffff8153a7cd&amp;gt;] wait_for_completion+0x1d/0x20^M
 [&amp;lt;ffffffffa0970f80&amp;gt;] target_stop_recovery_thread+0x50/0xe0 [ptlrpc]^M
 [&amp;lt;ffffffffa10ebf22&amp;gt;] mdt_iocontrol+0x6a2/0xa80 [mdt]^M
 [&amp;lt;ffffffff812337cf&amp;gt;] ? security_inode_permission+0x1f/0x30^M
 [&amp;lt;ffffffffa07843ec&amp;gt;] class_handle_ioctl+0x15fc/0x20d0 [obdclass]^M
 [&amp;lt;ffffffff811a5998&amp;gt;] ? do_filp_open+0x798/0xd20^M
 [&amp;lt;ffffffffa07682ab&amp;gt;] obd_class_ioctl+0x4b/0x190 [obdclass]^M
 [&amp;lt;ffffffff811a7f82&amp;gt;] vfs_ioctl+0x22/0xa0^M
 [&amp;lt;ffffffff811a8124&amp;gt;] do_vfs_ioctl+0x84/0x580^M
 [&amp;lt;ffffffff811a03f6&amp;gt;] ? final_putname+0x26/0x50^M
 [&amp;lt;ffffffff811a86a1&amp;gt;] sys_ioctl+0x81/0xa0^M
 [&amp;lt;ffffffff810e905e&amp;gt;] ? __audit_syscall_exit+0x25e/0x290^M
 [&amp;lt;ffffffff8100b0d2&amp;gt;] system_call_fastpath+0x16/0x1b^M
INFO: task lctl:7532 blocked &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more than 120 seconds.^M
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; 
&lt;p&gt;At this point the systems were rebooted. &lt;br/&gt;
After reboot, the MDT could not be mounted, as MGS reported a duplicate export:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Aug 18 11:24:44 lola-10 kernel: LDISKFS-fs (dm-1): mounted filesystem with ordered data mode. quota=on. Opts:
Aug 18 11:24:45 lola-10 kernel: LustreError: 6045:0:(tgt_lastrcvd.c:1474:tgt_clients_data_init()) soaked-MDT0002: duplicate export &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; client generation 1
Aug 18 11:24:45 lola-10 kernel: LustreError: 6045:0:(obd_config.c:578:class_setup()) setup soaked-MDT0002 failed (-114)
Aug 18 11:24:45 lola-10 kernel: LustreError: 6045:0:(obd_config.c:1671:class_config_llog_handler()) MGC192.168.1.108@o2ib10: cfg command failed: rc = -114
Aug 18 11:24:45 lola-10 kernel: Lustre:    cmd=cf003 0:soaked-MDT0002  1:soaked-MDT0002_UUID  2:2  3:soaked-MDT0002-mdtlov  4:f
Aug 18 11:24:45 lola-10 kernel:
Aug 18 11:24:45 lola-10 kernel: LustreError: 15c-8: MGC192.168.1.108@o2ib10: The configuration from log &lt;span class=&quot;code-quote&quot;&gt;&apos;soaked-MDT0002&apos;&lt;/span&gt; failed (-114). This may be the result of communication errors between &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; node and the MGS, a bad configuration, or other errors. See the syslog &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; more information.
Aug 18 11:24:45 lola-10 kernel: LustreError: 5951:0:(obd_mount_server.c:1352:server_start_targets()) failed to start server soaked-MDT0002: -114
Aug 18 11:24:45 lola-10 kernel: LustreError: 5951:0:(obd_mount_server.c:1844:server_fill_super()) Unable to start targets: -114
Aug 18 11:24:45 lola-10 kernel: LustreError: 5951:0:(obd_config.c:625:class_cleanup()) Device 4 not setup
Aug 18 11:24:46 lola-10 kernel: Lustre: server umount soaked-MDT0002 complete
Aug 18 11:24:46 lola-10 kernel: LustreError: 5951:0:(obd_mount.c:1453:lustre_fill_super()) Unable to mount  (-114)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
</comment>
                            <comment id="162551" author="pjones" created="Fri, 19 Aug 2016 17:37:39 +0000"  >&lt;p&gt;Hongchao&lt;/p&gt;

&lt;p&gt;Could you please look into this issue?&lt;/p&gt;

&lt;p&gt;THanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="162797" author="hongchao.zhang" created="Tue, 23 Aug 2016 11:13:49 +0000"  >&lt;p&gt;I have looked at the logs, but I have not found useful information to help identify the cause of the problem.&lt;/p&gt;

&lt;p&gt;the increased memory usage could be related to the continuous lock replay req to enqueue&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;...
00010000:00080000:26.0:1465394173.485502:0:6068:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff88017e947050 x1536423085514416/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:334/0 lens 328/0 e 0 to 0 dl 1465394179 ref 2 fl Interpret:/40/ffffffff rc 0/-1
00010000:00080000:27.0:1465394173.485521:0:6067:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff880247be9c50 x1536423085514432/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:334/0 lens 328/0 e 0 to 0 dl 1465394179 ref 2 fl Interpret:/40/ffffffff rc 0/-1

...

00010000:00080000:26.0:1465394179.502412:0:6068:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff88016a6c2450 x1536423085514928/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:340/0 lens 328/0 e 0 to 0 dl 1465394185 ref 2 fl Interpret:/40/ffffffff rc 0/-1
00010000:00080000:27.0:1465394179.502417:0:6067:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff8802828f5050 x1536423085514944/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:340/0 lens 328/0 e 0 to 0 dl 1465394185 ref 2 fl Interpret:/40/ffffffff rc 0/-1

...

00010000:00080000:27.0:1465394185.503466:0:6067:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff880141b1e850 x1536423085515456/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:346/0 lens 328/0 e 0 to 0 dl 1465394191 ref 2 fl Interpret:/40/ffffffff rc 0/-1
00010000:00080000:26.0:1465394185.503491:0:6068:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff88031ca74050 x1536423085515472/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:346/0 lens 328/0 e 0 to 0 dl 1465394191 ref 2 fl Interpret:/40/ffffffff rc 0/-1

...

00010000:00080000:27.0:1465394191.504477:0:6067:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff88023309b450 x1536423085516016/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:352/0 lens 328/0 e 0 to 0 dl 1465394197 ref 2 fl Interpret:/40/ffffffff rc 0/-1
00010000:00080000:10.0:1465394191.504487:0:6068:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff88023309bc50 x1536423085516032/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:352/0 lens 328/0 e 0 to 0 dl 1465394197 ref 2 fl Interpret:/40/ffffffff rc 0/-1

...

00010000:00080000:10.0:1465394209.507404:0:6068:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff880247b59850 x1536423085517600/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:370/0 lens 328/0 e 0 to 0 dl 1465394215 ref 2 fl Interpret:/40/ffffffff rc 0/-1
00010000:00080000:11.0:1465394209.507410:0:6067:0:(ldlm_lib.c:2733:target_queue_recovery_request()) @@@ queue lock replay req  req@ffff88017e9e1c50 x1536423085517616/t0(0) o101-&amp;gt;soaked-MDT0001-mdtlov_UUID@192.168.1.109@o2ib10:370/0 lens 328/0 e 0 to 0 dl 1465394215 ref 2 fl Interpret:/40/ffffffff rc 0/-1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;the problem that MDT can&apos;t be mounted (with error -114, -EALREADY) is the same problem described in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7794&quot; title=&quot;tgt_clients_data_init()) soaked-MDT0001: duplicate export for client generation 1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7794&quot;&gt;&lt;del&gt;LU-7794&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Is the stack trace of all processes available, it could help to find where the recovery is stuck.&lt;br/&gt;
Thanks!&lt;/p&gt;</comment>
                            <comment id="162812" author="heckes" created="Tue, 23 Aug 2016 14:13:01 +0000"  >&lt;p&gt;Hi Hongchao,&lt;br/&gt;
there&apos;s a crash file available that had been stored at: &lt;tt&gt;lhn.hpdd.intel.com:/var/crashdumps/lu-7836/lola-11/127.0.0.1-2016-06-04-00:44:52&lt;/tt&gt; (see above). It&apos;s easy to access if you have an account on Lola (DCO create that within 2 hours if you file a ticket containing your ssh-public-key). crashdump tools and lustre.so are installed on the head node.&lt;/p&gt;</comment>
                            <comment id="166693" author="di.wang" created="Wed, 21 Sep 2016 09:44:48 +0000"  >&lt;p&gt;Hmm, right now, we do not fail the connection between MDTs to make sure the FS will not corrupt silently.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Jun  7 08:40:11 lola-11 kernel: Lustre: soaked-MDT0002: Client f6c679b2-f46e-20c5-89f7-51193ed93a53 (at 192.168.1.121@o2ib100) reconnecting, waiting for 20 clients in recovery for 1:52
Jun  7 08:40:11 lola-11 kernel: Lustre: Skipped 99 previous similar messages
Jun  7 08:40:18 lola-11 sshd[7507]: Accepted publickey for root from 10.4.0.116 port 47943 ssh2
Jun  7 08:40:18 lola-11 sshd[7507]: pam_unix(sshd:session): session opened for user root by (uid=0)
Jun  7 08:40:18 lola-11 sshd[7507]: Received disconnect from 10.4.0.116: 11: disconnected by user
Jun  7 08:40:18 lola-11 sshd[7507]: pam_unix(sshd:session): session closed for user root
Jun  7 08:40:33 lola-11 sshd[7530]: Accepted publickey for root from 10.4.0.116 port 47963 ssh2
Jun  7 08:40:33 lola-11 sshd[7530]: pam_unix(sshd:session): session opened for user root by (uid=0)
Jun  7 08:40:33 lola-11 sshd[7530]: Received disconnect from 10.4.0.116: 11: disconnected by user
Jun  7 08:40:33 lola-11 sshd[7530]: pam_unix(sshd:session): session closed for user root
Jun  7 08:40:47 lola-11 kernel: Lustre: soaked-MDT0003: recovery is timed out, evict stale exports
Jun  7 08:40:47 lola-11 kernel: Lustre: 6183:0:(ldlm_lib.c:2016:target_recovery_overseer()) soaked-MDT0003 recovery is aborted by hard timeout
Jun  7 08:40:47 lola-11 kernel: Lustre: 6183:0:(ldlm_lib.c:2026:target_recovery_overseer()) recovery is aborted, evict exports in recovery
Jun  7 08:40:47 lola-11 kernel: Lustre: soaked-MDT0003: disconnecting 1 stale clients
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; 

&lt;p&gt;I just glanced the log a bit, and it seems MDT0002 and MDT0003 were recovering at the same time. I thought soak-test will wait the MDT recovery finish, before fail another MDT? no? I might miss sth. thanks.&lt;/p&gt;</comment>
                            <comment id="167847" author="tappro" created="Fri, 30 Sep 2016 08:05:34 +0000"  >&lt;p&gt;Frank, is patch from &lt;a href=&quot;http://review.whamcloud.com/#/c/13726/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/13726/&lt;/a&gt; applied in the build or not? If not, I&apos;d try with it.&lt;/p&gt;</comment>
                            <comment id="169237" author="di.wang" created="Wed, 12 Oct 2016 03:21:37 +0000"  >&lt;p&gt;It turns out one MDS is stuck here&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;mdt00_016     S 0000000000000005     0  6217      2 0x00000080
 ffff880404c37760 0000000000000046 0000000000000000 ffff880404c37724
 0000000000000000 ffff88043fe82800 00000db7708e9e5b 0000000000000286
 ffff880404c37700 ffffffff81089e8c ffff8804066d9068 ffff880404c37fd8
Call Trace:
 [&amp;lt;ffffffff81089e8c&amp;gt;] ? lock_timer_base+0x3c/0x70
 [&amp;lt;ffffffff8153a9b2&amp;gt;] schedule_timeout+0x192/0x2e0
 [&amp;lt;ffffffff81089fa0&amp;gt;] ? process_timeout+0x0/0x10
 [&amp;lt;ffffffffa0b09821&amp;gt;] ptlrpc_set_wait+0x321/0x960 [ptlrpc]
 [&amp;lt;ffffffffa0afe980&amp;gt;] ? ptlrpc_interrupted_set+0x0/0x120 [ptlrpc]
 [&amp;lt;ffffffff81067650&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffffa0b15d05&amp;gt;] ? lustre_msg_set_jobid+0xf5/0x130 [ptlrpc]
 [&amp;lt;ffffffffa0b09ee1&amp;gt;] ptlrpc_queue_wait+0x81/0x220 [ptlrpc]
 [&amp;lt;ffffffffa13d4bc2&amp;gt;] osp_remote_sync+0xf2/0x1e0 [osp]
 [&amp;lt;ffffffffa13ba821&amp;gt;] osp_xattr_get+0x681/0xf90 [osp]
 [&amp;lt;ffffffffa12eb5b5&amp;gt;] lod_xattr_get+0x185/0x760 [lod]
 [&amp;lt;ffffffffa134f917&amp;gt;] mdd_links_read+0x117/0x270 [mdd]
 [&amp;lt;ffffffffa1364996&amp;gt;] ? mdd_attr_set_internal+0xd6/0x2c0 [mdd]
 [&amp;lt;ffffffffa13515bc&amp;gt;] mdd_linkea_prepare+0x3ec/0x4d0 [mdd]
 [&amp;lt;ffffffffa13573f2&amp;gt;] mdd_link+0xde2/0x10c0 [mdd]
 [&amp;lt;ffffffffa1222a8f&amp;gt;] mdt_reint_link+0xb2f/0xce0 [mdt]
 [&amp;lt;ffffffff81299b7a&amp;gt;] ? strlcpy+0x4a/0x60
 [&amp;lt;ffffffffa12178cf&amp;gt;] ? ucred_set_jobid+0x5f/0x70 [mdt]
 [&amp;lt;ffffffffa121b04d&amp;gt;] mdt_reint_rec+0x5d/0x200 [mdt]
 [&amp;lt;ffffffffa1205d5b&amp;gt;] mdt_reint_internal+0x62b/0xa50 [mdt]
 [&amp;lt;ffffffffa120662b&amp;gt;] mdt_reint+0x6b/0x120 [mdt]
 [&amp;lt;ffffffffa0b790cc&amp;gt;] tgt_request_handle+0x8ec/0x1440 [ptlrpc]
 [&amp;lt;ffffffffa0b25821&amp;gt;] ptlrpc_main+0xd31/0x1800 [ptlrpc]
 [&amp;lt;ffffffff8106ee50&amp;gt;] ? pick_next_task_fair+0xd0/0x130
 [&amp;lt;ffffffff81539896&amp;gt;] ? schedule+0x176/0x3a0
 [&amp;lt;ffffffffa0b24af0&amp;gt;] ? ptlrpc_main+0x0/0x1800 [ptlrpc]
 [&amp;lt;ffffffff810a138e&amp;gt;] kthread+0x9e/0xc0
 [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
 [&amp;lt;ffffffff810a12f0&amp;gt;] ? kthread+0x0/0xc0
 [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Then other threads waiting for the journal&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;mdt_out01_008 D 000000000000000b     0  6281      2 0x00000080
 ffff8804030a7a40 0000000000000046 0000000000000000 ffff88079c63b5c0
 ffff8807fbe51ad8 ffff88082d91f400 00000bef4b83696a ffff880828bdba80
 ffff8804030a7a10 0000000100c3b96b ffff8803fbd31068 ffff8804030a7fd8
Call Trace:
 [&amp;lt;ffffffffa0fe7fca&amp;gt;] start_this_handle+0x25a/0x480 [jbd2]
 [&amp;lt;ffffffff811781fb&amp;gt;] ? cache_alloc_refill+0x15b/0x240
 [&amp;lt;ffffffff810a1820&amp;gt;] ? autoremove_wake_function+0x0/0x40
 [&amp;lt;ffffffffa0fe83d5&amp;gt;] jbd2_journal_start+0xb5/0x100 [jbd2]
 [&amp;lt;ffffffffa1036a36&amp;gt;] ldiskfs_journal_start_sb+0x56/0xe0 [ldiskfs]
 [&amp;lt;ffffffffa10870f1&amp;gt;] osd_trans_start+0x1e1/0x430 [osd_ldiskfs]
 [&amp;lt;ffffffffa0b7c53c&amp;gt;] out_tx_end+0x9c/0x5d0 [ptlrpc]
 [&amp;lt;ffffffffa0b81ec9&amp;gt;] out_handle+0x11d9/0x18d0 [ptlrpc]
 [&amp;lt;ffffffff8105e9b6&amp;gt;] ? enqueue_task+0x66/0x80
 [&amp;lt;ffffffff8105ab8d&amp;gt;] ? check_preempt_curr+0x6d/0x90
 [&amp;lt;ffffffffa080bc5a&amp;gt;] ? lc_watchdog_touch+0x7a/0x190 [libcfs]
 [&amp;lt;ffffffff8153afce&amp;gt;] ? mutex_lock+0x1e/0x50
 [&amp;lt;ffffffffa0b71eda&amp;gt;] ? req_can_reconstruct+0x6a/0x120 [ptlrpc]
 [&amp;lt;ffffffffa0b790cc&amp;gt;] tgt_request_handle+0x8ec/0x1440 [ptlrpc]
 [&amp;lt;ffffffffa0b25821&amp;gt;] ptlrpc_main+0xd31/0x1800 [ptlrpc]
 [&amp;lt;ffffffff8106ee50&amp;gt;] ? pick_next_task_fair+0xd0/0x130
 [&amp;lt;ffffffff81539896&amp;gt;] ? schedule+0x176/0x3a0
 [&amp;lt;ffffffffa0b24af0&amp;gt;] ? ptlrpc_main+0x0/0x1800 [ptlrpc]
 [&amp;lt;ffffffff810a138e&amp;gt;] kthread+0x9e/0xc0
 [&amp;lt;ffffffff8100c28a&amp;gt;] child_rip+0xa/0x20
 [&amp;lt;ffffffff810a12f0&amp;gt;] ? kthread+0x0/0xc0
 [&amp;lt;ffffffff8100c280&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Then it cause other MDTs can not reconnect to this MDT.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;LustreError: 11-0: soaked-MDT0001-osp-MDT0003: operation mds_connect to node 192.168.1.109@o2ib10 failed: rc = -114
LustreError: 11-0: soaked-MDT0001-osp-MDT0003: operation mds_connect to node 192.168.1.109@o2ib10 failed: rc = -114
LustreError: 11-0: soaked-MDT0001-osp-MDT0003: operation mds_connect to node 192.168.1.109@o2ib10 failed: rc = -114
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Then it cause the whole recovery stuck. I will cook a fix.&lt;/p&gt;</comment>
                            <comment id="169238" author="gerrit" created="Wed, 12 Oct 2016 03:32:48 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/23096&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/23096&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8250&quot; title=&quot;MDT recovery stalled on secondary node&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8250&quot;&gt;&lt;del&gt;LU-8250&lt;/del&gt;&lt;/a&gt; mdd: move linkea prepare out of transaction.&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: b67d62eefef30fbb3365da3382021820985b5e39&lt;/p&gt;</comment>
                            <comment id="169240" author="adilger" created="Wed, 12 Oct 2016 04:54:40 +0000"  >&lt;p&gt;It would be useful to add a check for an open journal handle in the declare code and in the ptlrpc code to detect any other code paths like this.  That can be done relatively easily with ldiskfs, check &lt;tt&gt;current-&amp;gt;journal_info&lt;/tt&gt; in &lt;tt&gt;osd_trans_declare_op()&lt;/tt&gt; and &lt;tt&gt;ptlrpc_set_wait()&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="169303" author="gerrit" created="Wed, 12 Oct 2016 16:44:53 +0000"  >&lt;p&gt;wangdi (di.wang@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/23111&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/23111&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8250&quot; title=&quot;MDT recovery stalled on secondary node&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8250&quot;&gt;&lt;del&gt;LU-8250&lt;/del&gt;&lt;/a&gt; osd: add journal info check&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: c8bc88c85d6f273b18f7741cc9c0cdc1bcfca4d5&lt;/p&gt;</comment>
                            <comment id="169464" author="di.wang" created="Thu, 13 Oct 2016 16:13:23 +0000"  >&lt;p&gt;There is another panic with 23111. Since this is a blocker of 2.9, I will create a new ticket (&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8704&quot; title=&quot;RPC sent inside mdd_create transaction.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8704&quot;&gt;LU-8704&lt;/a&gt;).&lt;/p&gt;</comment>
                            <comment id="170246" author="jgmitter" created="Tue, 18 Oct 2016 21:25:38 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/23111&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/23111&lt;/a&gt; has been moved to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8704&quot; title=&quot;RPC sent inside mdd_create transaction.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8704&quot;&gt;LU-8704&lt;/a&gt; per discussion with Di.&lt;/p&gt;</comment>
                            <comment id="171269" author="gerrit" created="Wed, 26 Oct 2016 23:02:34 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/23096/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/23096/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-8250&quot; title=&quot;MDT recovery stalled on secondary node&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-8250&quot;&gt;&lt;del&gt;LU-8250&lt;/del&gt;&lt;/a&gt; mdd: move linkea prepare out of transaction.&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: b596047cae1d3381cafae9c4132e1a84e99ca9d0&lt;/p&gt;</comment>
                            <comment id="171282" author="pjones" created="Wed, 26 Oct 2016 23:24:03 +0000"  >&lt;p&gt;Landed for 2.9&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="38325">LU-8428</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="40580">LU-8704</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="40741">LU-8714</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="21820" name="console-lola-11.log-20160608.bz2" size="64732" author="heckes" created="Wed, 8 Jun 2016 16:04:06 +0000"/>
                            <attachment id="21821" name="dmesg-lola-11-20160609-0811.bz2" size="26974" author="heckes" created="Wed, 8 Jun 2016 16:04:06 +0000"/>
                            <attachment id="21822" name="lola-11-lustre-log.20160608-0656.bz2" size="261" author="heckes" created="Wed, 8 Jun 2016 16:04:06 +0000"/>
                            <attachment id="21823" name="messages-lola-11.log-20160608.bz2" size="318328" author="heckes" created="Wed, 8 Jun 2016 16:04:06 +0000"/>
                            <attachment id="23168" name="soak.log.2016-06-11-1417.txt.bz2" size="374834" author="heckes" created="Wed, 21 Sep 2016 11:51:31 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzye2v:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>