<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:29:48 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2965] MDS evicted OST  after attempting recovery.</title>
                <link>https://jira.whamcloud.com/browse/LU-2965</link>
                <project id="10000" key="LU">Lustre</project>
                    <description></description>
                <environment></environment>
        <key id="17899">LU-2965</key>
            <summary>MDS evicted OST  after attempting recovery.</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="2" iconUrl="https://jira.whamcloud.com/images/icons/priorities/critical.svg">Critical</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="bobijam">Zhenyu Xu</assignee>
                                    <reporter username="simmonsja">James A Simmons</reporter>
                        <labels>
                            <label>LB</label>
                    </labels>
                <created>Thu, 14 Mar 2013 11:33:19 +0000</created>
                <updated>Tue, 23 Apr 2013 13:34:19 +0000</updated>
                            <resolved>Tue, 23 Apr 2013 13:34:19 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="54022" author="simmonsja" created="Thu, 14 Mar 2013 11:47:34 +0000"  >&lt;p&gt;During the large scale test we lost a OSS due to another bug. The OSS came back and entered recovery but the recovery process at some point had the MDS evict the OSS imports. This resulted in clients then being evicted.&lt;br/&gt;
I have the logs from a client, oss, and the mds.&lt;/p&gt;</comment>
                            <comment id="54155" author="pjones" created="Fri, 15 Mar 2013 19:54:08 +0000"  >&lt;p&gt;Bobijam&lt;/p&gt;

&lt;p&gt;Could you please look into this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="54232" author="bobijam" created="Mon, 18 Mar 2013 06:39:51 +0000"  >&lt;p&gt;I saw these Lnet error messages in ost log&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;
Mar  8 21:02:19 widow-oss8b4 kernel: [  680.829820] LNet: No route to 12345-15903@gni via 10.36.227.92@o2ib (all routers down)
Mar  8 21:02:19 widow-oss8b4 kernel: [  680.860759] LNet: No route to 12345-2632@gni via 10.36.227.92@o2ib (all routers down)
Mar  8 21:02:19 widow-oss8b4 kernel: [  680.876671] LNet: Skipped 12 previous similar messages
Mar  8 21:02:20 widow-oss8b4 kernel: [  681.633701] LNet: No route to 12345-19172@gni via 10.36.227.92@o2ib (all routers down)
Mar  8 21:02:20 widow-oss8b4 kernel: [  681.649998] LNet: Skipped 1 previous similar message
Mar  8 21:02:22 widow-oss8b4 kernel: [  683.930515] LNet: No route to 12345-16216@gni via 10.36.227.92@o2ib (all routers down)
Mar  8 21:02:22 widow-oss8b4 kernel: [  683.946526] LNet: Skipped 3 previous similar messages
Mar  8 21:02:25 widow-oss8b4 kernel: [  686.430669] LNet: No route to 12345-245@gni via 10.36.227.92@o2ib (all routers down)
Mar  8 21:02:25 widow-oss8b4 kernel: [  686.446617] LNet: Skipped 18 previous similar messages
Mar  8 21:02:29 widow-oss8b4 kernel: [  691.152386] LNet: No route to 12345-1391@gni via 10.36.227.92@o2ib (all routers down)
Mar  8 21:02:29 widow-oss8b4 kernel: [  691.175839] LNet: Skipped 10 previous similar messages
Mar  8 21:03:27 widow-oss8b4 kernel: [  749.157594] LustreError: 11-0: routed1-MDT0000-lwp-OST00eb: Communicating with 10.36.227.211@o2ib, operation mds_connect failed with -114.
Mar  8 21:03:27 widow-oss8b4 kernel: [  749.198062] LustreError: Skipped 5 previous similar messages
Mar  8 21:05:57 widow-oss8b4 kernel: [  899.157523] LustreError: 11-0: routed1-MDT0000-lwp-OST008b: Communicating with 10.36.227.211@o2ib, operation mds_connect failed with -114.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Is the network on this OSS working during the recovery?&lt;/p&gt;</comment>
                            <comment id="54690" author="simmonsja" created="Fri, 22 Mar 2013 18:59:31 +0000"  >&lt;p&gt;yes it was working. The clients managed to reconnect with no problems. I believe that those down routes are for the other file systems in the fabric that we took off line.&lt;/p&gt;</comment>
                            <comment id="54721" author="bobijam" created="Sat, 23 Mar 2013 19:06:35 +0000"  >&lt;p&gt;The MDS log shows that the OSS reconnection was using a different connection cookie, and MDS reject it, I&apos;m still investigating what could cause this situation, when you lost the OSS what happened to it and what did you to this OSS before you try to mount its OST?&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;already connected client routed1-MDT0000-lwp-OST008b_UUID (at 10.36.227.92@o2ib) with handle 0x1e3d407c957c8e7b. Rejecting client with the same UUID trying to reconnect with handle 0xd64e61dedb5d3691&lt;br/&gt;
Mar  8 20:58:25 widow-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;26179.518445&amp;#93;&lt;/span&gt; format at ldlm_lib.c:999:target_handle_connect doesn&apos;t end in newline&lt;br/&gt;
Mar  8 20:58:25 widow-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;26179.518448&amp;#93;&lt;/span&gt; Lustre: routed1-MDT0000: Rejecting reconnect from the known client routed1-MDT0000-lwp-OST008b_UUID (at 10.36.227.92@o2ib) because it is indicating it is a new client&lt;/p&gt;&lt;/blockquote&gt;</comment>
                            <comment id="54737" author="bobijam" created="Mon, 25 Mar 2013 05:41:50 +0000"  >&lt;p&gt;James,&lt;/p&gt;

&lt;p&gt;what is your initial issue? Is it that a client should not be evicted from OSS? Following is the situation which I interpret from the logs.&lt;/p&gt;

&lt;p&gt;When OSS reconnect MDS, there is some stale export remained on MDS, so that MDS evicted OST (and MDS would clear the stale export, the OST reconnect MDS again should be ok, this is a normal situation).&lt;/p&gt;

&lt;p&gt;and on OSS logs&lt;/p&gt;
&lt;blockquote&gt;

&lt;p&gt;Mar  8 21:06:02 widow-oss8b4 kernel: [  903.540784] Lustre: routed1-OST00eb: Recovery over after 7:32, of 8982 clients 8982 recovered and 0 were evicted.&lt;br/&gt;
Mar  8 21:06:02 widow-oss8b4 kernel: [  903.569906] Lustre: Skipped 5 previous similar messages&lt;br/&gt;
Mar  8 21:31:27 widow-oss8b4 kernel: [ 2428.640322] Lustre: routed1-OST00eb: haven&apos;t heard from client 61445347-9977-82cd-59dd-430903b6625f (at 10.36.227.198@o2ib) in 1354 seconds. I think it&apos;s dead, and I am evicting it. exp ffff880397414000, cur 1362796287 expire 1362795387 last 1362794933&lt;br/&gt;
Mar  8 21:31:30 widow-oss8b4 kernel: [ 2432.006547] Lustre: routed1-OST014b: haven&apos;t heard from client 61445347-9977-82cd-59dd-430903b6625f (at 10.36.227.198@o2ib) in 1357 seconds. I think it&apos;s dead, and I am evicting it. exp ffff88039ae54000, cur 1362796290 expire 1362795390 last 1362794933&lt;br/&gt;
Mar  8 21:31:32 widow-oss8b4 kernel: [ 2433.311741] Lustre: routed1-OST008b: haven&apos;t heard from client 61445347-9977-82cd-59dd-430903b6625f (at 10.36.227.198@o2ib) in 1358 seconds. I think it&apos;s dead, and I am evicting it. exp ffff880398ddd400, cur 1362796291 expire 1362795391 last 1362794933&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;the client get evicted by OSS is because OSS does not hear client&apos;s message for a long time, this is a problem out of the recovery (the client&apos;s network could be in trouble)&lt;/p&gt;

&lt;p&gt;the other part of log in OSS&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Mar  8 22:43:04 widow-oss8b4 kernel: [ 6726.231467] LustreError: 2212:0:(obd_mount.c:2115:server_put_super()) routed1-OST002b: failed to disconnect lwp. (rc=-110)&lt;br/&gt;
Mar  8 22:43:04 widow-oss8b4 kernel: [ 6726.262375] Lustre: Failing over routed1-OST002b&lt;br/&gt;
Mar  8 22:43:05 widow-oss8b4 kernel: [ 6726.700008] Lustre: server umount routed1-OST002b complete&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;this part indicates that OST002b cannot connect to MDS (the same reason as I described, MDS has OST002b stale export), so that OSS failed to mount OST002b, then OSS has to umount it.&lt;/p&gt;</comment>
                            <comment id="54758" author="simmonsja" created="Mon, 25 Mar 2013 15:07:51 +0000"  >&lt;p&gt;The issue was the drop off of clients after recovery was &quot;done&quot;. As you can see all clients did recover but then they started to get evicted.&lt;/p&gt;</comment>
                            <comment id="54763" author="simmonsja" created="Mon, 25 Mar 2013 16:07:31 +0000"  >&lt;p&gt;Client that was evicted after recover. This log is for the node for the whole test shot so you the eviction problem look at 21:30 area.&lt;/p&gt;</comment>
                            <comment id="54764" author="bobijam" created="Mon, 25 Mar 2013 16:08:42 +0000"  >&lt;p&gt;Can specify a client example for me?&lt;/p&gt;</comment>
                            <comment id="54765" author="simmonsja" created="Mon, 25 Mar 2013 16:08:59 +0000"  >&lt;p&gt;I attached a log for one of the nodes that was evicted. Due to these problems recovery took the full 30 minutes instead of finishing quickly.&lt;/p&gt;</comment>
                            <comment id="54824" author="bobijam" created="Tue, 26 Mar 2013 07:20:06 +0000"  >&lt;p&gt;Can you specify the timestamp of the troubled recovery process, from the c4-0c0s5n0.log&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;2013-03-08 21:03:28&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;c4-0c0s5n0&amp;#93;&lt;/span&gt;Lustre: 7635:0:(client.c:1866:ptlrpc_expire_one_request()) Skipped 5 previous similar messages&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;2013-03-08 21:06:04&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;c4-0c0s5n0&amp;#93;&lt;/span&gt;Lustre: routed1-OST00eb-osc-ffff88081e7e4000: Connection restored to routed1-OST00eb (at 10.36.227.92@o2ib)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;2013-03-08 21:06:04&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;c4-0c0s5n0&amp;#93;&lt;/span&gt;Lustre: routed1-OST008b-osc-ffff88081e7e4000: Connection restored to routed1-OST008b (at 10.36.227.92@o2ib)&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;2013-03-08 21:28:08&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;c4-0c0s5n0&amp;#93;&lt;/span&gt;LustreError: 11-0: routed1-MDT0000-mdc-ffff88081e7e4000: Communicating with 10.36.227.211@o2ib, operation obd_ping failed with -107.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;2013-03-08 21:28:08&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;c4-0c0s5n0&amp;#93;&lt;/span&gt;Lustre: routed1-MDT0000-mdc-ffff88081e7e4000: Connection to routed1-MDT0000 (at 10.36.227.211@o2ib) was lost; in progress operations using this service will wait for recovery to complete&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;2013-03-08 21:28:08&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;c4-0c0s5n0&amp;#93;&lt;/span&gt;LustreError: 167-0: routed1-MDT0000-mdc-ffff88081e7e4000: This client was evicted by routed1-MDT0000; in progress operations using this service will fail.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;2013-03-08 21:28:08&amp;#93;&lt;/span&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;c4-0c0s5n0&amp;#93;&lt;/span&gt;Lustre: routed1-MDT0000-mdc-ffff88081e7e4000: Connection restored to routed1-MDT0000 (at 10.36.227.211@o2ib)&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I only see that the client restored connection to OST00&lt;span class=&quot;error&quot;&gt;&amp;#91;eb|8b&amp;#93;&lt;/span&gt; at 2013-03-08 21:06:04, and 24 minutes later lost connection with MDS, and restored to it again soon. Then at 22:32 it is umounted (I think by manually).&lt;/p&gt;</comment>
                            <comment id="55001" author="simmonsja" created="Thu, 28 Mar 2013 13:46:31 +0000"  >&lt;p&gt;We saw the problems around 21:07 for the client evicts. The MDS had these errors which lead to the evictions.&lt;/p&gt;

&lt;p&gt;Mar  8 21:12:54 widow-mds1 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;27048.704079&amp;#93;&lt;/span&gt; LustreError: 0:0:(ldlm_lockd.c:391:waiting_locks_callback()) ### lock callback timer expired after 379s: evicting client at 3167@gni  ns: mdt-ffff880e55b20000 lock: ffff880c3f748600/0x4453d516b35bd568 lrc: 3/0,0 mode: CR/CR res: 8589939214/48084 bits 0x9 rrc: 14 type: IBT flags: 0x200000000020 nid: 3167@gni remote: 0x1f865dc509f72511 expref: 11 pid: 28552 timeout: 4321715043 lvb_type: 0&lt;/p&gt;

&lt;p&gt;To make clear what happened I have attached the MDS logs for the entire test shot.&lt;/p&gt;</comment>
                            <comment id="55163" author="bobijam" created="Mon, 1 Apr 2013 04:33:09 +0000"  >&lt;p&gt;the OSS has completed recover around 21:06, but 25 minutes later, MDS and OSS all reports that they lost connection with some clients&lt;/p&gt;

&lt;div class=&quot;panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;panelHeader&quot; style=&quot;border-bottom-width: 1px;&quot;&gt;&lt;b&gt;oss log&lt;/b&gt;&lt;/div&gt;&lt;div class=&quot;panelContent&quot;&gt;
&lt;p&gt;Mar  8 21:06:02 widow-oss8b4 kernel: [  903.540784] Lustre: routed1-OST00eb: Recovery over after 7:32, of 8982 clients 8982 recovered and 0 were evicted.&lt;br/&gt;
Mar  8 21:06:02 widow-oss8b4 kernel: [  903.569906] Lustre: Skipped 5 previous similar messages&lt;br/&gt;
Mar  8 21:31:27 widow-oss8b4 kernel: [ 2428.640322] Lustre: routed1-OST00eb: haven&apos;t heard from client 61445347-9977-82cd-59dd-430903b6625f (at 10.36.227.198@o2ib) in 1354 seconds. I think it&apos;s dead, and I am evicting it. exp ffff880397414000, cur 1362796287 expire 1362795387 last 1362794933&lt;/p&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;and all logs provided do not show details during the 25 minutes, could something happened about the clients&apos; network?  &lt;/p&gt;</comment>
                            <comment id="55298" author="green" created="Tue, 2 Apr 2013 15:21:49 +0000"  >&lt;p&gt;I just confirmed with James that the actual problem is &quot;even though recovery completed just fine after a server failure, clients were evicted later for some reason&quot;.&lt;/p&gt;

&lt;p&gt;Given the timing of 1354 seconds - that&apos;s exactly how much passed since recovery - appears the clients were not pinging?&lt;br/&gt;
Were you already playing with suppress pinging at the time, I wonder?&lt;/p&gt;</comment>
                            <comment id="55299" author="simmonsja" created="Tue, 2 Apr 2013 15:56:32 +0000"  >&lt;p&gt;Nope. I didn&apos;t test suppress pings at that time. I started the suppress pings test 2 hours later than when all the recover problems happened.&lt;/p&gt;</comment>
                            <comment id="55301" author="simmonsja" created="Tue, 2 Apr 2013 16:00:06 +0000"  >&lt;p&gt;Looking at it logs at 175 seconds after every client recovered some stopped pinging the servers.&lt;/p&gt;</comment>
                            <comment id="55356" author="liwei" created="Wed, 3 Apr 2013 06:18:16 +0000"  >&lt;p&gt;James,&lt;/p&gt;

&lt;p&gt;According to the oss-test-shot.log, only one client was evicted by the OSS around 21:30: &quot;61445347-9977-82cd-59dd-430903b6625f (at 10.36.227.198@o2ib)&quot;.  It seems c4-0c0s5n0 was &quot;1546@gni&quot;, not the o2ib client.  Is the log on the o2ib client still available?&lt;/p&gt;

&lt;p&gt;Also, how did you infer that some clients had stopped pinging the servers, please?&lt;/p&gt;</comment>
                            <comment id="55388" author="simmonsja" created="Wed, 3 Apr 2013 14:49:46 +0000"  >&lt;p&gt;10.36.227.198 was our 2.4 non cray client. I attached the log (spare.log) here.&lt;/p&gt;</comment>
                            <comment id="55407" author="green" created="Wed, 3 Apr 2013 17:45:15 +0000"  >&lt;p&gt;I guess the assumption clients stopped pinging is because they were evicted for not pinging &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/wink.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;</comment>
                            <comment id="55686" author="liwei" created="Mon, 8 Apr 2013 01:06:02 +0000"  >&lt;p&gt;From spare.log (why is it called &quot;spare&quot;, by the way?):&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Mar  8 21:06:01 widow-spare06 kernel: [ 7478.823574] Lustre: routed1-OST00eb-osc-ffff8804319c7000: Connection restored to routed1-OST00eb (at 10.36.227.92@o2ib)&lt;br/&gt;
Mar  8 21:06:01 widow-spare06 kernel: [ 7478.845592] Lustre: routed1-OST01ab-osc-ffff8804319c7000: Connection restored to routed1-OST01ab (at 10.36.227.92@o2ib)&lt;br/&gt;
Mar  8 21:12:33 widow-spare06 kernel: imklog 5.8.10, log source = /proc/kmsg started.&lt;br/&gt;
Mar  8 21:12:33 widow-spare06 kernel: [    0.000000] Initializing cgroup subsys cpuset&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;...&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  8 21:13:34 widow-spare06 kernel: [  120.446824] ipmi device interface&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;EOF&amp;#93;&lt;/span&gt;&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;The client completed recoveries at least with OST00eb and OST01ab, but restarted for some reason.  OST00eb last heard from the client at around 21:08:53---two to three minutes after the recovery finished.  If this was not something else, the timing was close to the 150s ping interval, suggesting the request heard was probably a ping.  The next ping should happen 150s to 300s later, which correspond to 21:11:23 and 21:13:53.  Could this eviction merely because the client was simply not remounted after the restart?  Any log available after 21:13:34?&lt;/p&gt;</comment>
                            <comment id="55801" author="jlevi" created="Mon, 8 Apr 2013 20:13:18 +0000"  >&lt;p&gt;Reducing priority until more information is available to understand the issue.&lt;/p&gt;</comment>
                            <comment id="56807" author="simmonsja" created="Tue, 23 Apr 2013 13:30:16 +0000"  >&lt;p&gt;For our last test shot IR worked flawlessly. I think we can close this ticket. If we encounter the bug again this ticket can be reopened.&lt;/p&gt;</comment>
                            <comment id="56808" author="pjones" created="Tue, 23 Apr 2013 13:34:19 +0000"  >&lt;p&gt;Thanks James - that is good news!&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="12422" name="c4-0c0s5n0.log" size="34126" author="simmonsja" created="Mon, 25 Mar 2013 16:07:31 +0000"/>
                            <attachment id="12297" name="client.log" size="4061" author="simmonsja" created="Thu, 14 Mar 2013 11:47:27 +0000"/>
                            <attachment id="12438" name="mds-test-shot-whole.log" size="1331818" author="simmonsja" created="Thu, 28 Mar 2013 13:46:07 +0000"/>
                            <attachment id="12296" name="mds-test-shot.log" size="8060" author="simmonsja" created="Thu, 14 Mar 2013 11:47:27 +0000"/>
                            <attachment id="12295" name="oss-test-shot.log" size="8158" author="simmonsja" created="Thu, 14 Mar 2013 11:47:27 +0000"/>
                            <attachment id="12476" name="spare.log" size="890104" author="simmonsja" created="Wed, 3 Apr 2013 14:49:46 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvldz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>7230</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>