<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:13:40 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7990]  Large bulk IO support</title>
                <link>https://jira.whamcloud.com/browse/LU-7990</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt; Add large bulk IO support, e.g. make the ptlrpc be able to size 16MB IO, to improve the performance.&lt;/p&gt;</description>
                <environment></environment>
        <key id="35844">LU-7990</key>
            <summary> Large bulk IO support</summary>
                <type id="2" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11311&amp;avatarType=issuetype">New Feature</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="jay">Jinshan Xiong</assignee>
                                    <reporter username="cengku9660">Gu Zheng</reporter>
                        <labels>
                            <label>patch</label>
                    </labels>
                <created>Wed, 6 Apr 2016 10:34:59 +0000</created>
                <updated>Tue, 8 Dec 2020 05:53:20 +0000</updated>
                            <resolved>Thu, 5 May 2016 19:09:29 +0000</resolved>
                                                    <fixVersion>Lustre 2.9.0</fixVersion>
                    <fixVersion>Lustre 2.11.0</fixVersion>
                    <fixVersion>Lustre 2.10.2</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>15</watches>
                                                                            <comments>
                            <comment id="148071" author="gerrit" created="Thu, 7 Apr 2016 02:04:45 +0000"  >&lt;p&gt;Gu Zheng (gzheng@ddn.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/19366&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/19366&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; rpc: increase bulk size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 17aaa0120997e1b6428c449454e45be9c11ef62f&lt;/p&gt;</comment>
                            <comment id="148072" author="gerrit" created="Thu, 7 Apr 2016 02:04:46 +0000"  >&lt;p&gt;Gu Zheng (gzheng@ddn.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/19367&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/19367&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; ofd: add a parameter to adjust preferred brw size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: bf1ebfadacd9f77c0e46103c1c76b8da9e1cb1b0&lt;/p&gt;</comment>
                            <comment id="148073" author="gerrit" created="Thu, 7 Apr 2016 02:04:47 +0000"  >&lt;p&gt;Gu Zheng (gzheng@ddn.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/19368&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/19368&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; clio: revise readahead to support 16MB IO&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: ed70b211cbed6d74487707746f098be98787b777&lt;/p&gt;</comment>
                            <comment id="149252" author="gerrit" created="Sun, 17 Apr 2016 20:52:12 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/19366/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/19366/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; rpc: increase bulk size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 51b32ac2b9b86a600e92ab03f387717e526153d3&lt;/p&gt;</comment>
                            <comment id="149278" author="shadow" created="Mon, 18 Apr 2016 13:11:09 +0000"  >&lt;p&gt;I like to see DDN work to decrease a overall Lustre performance. Last merged patch replace a kmalloc allocations with vmalloc so dramatically increase vmalloc lock contention for kernels without per CPU vmalloc code. &lt;/p&gt;</comment>
                            <comment id="149325" author="simmonsja" created="Mon, 18 Apr 2016 20:27:04 +0000"  >&lt;p&gt;Ouch that needs to be fixed.&lt;/p&gt;</comment>
                            <comment id="149376" author="jay" created="Tue, 19 Apr 2016 05:05:03 +0000"  >&lt;p&gt;I guess you are referring to the memory allocation in ptlrpc_new_bulk(). With 16M RPC support, the maximum value of nfrags is 4096, and this will lead to a memory allocation of 64KB in maximum. This is less than the maximum memory size kmalloc() can allocate in modern systems. If the memory allocation can;t be fulfilled by kmalloc(), it will then fall down to vmalloc().  I didn&apos;t see any problem with it.&lt;/p&gt;</comment>
                            <comment id="149616" author="paf" created="Thu, 21 Apr 2016 02:10:14 +0000"  >&lt;p&gt;Is it?  What is that maximum size?  I thought it was page size times two?&lt;/p&gt;</comment>
                            <comment id="149797" author="lixi" created="Fri, 22 Apr 2016 03:55:18 +0000"  >&lt;p&gt;Hi Alexey, did you actually see a performance degression? Performance is really important to us, so please open a ticket if so. &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/wink.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt; I do agree with Jinshan on the change of ptlrpc_new_bulk().&lt;/p&gt;</comment>
                            <comment id="149980" author="gerrit" created="Mon, 25 Apr 2016 04:16:47 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/19367/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/19367/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; ofd: add a parameter to adjust preferred brw size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 60032c840c3764755861a61a5ffd6d2ccd5446c9&lt;/p&gt;</comment>
                            <comment id="149991" author="shadow" created="Mon, 25 Apr 2016 07:26:42 +0000"  >&lt;p&gt;Li Xi,&lt;/p&gt;

&lt;p&gt;we have lots problems with vmalloc on RHEL6 kernels. That kernel have a single spinlock to protect a vmalloc allocations.&lt;br/&gt;
So it&apos;s very high loaded spinlock on server side when we need an allocate a bulk for each BRW request.&lt;br/&gt;
Typical performance down is 20% in that case. Probably you run own tests on less loaded system or at system with kernels 3.8 and up, where some vmalloc improvements merged into kernel, in that case you may don&apos;t hit it problem.&lt;/p&gt;</comment>
                            <comment id="149992" author="lixi" created="Mon, 25 Apr 2016 07:31:26 +0000"  >&lt;p&gt;Hi Alexey, thanks for sharing this information. We will run benchmarks to check this too.&lt;/p&gt;</comment>
                            <comment id="149994" author="shadow" created="Mon, 25 Apr 2016 07:54:20 +0000"  >&lt;p&gt;From my point view, you should look to very different side if you want to increase a size of transfer.&lt;br/&gt;
Currently we use a single page per element in transfer, it increase overhead on each parts. Size of array, number WR / SGE requests on IB transfer, ... but two main points exist.&lt;br/&gt;
1) one object &amp;lt;= page size limitation exist only for socklnd, due bug / feauture to use an tcp_sendpage, but if we kill it and return to use kernel_sockmsg or tcp_sendpages directly we able to send more data via single call. IB lnd, GNI lnd, OmniPath .. all able to send up 2Gb per single SGE / WR entry.&lt;br/&gt;
2) OSC always send a data via continues region. Yes, region is continues in virtual memory, but we need anyway remap pages to DMA area when we need a prepare a SGE list, from other side it have 16 continues regions.&lt;/p&gt;

&lt;p&gt;So from my point of view - bulk code should be changed to access an OSC view as input and have a region to SGE page conversion inside of LNet as SGE list don&apos;t have a limitation to number of segments and may better controlled by low level drivers.&lt;br/&gt;
It&apos;s decrease a memory usage on OST side also.&lt;/p&gt;</comment>
                            <comment id="150110" author="simmonsja" created="Mon, 25 Apr 2016 21:42:55 +0000"  >&lt;p&gt;You make valid points Alexey. I think I like to bring up is that it has been recommend that we move from kernel_sock** to using the netlink api. I currently don&apos;t have the cycles for that but its something we should look into.&lt;/p&gt;</comment>
                            <comment id="150144" author="adilger" created="Tue, 26 Apr 2016 01:46:34 +0000"  >&lt;p&gt;In a related question - does the large BRW RPC size also increase the initial ocd_grant request, so that the client can at least form one full-size RPC to the OST after a connect?  The default is currently 2MB (i.e. 2x 1MB RPCs) but this initial grant should probably be increased to 8MB or more when the RPC size is increased.&lt;/p&gt;</comment>
                            <comment id="150195" author="ihara" created="Tue, 26 Apr 2016 13:22:44 +0000"  >&lt;p&gt;Alexey, would you please give us speicfic IO pattern you think performance drops? we had a lot of benchmark and didn&apos;t see any perforamnce regresions so far and singinicant improved performance, instead. &lt;/p&gt;</comment>
                            <comment id="150228" author="adilger" created="Tue, 26 Apr 2016 15:24:15 +0000"  >&lt;p&gt;The &lt;tt&gt;ptlrpc_new_bulk()&lt;/tt&gt; code is changed from calling &lt;tt&gt;OBD_ALLOC()&lt;/tt&gt; to use &lt;tt&gt;OBD_ALLOC_LARGE()&lt;/tt&gt;, which in Lustre 2.7.55 and later is:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;#define OBD_ALLOC_LARGE(ptr, size)                                            \
&lt;span class=&quot;code-keyword&quot;&gt;do&lt;/span&gt; {                                                                          \
        OBD_ALLOC_GFP(ptr, size, GFP_NOFS | __GFP_NOWARN);                    \
        &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (ptr == NULL)                                                      \
                OBD_VMALLOC(ptr, size);                                       \
} &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; (0)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;so it will try &lt;tt&gt;kmalloc()&lt;/tt&gt; first at whatever size is requested, and only if this fails will it fall back to &lt;tt&gt;vmalloc()&lt;/tt&gt; if there isn&apos;t a large-order allocation available to fulfill this request.  For a fragmented 16MB RPC this will be 64KB at most, so it is likely that kmalloc() will succeed to get an order-4 allocation in most cases.  Also, the size of this allocation is driven by the actual RPC size, so it won&apos;t do a larger allocation if large RPCs are not enabled.  I don&apos;t see any significant problem in the current implementation.&lt;/p&gt;

&lt;p&gt;Alexey, there is work being done on the upstream kernel to allow higher-order page allocations for IO, which would reduce the number of fragments in the IB SG list and could optimize both Lustre RPCs and IB-connected storage like SRP.  It would be welcome if you would investigate your proposal to use higher-order allocations or virtual mappings to reduce the SG list and to update the socklnd code to be more efficient.  Patches should probably be submitted under a different LU ticket.&lt;/p&gt;</comment>
                            <comment id="150240" author="jay" created="Tue, 26 Apr 2016 16:27:48 +0000"  >&lt;p&gt;It sounds like there will be significant changes from VFS level for the support of large page order allocation. Right now kernel dirties pages one by one therefore the order of page allocation is always 1.&lt;/p&gt;

&lt;p&gt;Alexey, It&apos;s good to see this problem from a higher level. Please document your proposal in detail so that we can revisit this later after we have sufficient support from kernel.&lt;/p&gt;</comment>
                            <comment id="150242" author="jay" created="Tue, 26 Apr 2016 16:31:44 +0000"  >&lt;p&gt;Andreas - I&apos;m reluctant to increase the initial grant because I would like to be conservative before 16M RPC is widely used. Without this change, the first RPC would be small in size, but after that the RPC size should be 16MB, not a big deal.&lt;/p&gt;</comment>
                            <comment id="150252" author="shadow" created="Tue, 26 Apr 2016 17:08:12 +0000"  >&lt;p&gt;Andreas,&lt;/p&gt;

&lt;p&gt;64k allocations failed easly after some time. If you look into several LNet OOM tickets you may see it. So any aged system including servers will switch to use vmalloc for this code.&lt;/p&gt;

&lt;p&gt;As about SG lists - server side R/O cache kills an IB OFED cache as pool code can&apos;t find older map. I worked on other possibilities  time to time. when looked to rework an o2ib LND.&lt;/p&gt;</comment>
                            <comment id="150264" author="simmonsja" created="Tue, 26 Apr 2016 17:35:01 +0000"  >&lt;p&gt;Over time these larger allocations are going to fail due to the increasing memory fragmentation so the vmalloc penalty will show up.&lt;/p&gt;</comment>
                            <comment id="150978" author="gerrit" created="Wed, 4 May 2016 15:02:01 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/19368/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/19368/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; clio: revise readahead to support 16MB IO&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: d8467ab8a2ca15fbbd5be3429c9cf9ceb0fa78b8&lt;/p&gt;</comment>
                            <comment id="151225" author="jgmitter" created="Thu, 5 May 2016 19:09:29 +0000"  >&lt;p&gt;Patches have landed to master for 2.9.0&lt;/p&gt;</comment>
                            <comment id="167684" author="cheneva1" created="Thu, 29 Sep 2016 02:19:36 +0000"  >&lt;p&gt;Add &lt;a href=&quot;https://jira.whamcloud.com/browse/LUDOC-356&quot; title=&quot;Lustre Manual - Large Bulk IO&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LUDOC-356&quot;&gt;&lt;del&gt;LUDOC-356&lt;/del&gt;&lt;/a&gt; into lustre manual tracking for Large bulk IO feature.&lt;/p&gt;</comment>
                            <comment id="194511" author="gerrit" created="Thu, 4 May 2017 22:38:36 +0000"  >&lt;p&gt;Andreas Dilger (andreas.dilger@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/26955&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/26955&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; llite: increase whole-file readahead to RPC size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 25f80fc94b7988e9d356de200573027bc5178a32&lt;/p&gt;</comment>
                            <comment id="211761" author="gerrit" created="Tue, 24 Oct 2017 07:18:29 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/26955/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/26955/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; llite: increase whole-file readahead to RPC size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 627d0133d9d7bef908544537a8a2b2093719fc72&lt;/p&gt;</comment>
                            <comment id="211820" author="gerrit" created="Tue, 24 Oct 2017 15:16:39 +0000"  >&lt;p&gt;Minh Diep (minh.diep@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/29738&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/29738&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; llite: increase whole-file readahead to RPC size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_10&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 9482cd464b354cd1a3edddb158d42eea2d733419&lt;/p&gt;</comment>
                            <comment id="212091" author="gerrit" created="Thu, 26 Oct 2017 16:09:30 +0000"  >&lt;p&gt;John L. Hammond (john.hammond@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/29738/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/29738/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7990&quot; title=&quot; Large bulk IO support&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7990&quot;&gt;&lt;del&gt;LU-7990&lt;/del&gt;&lt;/a&gt; llite: increase whole-file readahead to RPC size&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_10&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 5c1877c4143bddcd3c5c4f2397db5622c2116fce&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="40172">LUDOC-356</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="45548">LU-9356</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="22860">LU-4533</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="37136">LU-8193</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="32063">LU-7140</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzy6wn:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                                                                                </customfields>
    </item>
</channel>
</rss>