<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:28:26 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2816] sanity-benchmark test_bonnie slow after 4MB BRW RPC patch</title>
                <link>https://jira.whamcloud.com/browse/LU-2816</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for sarah &amp;lt;sarah@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/91d8ae78-755b-11e2-bf59-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/91d8ae78-755b-11e2-bf59-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_bonnie failed with the following error:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;test failed to respond and timed out&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;From the stdout log, I found the test ran extremely slow compare to the success run:&lt;/p&gt;

&lt;p&gt;It took almost 45 minutes for the failed test:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;16:44:39:== sanity-benchmark test bonnie: bonnie++ == 16:44:30 (1360629870)
16:44:39:min OST has 1809000kB available, using 3845408kB file size
16:44:39:debug=0
16:44:39:running as uid/gid/euid/egid 500/500/500/500, groups:
16:44:39: [touch] [/mnt/lustre/d0_runas_test/f16388]
16:44:39:running as uid/gid/euid/egid 500/500/500/500, groups:
16:44:39: [bonnie++] [-f] [-r] [0] [-s3755] [-n] [10] [-u500:500] [-d/mnt/lustre/d0.bonnie]
16:44:39:Using uid:500, gid:500.
16:47:56:Writing intelligently...done
17:29:18:Rewriting...done
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Only 9 minutes for the success test:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;01:31:59:== sanity-benchmark test bonnie: bonnie++ == 01:31:52 (1359624712)
01:31:59:min OST has 1783256kB available, using 3845408kB file size
01:31:59:debug=0
01:31:59:running as uid/gid/euid/egid 500/500/500/500, groups:
01:31:59: [touch] [/mnt/lustre/d0_runas_test/f22196]
01:31:59:running as uid/gid/euid/egid 500/500/500/500, groups:
01:31:59: [bonnie++] [-f] [-r] [0] [-s3755] [-n] [10] [-u500:500] [-d/mnt/lustre/d0.bonnie]
01:31:59:Using uid:500, gid:500.
01:35:21:Writing intelligently...done
01:40:56:Rewriting...done
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="17582">LU-2816</key>
            <summary>sanity-benchmark test_bonnie slow after 4MB BRW RPC patch</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="jay">Jinshan Xiong</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                            <label>HB</label>
                    </labels>
                <created>Thu, 14 Feb 2013 21:47:00 +0000</created>
                <updated>Wed, 13 Mar 2013 17:19:06 +0000</updated>
                            <resolved>Wed, 13 Mar 2013 15:30:27 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                    <fixVersion>Lustre 2.4.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="52464" author="adilger" created="Fri, 15 Feb 2013 13:50:19 +0000"  >&lt;p&gt;A test run that was fast before the 4MB patch landing (commit b7ee9ce814f37475cdfc6d4def83906b8b41c1d8):&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/sub_tests/e3aca9ba-7665-11e2-bc2f-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/sub_tests/e3aca9ba-7665-11e2-bc2f-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Version  1.96       ------Sequential Output------ --Sequential Input- --Random-
Concurrency   1     -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks--
Machine        Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  /sec %CP
client-24vm2. 3755M           23662   6 12159   7           21428   5  64.2  23
Latency                        4517ms    3199ms              2569ms     655ms
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;After 4MB patch landing the read performance is much slower (commit 86097b23c959c50a81f7b60ce969bcfece8839aa):&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/sub_tests/cbd6a396-73f8-11e2-9df1-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/sub_tests/cbd6a396-73f8-11e2-9df1-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Version  1.96       ------Sequential Output------ --Sequential Input- --Random-
Concurrency   1     -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks--
Machine        Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  /sec %CP
client-29vm6. 3753M           26102  37  1526   4            1858   3 858.0  63
Latency                         430ms    1412ms             59478us   45500us
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I&apos;m not sure why rewrite would be slower, when the initial write performance is actually faster.  It might have something to do with the returned st_blocksize, or possibly tying the readahead chunk size (RAS_INCREASE_STEP) to st_blocksize instead of hard-coding it at 1MB.&lt;/p&gt;</comment>
                            <comment id="52500" author="doug" created="Fri, 15 Feb 2013 18:35:44 +0000"  >&lt;p&gt;Hi Jinshan,&lt;/p&gt;

&lt;p&gt;This is a high blocker.  Can you determine if this is something you can look at and if not, let me know.&lt;/p&gt;</comment>
                            <comment id="52506" author="keith" created="Fri, 15 Feb 2013 19:30:09 +0000"  >&lt;p&gt;The below stuck out at me on a quick review of this LU.  &lt;/p&gt;

&lt;p&gt;Both report (with a little spacing help):&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;OST has 1809000kB available, 
using   3845408kB file size
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;We are creating a file that will not fit in the OST Space for this test? &lt;/p&gt;

&lt;p&gt;It seem this is case for good and bad runs both but it stuck me as odd. &lt;/p&gt;

&lt;p&gt;Also on the OST from the system log it looks like memory pressure was on. &lt;/p&gt;

&lt;p&gt;In the backtraces pte_fault is pretty common path for programs to be stuck in and a few seconds after the systems showstate&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Feb 11 17:46:06 client-16vm4 kernel: cannot allocate a tage (399)
Feb 11 17:46:06 client-16vm4 kernel: cannot allocate a tage (399)
Feb 11 17:46:06 client-16vm4 kernel: cannot allocate a tage (399)
Feb 11 17:46:06 client-16vm4 kernel: cannot allocate a tage (399)
Feb 11 17:46:06 client-16vm4 kernel: cannot allocate a tage (399)
Feb 11 17:46:06 client-16vm4 kernel: cannot allocate a tage (399)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;This system may have just been under memory pressure and swapping the whole time. It is hard to say from the logs for sure it was not OOM but things can slow down alot before that point is reached. &lt;/p&gt;</comment>
                            <comment id="52528" author="adilger" created="Sat, 16 Feb 2013 03:17:30 +0000"  >&lt;p&gt;Keith, note that the OST size reported is for one OST, but there are 3 OSTs in total, hence 3x file size. &lt;/p&gt;

&lt;p&gt;I agree that memory pressure can slow things down, and the 4MB patch will chase larger allocations. Alex has a patch that could help this. &lt;/p&gt;

&lt;p&gt;I&apos;m also wary about the readahead changes I had to make for this patch. They may be causing larger reads than needed.  It might also relate to a larger st_blocksize reported to userspace for files (4MB vs. 2MB).&lt;/p&gt;</comment>
                            <comment id="52850" author="sarah" created="Thu, 21 Feb 2013 21:35:45 +0000"  >&lt;p&gt;also seen in interop test between 2.3.0 server and 2.4 client:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/2746a972-7572-11e2-93d9-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/2746a972-7572-11e2-93d9-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="53271" author="jay" created="Mon, 4 Mar 2013 12:22:07 +0000"  >&lt;p&gt;I&apos;m looking at this issue.&lt;/p&gt;</comment>
                            <comment id="53619" author="jay" created="Fri, 8 Mar 2013 14:48:45 +0000"  >&lt;p&gt;Performance test with 1MB vs 4MB RPC and writethrough cache settings&lt;/p&gt;</comment>
                            <comment id="53620" author="jay" created="Fri, 8 Mar 2013 14:55:06 +0000"  >&lt;p&gt;Please check the following picture for 1MB vs. 4MB RPC w/ Writethrough settings:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;image-wrap&quot; style=&quot;&quot;&gt;&lt;img src=&quot;http://jira.whamcloud.com/secure/attachment/12286/Untitled.png&quot; style=&quot;border: 0px solid black&quot; /&gt;&lt;/span&gt;&lt;/p&gt;

&lt;p&gt;From this testing, if writethrough_cache is disabled on the OST side, the performance is not dropped significantly with 4MB RPC size, 10% less I would say. However, if write through_cache is enabled, I can see huge performance drop with both settings - but 4MB cache is worse.&lt;/p&gt;

&lt;p&gt;So I guess this is all about memory pressure, the memory is almost used up if write through is enabled.&lt;/p&gt;

&lt;p&gt;About the testing env: This test is run on rosso, with 16 CPUs and 48G memory, I used memory disk on OST and applied patch 5164 to saturate QDR network. Iozone was used to generate data w/ 4MB block size.&lt;/p&gt;</comment>
                            <comment id="53623" author="jay" created="Fri, 8 Mar 2013 15:13:37 +0000"  >&lt;p&gt;Initially I was trying to reproduce this issue but I found the results are interesting, so I completed the performance under different settings.&lt;/p&gt;

&lt;p&gt;I will continue working on this issue itself.&lt;/p&gt;</comment>
                            <comment id="53643" author="adilger" created="Sat, 9 Mar 2013 05:46:27 +0000"  >&lt;p&gt;I suspect that 4MB RPC size will be most interesting for DDN RAID-6 devices and similar. Running on RAM disk will not show any of the overhead of seeking or better streaming IO. &lt;/p&gt;</comment>
                            <comment id="53644" author="adilger" created="Sat, 9 Mar 2013 05:51:31 +0000"  >&lt;p&gt;Jinshan, also note that the bug isn&apos;t about how to get 4MB performance to be the same or better than 1MB performance, since the default is going to stay at 1MB for the 2.4 release. Rather, the performance at 1MB has gone down after the 4MB path landed, even though the RPC size stayed at 1MB. Also, it may be that the performance loss is for reads and not writes.&lt;/p&gt;

&lt;p&gt;Please compare performance from before and after the patch, and look at the test times for before/after in Maloo.&lt;/p&gt;

&lt;p&gt;Maybe Alex&apos;s patch to dynamically allocate the OST memory has removed this problem already?  I&apos;m not sure if it recently landed, or is still to be landed...&lt;/p&gt;</comment>
                            <comment id="53646" author="jay" created="Sat, 9 Mar 2013 10:10:47 +0000"  >&lt;p&gt;I see. I will perform the test. I reset the tree to before commit 0f8b7f951e7f43dbf389a431afea2091388a805e(&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1431&quot; title=&quot;Support for larger than 1MB sequential I/O RPCs&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1431&quot;&gt;&lt;del&gt;LU-1431&lt;/del&gt;&lt;/a&gt; ptlrpc: Support for over 1MB bulk I/O RPC) but didn&apos;t see much difference on write performance with 1MB RPC. I will try read.&lt;/p&gt;</comment>
                            <comment id="53658" author="adilger" created="Sun, 10 Mar 2013 04:14:54 +0000"  >&lt;p&gt;Jinshan, note that if you do find a read performance drop with the 4MZb patch, it is worthwhile to see if Alex&apos;s &lt;a href=&quot;http://review.whamcloud.com/5444&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5444&lt;/a&gt; patch will fix it. This should reduce the memory pressure significantly. &lt;/p&gt;

&lt;p&gt;Also, it is worthwhile to do a Maloo search for the sanity-benchmark test-bonnie results and compare the times. It may be that the performance got worse for a time and the got better. &lt;/p&gt;</comment>
                            <comment id="53732" author="jay" created="Mon, 11 Mar 2013 19:31:09 +0000"  >&lt;p&gt;I confirmed that there is significant read performance degradation after 4MB RPC patch is landed, the read performance dropped from ~3GB/s to ~1GB/s. The patch 5444 didn&apos;t help. I&apos;m working on it.&lt;/p&gt;</comment>
                            <comment id="53733" author="jay" created="Mon, 11 Mar 2013 19:31:37 +0000"  >&lt;p&gt;Also, I didn&apos;t see significant performance loss for write case.&lt;/p&gt;</comment>
                            <comment id="53743" author="adilger" created="Mon, 11 Mar 2013 20:36:30 +0000"  >&lt;p&gt;There were some changes to the readahead code as a result of the 4MB patch.  In particular, the changes in &lt;a href=&quot;http://review.whamcloud.com/5230&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5230&lt;/a&gt; were done by me because of PTLRPC_MAX_BRW_PAGES increasing to a large value.  Instead of allowing the readahead window to grow immediately to potentially 32MB, the RAS_INCREASE_STEP is limited to (1 &amp;lt;&amp;lt; inode-&amp;gt;i_blkbits) = (1 &amp;lt;&amp;lt; LL_MAX_BLKSIZE_BITS) = 4MB, instead of the previous 1MB value.&lt;/p&gt;

&lt;p&gt;It would be a simple first test to see if changing LL_MAX_BLKSIZE_BITS to 20 or 21 fixes the read performance issue.  This can&apos;t be the final fix, since this will affect st_blksize and the IO size that &quot;cp&quot; and other tools use, but it will at least help narrow down the problem.&lt;/p&gt;

&lt;p&gt;Probably a better long-term solution is to change inode-&amp;gt;i_blkbits to be (2 * lmm_stripe_size) instead of being tied to PTLRPC_MAX_BRW_PAGES or LL_MAX_BLKSIZE_BITS.  This was originally working in Lustre (lsm_xfersize), but was changed because of some minor NFS issues in &lt;a href=&quot;https://bugzilla.lustre.org/show_bug.cgi?id=6062&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://bugzilla.lustre.org/show_bug.cgi?id=6062&lt;/a&gt;, but should be based on lmm_stripe_size again.  The main difference compared to the old code would be the default lmm_xfersize should be 2x the default stripe size, so that it will match under normal usage (i.e. for SPEC SFS NFS testing), but it can be tuned on a per-file basis for applications that are badly behaved (e.g. &lt;a href=&quot;https://bugzilla.lustre.org/show_bug.cgi?id=12739&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://bugzilla.lustre.org/show_bug.cgi?id=12739&lt;/a&gt;).&lt;/p&gt;</comment>
                            <comment id="53747" author="jay" created="Mon, 11 Mar 2013 21:52:14 +0000"  >&lt;blockquote&gt;
&lt;p&gt;There were some changes to the readahead code as a result of the 4MB patch. In particular, the changes in &lt;a href=&quot;http://review.whamcloud.com/5230&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5230&lt;/a&gt; were done by me because of PTLRPC_MAX_BRW_PAGES increasing to a large value. Instead of allowing the readahead window to grow immediately to potentially 32MB, the RAS_INCREASE_STEP is limited to (1 &amp;lt;&amp;lt; inode-&amp;gt;i_blkbits) = (1 &amp;lt;&amp;lt; LL_MAX_BLKSIZE_BITS) = 4MB, instead of the previous 1MB value.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I reverted this change but it didn&apos;t help. I&apos;m now on the way to revert change of ptlrpc desc change, I guess this would be the major cause of performance loss. Stay tuned.&lt;/p&gt;</comment>
                            <comment id="53752" author="keith" created="Mon, 11 Mar 2013 23:20:09 +0000"  >&lt;p&gt;Do you have the parameters you were using for iozone?  Do you have the raw data for all of the iozone output? &lt;/p&gt;

&lt;p&gt;Also you say your disk was a ram disk?  I guess that explains why cached is so much slower (that is non-intuitive for spindles). &lt;/p&gt;

&lt;p&gt;Also you could add the data for pre 4MB BRW patch data so we know what things looked like before the changes? &lt;/p&gt;</comment>
                            <comment id="53810" author="jay" created="Tue, 12 Mar 2013 15:19:09 +0000"  >&lt;p&gt;I&apos;ve known the root cause of this problem - it&apos;s a simple one:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;diff --git a/lustre/llite/rw.c b/lustre/llite/rw.c
index cbd1c91..8cfa70e 100644
--- a/lustre/llite/rw.c
+++ b/lustre/llite/rw.c
@@ -576,7 +576,7 @@ &lt;span class=&quot;code-keyword&quot;&gt;static&lt;/span&gt; &lt;span class=&quot;code-object&quot;&gt;int&lt;/span&gt; ll_read_ahead_page(&lt;span class=&quot;code-keyword&quot;&gt;const&lt;/span&gt; struct lu_env *env, struct cl_io *io,
  * know what the actual RPC size is.  If &lt;span class=&quot;code-keyword&quot;&gt;this&lt;/span&gt; needs to change, it makes more
  * sense to tune the i_blkbits value &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; the file based on the OSTs it is
  * striped over, rather than having a constant value &lt;span class=&quot;code-keyword&quot;&gt;for&lt;/span&gt; all files here. */
-#define RAS_INCREASE_STEP(inode) (1UL &amp;lt;&amp;lt; inode-&amp;gt;i_blkbits)
+#define RAS_INCREASE_STEP(inode) (1UL &amp;lt;&amp;lt; (inode-&amp;gt;i_blkbits - CFS_PAGE_SHIFT))
 
 &lt;span class=&quot;code-keyword&quot;&gt;static&lt;/span&gt; inline &lt;span class=&quot;code-object&quot;&gt;int&lt;/span&gt; stride_io_mode(struct ll_readahead_state *ras)
 {
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;But there is still a 10%~20% performance loss on read after applying this patch, I&apos;d like to spend more time on it.&lt;/p&gt;</comment>
                            <comment id="53820" author="jay" created="Tue, 12 Mar 2013 18:44:10 +0000"  >&lt;p&gt;I did some research and it turned out that we need to adjust the max_read_ahead_mb and max_read_ahead_per_file_mb to match 4MB size RPC and 4MB increase step. Otherwise the readahead budget will be used up and produce huge number of RA_STAT_MAX_IN_FLIGHT. After I increased those two parameters I didn&apos;t see significant performance loss on read.&lt;/p&gt;

&lt;p&gt;However, adjusting those two parameters need lot of work to do to make sure they are okay with other user cases. So I&apos;d like to set RAS_INCREASE_STEP back to 1MB and after 4MB RPC is enabled by default, we should spend more time on that.&lt;/p&gt;</comment>
                            <comment id="53821" author="jay" created="Tue, 12 Mar 2013 18:57:35 +0000"  >&lt;p&gt;patch is at: &lt;a href=&quot;http://review.whamcloud.com/5691&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/5691&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="53939" author="jay" created="Wed, 13 Mar 2013 15:30:27 +0000"  >&lt;p&gt;Let&apos;s fix it for now. Readahead performance tweak will be addressed in another ticket.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="17882">LU-2957</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="14521">LU-1431</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="12286" name="Untitled.png" size="98097" author="jay" created="Fri, 8 Mar 2013 14:48:45 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvj67:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>6820</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>