<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:16:21 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-1408] single client&apos;s performance regression test</title>
                <link>https://jira.whamcloud.com/browse/LU-1408</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This is a single client performance regression on 2.2 compared to 2.1.2 or 1.8.x.&lt;/p&gt;

&lt;p&gt;I filed &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-744&quot; title=&quot;Single client&amp;#39;s performance degradation on 2.1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-744&quot;&gt;&lt;del&gt;LU-744&lt;/del&gt;&lt;/a&gt; before for another single client&apos;s performance regression, but that regression also happened on 2.1.x as well as 2.2 when the amount of file size is larger than client&apos;s memory size.&lt;/p&gt;

&lt;p&gt;So, this regression might not be related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-744&quot; title=&quot;Single client&amp;#39;s performance degradation on 2.1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-744&quot;&gt;&lt;del&gt;LU-744&lt;/del&gt;&lt;/a&gt;, but there is an another regression on 2.2 even if amount of file size is smaller than client&apos;s memory size.&lt;/p&gt;</description>
                <environment>lustre-2.2, b2_1 and lustre-1.8.7&lt;br/&gt;
CentOS6.2 on both servers and clients </environment>
        <key id="14437">LU-1408</key>
            <summary>single client&apos;s performance regression test</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="hongchao.zhang">Hongchao Zhang</assignee>
                                    <reporter username="ihara">Shuichi Ihara</reporter>
                        <labels>
                    </labels>
                <created>Tue, 15 May 2012 09:41:29 +0000</created>
                <updated>Mon, 2 Jun 2014 15:02:01 +0000</updated>
                            <resolved>Fri, 20 Jul 2012 08:53:40 +0000</resolved>
                                                    <fixVersion>Lustre 2.3.0</fixVersion>
                    <fixVersion>Lustre 2.6.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="38815" author="ihara" created="Tue, 15 May 2012 09:47:30 +0000"  >&lt;p&gt;test script with IOR&lt;/p&gt;</comment>
                            <comment id="38816" author="ihara" created="Tue, 15 May 2012 09:49:48 +0000"  >&lt;p&gt;Here is test scripts and an initial single client performance testing on various lustre version.&lt;br/&gt;
&lt;a href=&quot;http://jira.whamcloud.com/secure/attachment/11360/test-script.sh&quot; class=&quot;external-link&quot; rel=&quot;nofollow&quot;&gt;http://jira.whamcloud.com/secure/attachment/11360/test-script.sh&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;http://jira.whamcloud.com/secure/attachment/11303/lustre-singleclient-comparison.xlsx&quot; class=&quot;external-link&quot; rel=&quot;nofollow&quot;&gt;http://jira.whamcloud.com/secure/attachment/11303/lustre-singleclient-comparison.xlsx&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Servers are running with lustre-2.2 on CentOS6.2, just test each checksum algo with various lustre version.&lt;/p&gt;</comment>
                            <comment id="38905" author="ihara" created="Wed, 16 May 2012 09:26:33 +0000"  >&lt;p&gt;During test for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1408&quot; title=&quot;single client&amp;#39;s performance regression test&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1408&quot;&gt;&lt;del&gt;LU-1408&lt;/del&gt;&lt;/a&gt;, I saw another big performance differences between b2_1 branch and 2.1.2RC0 tag on the single client. Please refer to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1413&quot; title=&quot;difference of single client&amp;#39;s performance between b2_1 and 2.1.2RC0&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1413&quot;&gt;&lt;del&gt;LU-1413&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="38909" author="pjones" created="Wed, 16 May 2012 09:47:36 +0000"  >&lt;p&gt;Oleg&lt;/p&gt;

&lt;p&gt;Could you please suggest some steps here and perhaps someone else can assist in executing them?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="39102" author="ihara" created="Sun, 20 May 2012 06:45:02 +0000"  >
&lt;p&gt;This regression is related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1413&quot; title=&quot;difference of single client&amp;#39;s performance between b2_1 and 2.1.2RC0&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1413&quot;&gt;&lt;del&gt;LU-1413&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;As I commented on &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1413&quot; title=&quot;difference of single client&amp;#39;s performance between b2_1 and 2.1.2RC0&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1413&quot;&gt;&lt;del&gt;LU-1413&lt;/del&gt;&lt;/a&gt;, we also saw the single client&apos;s performance regression on the latest b2_1 branch, but no regression on 2.1.2RC0 branch. The regression started from commit b9cbe3616b6e0b44c7835b1aec65befb85f848f9 (&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; debug: reduce stack usage). Please see &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1413&quot; title=&quot;difference of single client&amp;#39;s performance between b2_1 and 2.1.2RC0&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1413&quot;&gt;&lt;del&gt;LU-1413&lt;/del&gt;&lt;/a&gt; for these test results.&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;http://jira.whamcloud.com/secure/attachment/11303/lustre-singleclient-comparison.xlsx&quot; class=&quot;external-link&quot; rel=&quot;nofollow&quot;&gt;http://jira.whamcloud.com/secure/attachment/11303/lustre-singleclient-comparison.xlsx&lt;/a&gt;&lt;br/&gt;
Above, my testing on b2_2 and b2_1 and 1.8.7, I used 2.1.2RC0 branch for b2_1 testing.&lt;/p&gt;

&lt;p&gt;This is why 2.1.2 numbers were faster than b2_2, because &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; was not landed in 2.1.2RC0 yet, but landed in b2_2.&lt;/p&gt;

&lt;p&gt;In order to verify, I removed commit for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; from b2_2 and measured the performance.&lt;br/&gt;
(IOR with 4 processes with checksum=disable)&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;               WRITE(GB/s)  READ(GB/s)
b2_2              1.4          1.4
b2_2/wo LU969     2.4          2.8
b2_1              1.4          1.4
b2_1/wo LU969     2.7          3.2

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;I will try to run full testing on FDR infiniband.&lt;/p&gt;</comment>
                            <comment id="39103" author="ihara" created="Sun, 20 May 2012 09:24:19 +0000"  >&lt;p&gt;This is benchmark results on 2.1.2RC0, b2_2 and b2_2 without &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; patches.&lt;br/&gt;
RHEL6.2, Lustre-2.2 for servers. RHEL6.2 on an client (96GB memory, 12 CPU cores) with QDR Infiniband.&lt;/p&gt;

&lt;p&gt;All performance were improved by disabling &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; patches.&lt;/p&gt;</comment>
                            <comment id="39107" author="ihara" created="Sun, 20 May 2012 12:41:38 +0000"  >&lt;p&gt;benchmark results with correct parameters.&lt;/p&gt;</comment>
                            <comment id="39108" author="ihara" created="Sun, 20 May 2012 12:48:10 +0000"  >&lt;p&gt;&lt;a href=&quot;http://jira.whamcloud.com/secure/attachment/11413/single-client-perforamnce-LU1408-rev2.xlsx&quot; class=&quot;external-link&quot; rel=&quot;nofollow&quot;&gt;http://jira.whamcloud.com/secure/attachment/11413/single-client-perforamnce-LU1408-rev2.xlsx&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is correct benchmark results with the correct parameters. max_rpcs_in_flight=256 was missing on the previous benchmark results when the checksum was enabled.&lt;br/&gt;
As far as I can see the new results, when I removed &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; patches from b2_2, all numbers were alomost close to 2.1.2RC0&apos;s numbers.&lt;/p&gt;

&lt;p&gt;I will run benchmark on FDR Infiniband to see the maximum single perforamnce on b2_2.&lt;/p&gt;
</comment>
                            <comment id="39109" author="pjones" created="Sun, 20 May 2012 14:23:18 +0000"  >&lt;p&gt;Ihahara&lt;/p&gt;

&lt;p&gt;Thanks for this investigation work!&lt;/p&gt;


&lt;p&gt;Hongchao&lt;/p&gt;

&lt;p&gt;Could you please look into why the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; changes are negatively affecting performance?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="39119" author="hongchao.zhang" created="Mon, 21 May 2012 06:21:03 +0000"  >&lt;p&gt;this could be related to the modification of __CHECK_STACK,&lt;/p&gt;

&lt;p&gt;Hi,Ihahara&lt;br/&gt;
could you please help to check whether there are much more logs in server side when testing with patches of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt;? thanks very much!&lt;/p&gt;</comment>
                            <comment id="39138" author="green" created="Mon, 21 May 2012 14:08:08 +0000"  >&lt;p&gt;I suspect the root problem is because we are filling the debug structure every time now regardless of if it hits or not.&lt;/p&gt;

&lt;p&gt;So perhaps if we move the structure filling until after the check, all should be fine?&lt;/p&gt;

&lt;p&gt;Soemthing like this:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;diff --git a/libcfs/include/libcfs/libcfs_debug.h b/libcfs/include/libcfs/libcfs_debug.h
index 8a366f9..72171ad 100644
--- a/libcfs/include/libcfs/libcfs_debug.h
+++ b/libcfs/include/libcfs/libcfs_debug.h
@@ -203,12 +203,13 @@ static inline int cfs_cdebug_show(unsigned int mask, unsigned int subsystem)
 
 #define __CDEBUG(cdls, mask, format, ...)                               \
 do {                                                                    \
-        LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, cdls);                \
                                                                         \
-        CFS_CHECK_STACK(&amp;amp;msgdata, mask, cdls);                          \
+        if (cfs_cdebug_show(mask, DEBUG_SUBSYSTEM)) {                   \
+                LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, cdls);        \
                                                                         \
-        if (cfs_cdebug_show(mask, DEBUG_SUBSYSTEM))                     \
+                CFS_CHECK_STACK(&amp;amp;msgdata, mask, cdls);                  \
                 libcfs_debug_msg(&amp;amp;msgdata, format, ## __VA_ARGS__);     \
+        }                                                               \
 } while (0)
 
 #define CDEBUG(mask, format, ...) __CDEBUG(NULL, mask, format, ## __VA_ARGS__)
diff --git a/lustre/include/cl_object.h b/lustre/include/cl_object.h
index 47782be..ad1cf7a 100644
--- a/lustre/include/cl_object.h
+++ b/lustre/include/cl_object.h
@@ -1065,9 +1065,9 @@ struct cl_page_operations {
  */
 #define CL_PAGE_DEBUG(mask, env, page, format, ...)                     \
 do {                                                                    \
-        LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);                \
                                                                         \
         if (cfs_cdebug_show(mask, DEBUG_SUBSYSTEM)) {                   \
+                LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);        \
                 cl_page_print(env, &amp;amp;msgdata, lu_cdebug_printer, page);  \
                 CDEBUG(mask, format , ## __VA_ARGS__);                  \
         }                                                               \
@@ -1078,9 +1078,9 @@ do {                                                                    \
  */
 #define CL_PAGE_HEADER(mask, env, page, format, ...)                          \
 do {                                                                          \
-        LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);                      \
                                                                               \
         if (cfs_cdebug_show(mask, DEBUG_SUBSYSTEM)) {                         \
+                LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);              \
                 cl_page_header_print(env, &amp;amp;msgdata, lu_cdebug_printer, page); \
                 CDEBUG(mask, format , ## __VA_ARGS__);                        \
         }                                                                     \
@@ -1789,9 +1789,9 @@ struct cl_lock_operations {
 
 #define CL_LOCK_DEBUG(mask, env, lock, format, ...)                     \
 do {                                                                    \
-        LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);                \
                                                                         \
         if (cfs_cdebug_show(mask, DEBUG_SUBSYSTEM)) {                   \
+                LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);        \
                 cl_lock_print(env, &amp;amp;msgdata, lu_cdebug_printer, lock);  \
                 CDEBUG(mask, format , ## __VA_ARGS__);                  \
         }                                                               \
diff --git a/lustre/include/lu_object.h b/lustre/include/lu_object.h
index 0fd61fb..b97a249 100644
--- a/lustre/include/lu_object.h
+++ b/lustre/include/lu_object.h
@@ -763,9 +763,9 @@ int lu_cdebug_printer(const struct lu_env *env,
  */
 #define LU_OBJECT_DEBUG(mask, env, object, format, ...)                   \
 do {                                                                      \
-        LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);                  \
                                                                           \
         if (cfs_cdebug_show(mask, DEBUG_SUBSYSTEM)) {                     \
+                LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);          \
                 lu_object_print(env, &amp;amp;msgdata, lu_cdebug_printer, object);\
                 CDEBUG(mask, format , ## __VA_ARGS__);                    \
         }                                                                 \
@@ -776,9 +776,9 @@ do {                                                                      \
  */
 #define LU_OBJECT_HEADER(mask, env, object, format, ...)                \
 do {                                                                    \
-        LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);                \
                                                                         \
         if (cfs_cdebug_show(mask, DEBUG_SUBSYSTEM)) {                   \
+                LIBCFS_DEBUG_MSG_DATA_DECL(msgdata, mask, NULL);        \
                 lu_object_header_print(env, &amp;amp;msgdata, lu_cdebug_printer,\
                                        (object)-&amp;gt;lo_header);            \
                 lu_cdebug_printer(env, &amp;amp;msgdata, &quot;\n&quot;);                 \
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="39210" author="hongchao.zhang" created="Tue, 22 May 2012 13:10:09 +0000"  >&lt;p&gt;at Toro(1 client, 1 MDT, 6 OST), several tests of b2_2 and previous b2_1(without the patch of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt;), the read/write&lt;br/&gt;
performance is almost the same, then the affect of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; patch could be related to your test environment,&lt;/p&gt;

&lt;p&gt;could you please help to test with the following patch (against &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; patch)? Thanks in advance!&lt;/p&gt;

&lt;p&gt;diff --git a/libcfs/include/libcfs/linux/libcfs.h b/libcfs/include/libcfs/linux/libcfs.h&lt;br/&gt;
index ce07e80..0dadd84 100644&lt;br/&gt;
&amp;#8212; a/libcfs/include/libcfs/linux/libcfs.h&lt;br/&gt;
+++ b/libcfs/include/libcfs/linux/libcfs.h&lt;br/&gt;
@@ -79,7 +79,8 @@&lt;/p&gt;

&lt;p&gt; #define __CHECK_STACK(msgdata, mask, cdls)                              \&lt;br/&gt;
 do {                                                                    \&lt;/p&gt;
&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;if (unlikely(CDEBUG_STACK() &amp;gt; libcfs_stack)) {                  \&lt;br/&gt;
+        if (unlikely(CDEBUG_STACK() &amp;gt; 3 * THREAD_SIZE / 4 &amp;amp;&amp;amp;            \&lt;br/&gt;
+                     CDEBUG_STACK() &amp;gt; libcfs_stack)) {                  \&lt;br/&gt;
                 libcfs_stack = CDEBUG_STACK();                          \&lt;br/&gt;
                 (msgdata)-&amp;gt;msg_mask = D_WARNING;                        \&lt;br/&gt;
                 (msgdata)-&amp;gt;msg_cdls = NULL;                             \&lt;/li&gt;
&lt;/ul&gt;
</comment>
                            <comment id="39275" author="hongchao.zhang" created="Wed, 23 May 2012 10:34:25 +0000"  >&lt;p&gt;Hi Ihahara,&lt;/p&gt;

&lt;p&gt;could you please help to test this patch (&lt;a href=&quot;http://review.whamcloud.com/#change,2889&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2889&lt;/a&gt;) ATM, thanks!&lt;/p&gt;</comment>
                            <comment id="39277" author="simmonsja" created="Wed, 23 May 2012 11:15:54 +0000"  >&lt;p&gt;ORNL is also testing this patch at this time on our test beds. Will post results soon. Thanks for the patch.&lt;/p&gt;</comment>
                            <comment id="39285" author="ihara" created="Wed, 23 May 2012 14:44:08 +0000"  >&lt;p&gt;Hongchao, I&apos;m very sorry, our test system is shutdowning down on this week, and I was looking for other system, but not luck.&lt;/p&gt;

&lt;p&gt;James, I appreciate your helsp for testing!&lt;/p&gt;</comment>
                            <comment id="39287" author="ihara" created="Wed, 23 May 2012 14:51:10 +0000"  >&lt;p&gt;James, &lt;br/&gt;
For tesitng, there is my IOR command. It&apos;s basic, but the block size is needed to optimize by number of process for less than client&apos;s memory.&lt;br/&gt;
mpirun -np 4 IOR -b Xg -t 1m -F -C -w -r -e -vv -o /lustre/ior.out/file&lt;/p&gt;</comment>
                            <comment id="39290" author="simmonsja" created="Wed, 23 May 2012 15:11:55 +0000"  >&lt;p&gt;Got it. So if the block size is greater than the client memory does this problem still exist. I did a earlier running with a block size much larger than the client memory. In that case I got results similar with and without the patch.&lt;/p&gt;</comment>
                            <comment id="39291" author="ihara" created="Wed, 23 May 2012 15:16:55 +0000"  >&lt;p&gt;Yes, in case of larger than client&apos;s memory size, that problem is still exist and it was filed on &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-744&quot; title=&quot;Single client&amp;#39;s performance degradation on 2.1&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-744&quot;&gt;&lt;del&gt;LU-744&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="39360" author="cliffw" created="Thu, 24 May 2012 22:32:27 +0000"  >&lt;p&gt;Tests run on build 6296, 106 clients  &lt;a href=&quot;http://review.whamcloud.com/#change,2889&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2889&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Kernel: 2.6.32-220.13.1.el6_lustre.g10a847d.x86_64
Lustre: jenkins-arch=x86_64,build_type=server,distro=el6,ib_stack=inkernel-6296-gc1ba127-PRISTINE-2.6.32-220.13.1.el6_lustre.g10a847d.x86_64

parallel-scale.test_iorssf.test_log.hyperion244.log
        clients            = 848 (8 per node)
        aggregate filesize = 848 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        5992.72    5561.75    5837.06     195.23  148.93573 0 848 8 3 0 0 1 0 0 1 1073741824 1048576 910533066752 POSIX 0
 read         5092.44    4625.07    4896.03     197.95  177.65520 0 848 8 3 0 0 1 0 0 1 1073741824 1048576 910533066752 POSIX 0

 Finished: Thu May 24 11:40:39 2012

parallel-scale.test_iorfpp.test_log.hyperion244.log
        clients            = 848 (8 per node)
        aggregate filesize = 848 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        6858.42    6656.15    6766.77      83.66  128.34560 0 848 8 3 1 0 1 0 0 1 1073741824 1048576 910533066752 POSIX 0
 read         6196.60    6069.53    6142.16      53.44  141.38652 0 848 8 3 1 0 1 0 0 1 1073741824 1048576 910533066752 POSIX 0

 Finished: Thu May 24 11:58:08 2012
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="39361" author="cliffw" created="Thu, 24 May 2012 22:34:16 +0000"  >&lt;p&gt;Tests run on build 6296, 50 clients: (includes patch)&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Kernel: 2.6.32-220.13.1.el6_lustre.g10a847d.x86_64
Lustre: jenkins-arch=x86_64,build_type=server,distro=el6,ib_stack=inkernel-6296-gc1ba127-PRISTINE-2.6.32-220.13.1.el6_lustre.g10a847d.x86_64

parallel-scale.test_iorssf.test_log.hyperion244.log
        clients            = 400 (8 per node)
        aggregate filesize = 400 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write         716.36     610.43     659.39      43.61  623.87034 0 400 8 3 0 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0
 read         5380.47    5285.04    5336.81      39.38   76.75416 0 400 8 3 0 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0

 Finished: Thu May 24 14:16:16 2012

parallel-scale.test_iorfpp.test_log.hyperion244.log
        clients            = 400 (8 per node)
        aggregate filesize = 400 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        6763.45     176.72    4552.72    3094.36  813.10982 0 400 8 3 1 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0
 read         5981.72    5336.92    5766.69     303.89   71.23411 0 400 8 3 1 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0

 Finished: Thu May 24 15:02:52 2012
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="39362" author="cliffw" created="Thu, 24 May 2012 22:47:45 +0000"  >&lt;p&gt;Tip of 2.1, includes patch, 50 clients: &lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Kernel: 2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64
Lustre: jenkins-g696f7f2-PRISTINE-2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64

parallel-scale.test_iorssf.test_log.hyperion244.log
        clients            = 400 (8 per node)
        aggregate filesize = 400 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        5179.29    4777.87    4998.78     166.35   82.03194 0 400 8 3 0 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0
 read         5334.62    4381.00    4986.10     429.51   82.80132 0 400 8 3 0 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0

 Finished: Thu May 24 19:36:41 2012

parallel-scale.test_iorfpp.test_log.hyperion244.log
        clients            = 400 (8 per node)
        aggregate filesize = 400 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        6397.31    6191.68    6273.38      89.10   65.30478 0 400 8 3 1 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0
 read         6289.52    5801.98    5992.59     212.76   68.43547 0 400 8 3 1 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0

 Finished: Thu May 24 19:44:19 2012
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt; </comment>
                            <comment id="39364" author="cliffw" created="Thu, 24 May 2012 23:25:52 +0000"  >&lt;p&gt;Tip of 2.1 (build #81)&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Kernel: 2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64
Lustre: jenkins-g696f7f2-PRISTINE-2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64

parallel-scale.test_iorssf.test_log.hyperion244.log
        clients            = 840 (8 per node)
        aggregate filesize = 840 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        5770.81    5199.94    5412.42     254.87  159.26572 0 840 8 3 0 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0
 read         5151.68    5057.77    5117.10      42.14  168.10677 0 840 8 3 0 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0

 Finished: Thu May 24 20:06:38 2012

parallel-scale.test_iorfpp.test_log.hyperion244.log
        clients            = 840 (8 per node)
        aggregate filesize = 840 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        6177.47    5906.30    6009.84     119.62  143.18126 0 840 8 3 1 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0
 read         6144.79    5520.26    5738.26     287.71  150.26389 0 840 8 3 1 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0

 Finished: Thu May 24 20:23:21 2012
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="39372" author="cliffw" created="Fri, 25 May 2012 01:20:32 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/2901&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/2901&lt;/a&gt; -  build without the offending patch 50 clients&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Kernel: 2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64
Lustre: jenkins-ga944961-PRISTINE-2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64

parallel-scale.test_iorssf.test_log.hyperion244.log
        clients            = 400 (8 per node)
        aggregate filesize = 400 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        5489.07    4984.54    5268.71     210.86   77.86901 0 400 8 3 0 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0
 read         5410.81    5175.37    5257.66     108.40   77.93804 0 400 8 3 0 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0

 Finished: Thu May 24 22:09:52 2012

parallel-scale.test_iorfpp.test_log.hyperion244.log
        clients            = 400 (8 per node)
        aggregate filesize = 400 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        6250.72    3615.96    5349.70    1226.26   81.68548 0 400 8 3 1 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0
 read         6329.04    5786.43    6014.79     229.69   68.19637 0 400 8 3 1 0 1 0 0 1 1073741824 1048576 429496729600 POSIX 0

 Finished: Thu May 24 22:18:16 2012
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="39375" author="cliffw" created="Fri, 25 May 2012 02:09:15 +0000"  >&lt;p&gt;105 clients&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
Kernel: 2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64
Lustre: jenkins-ga944961-PRISTINE-2.6.32-220.17.1.el6_lustre.g636ddbf.x86_64

parallel-scale.test_iorssf.test_log.hyperion244.log
        clients            = 840 (8 per node)
        aggregate filesize = 840 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        5372.92    5306.97    5340.69      26.95  161.06181 0 840 8 3 0 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0
 read         5136.44    5110.41    5121.05      11.14  167.96625 0 840 8 3 0 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0

 Finished: Thu May 24 22:39:19 2012

parallel-scale.test_iorfpp.test_log.hyperion244.log
        clients            = 840 (8 per node)
        aggregate filesize = 840 GiB
        blocksize          = 1 GiB
        xfersize           = 1 MiB
 Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt blksiz xsize aggsize API RefNum
 write        5720.72    5560.95    5632.40      66.31  152.73755 0 840 8 3 1 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0
 read         6225.76    5433.65    5778.62     331.35  149.33184 0 840 8 3 1 0 1 0 0 1 1073741824 1048576 901943132160 POSIX 0

 Finished: Thu May 24 22:56:19 2012
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="39385" author="ihara" created="Fri, 25 May 2012 03:57:27 +0000"  >&lt;p&gt;Please test on the single client with multiple thread, instead of multiple nodes.&lt;br/&gt;
The original problem of this is single client&apos;s performance regression. So, even we could get better performance on the multiple clients, still need to make sure really single client&apos;s perforamnce regression is gone by patches. &lt;/p&gt;</comment>
                            <comment id="39400" author="simmonsja" created="Fri, 25 May 2012 08:42:31 +0000"  >&lt;p&gt;Oh I see. Just like Cliff I was not seeing really big difference between with and without &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1408&quot; title=&quot;single client&amp;#39;s performance regression test&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1408&quot;&gt;&lt;del&gt;LU-1408&lt;/del&gt;&lt;/a&gt; patch. Also like Cliff I was testing with multiple nodes of various thread counts per node. Thank for you for clarifying this. Will do another round of testing to see if this makes a difference.&lt;/p&gt;</comment>
                            <comment id="39411" author="simmonsja" created="Fri, 25 May 2012 11:23:39 +0000"  >&lt;p&gt;Finished a run against master (2.2.53). Cliff can you try running on a single client as well. The results I got were:&lt;/p&gt;

&lt;p&gt;*************************************************************************************&lt;br/&gt;
 No &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1408&quot; title=&quot;single client&amp;#39;s performance regression test&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1408&quot;&gt;&lt;del&gt;LU-1408&lt;/del&gt;&lt;/a&gt; patch&lt;br/&gt;
*************************************************************************************&lt;/p&gt;

&lt;p&gt;16 Threads one single node&lt;br/&gt;
Operation  Max (MiB)  Min (MiB)  Mean (MiB)   Std Dev  Max (OPs)  Min (OPs)     Mean (OPs)   Std Dev  Mean (s)&lt;br/&gt;
---------  ---------  ---------  ----------   -------  ---------  ---------     ----------   -------  --------&lt;br/&gt;
write         607.24     568.29      596.03     14.05     151.81     142.07         149.01      3.51 110.01801&lt;br/&gt;
read          723.09     700.82      714.84      8.14     180.77     175.20         178.71      2.04  91.69101&lt;/p&gt;

&lt;p&gt;Max Write: 607.24 MiB/sec (636.74 MB/sec)&lt;br/&gt;
Max Read:  723.09 MiB/sec (758.21 MB/sec)&lt;/p&gt;

&lt;p&gt;*************************************************************************************&lt;br/&gt;
 With &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1408&quot; title=&quot;single client&amp;#39;s performance regression test&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1408&quot;&gt;&lt;del&gt;LU-1408&lt;/del&gt;&lt;/a&gt; patch&lt;br/&gt;
*************************************************************************************&lt;br/&gt;
16 Threads one single node&lt;br/&gt;
Operation  Max (MiB)  Min (MiB)  Mean (MiB)   Std Dev  Max (OPs)  Min (OPs)     Mean (OPs)   Std Dev  Mean (s)&lt;br/&gt;
---------  ---------  ---------  ----------   -------  ---------  ---------     ----------   -------  --------&lt;br/&gt;
write         611.51     445.17      569.68     62.68     152.88     111.29     142.42         15.67 116.72391&lt;br/&gt;
read          771.25     746.29      757.22      9.74     192.81     186.57     189.30          2.44  86.56250&lt;/p&gt;

&lt;p&gt;Max Write: 611.51 MiB/sec (641.21 MB/sec)&lt;br/&gt;
Max Read:  771.25 MiB/sec (808.71 MB/sec)&lt;/p&gt;</comment>
                            <comment id="39412" author="ihara" created="Fri, 25 May 2012 12:14:42 +0000"  >&lt;p&gt;hmm.. it seems to be pretty lower than I had even with patch. Server and clients are connected with QDR Infiniband? and how many CPU cores on the client? and disabled checksum just in case?&lt;/p&gt;</comment>
                            <comment id="39413" author="ihara" created="Fri, 25 May 2012 12:14:42 +0000"  >&lt;p&gt;hmm.. it seems to be pretty lower than I had even with patch. Server and clients are connected with QDR Infiniband? and how many CPU cores on the client? and disabled checksum just in case?&lt;/p&gt;</comment>
                            <comment id="39415" author="simmonsja" created="Fri, 25 May 2012 12:31:08 +0000"  >&lt;p&gt;Yes the fiber is QDR. Each node has 8 Intel Xeon CPU core (mode E5520) at 2.27GHz. Actually for this set of test I had checksums on. I just realized that. No matter; with checksum on or off the results are pretty close to each other.&lt;/p&gt;</comment>
                            <comment id="39434" author="ihara" created="Sat, 26 May 2012 03:58:35 +0000"  >&lt;p&gt;ok, our system is back and tested &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-1408&quot; title=&quot;single client&amp;#39;s performance regression test&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-1408&quot;&gt;&lt;del&gt;LU-1408&lt;/del&gt;&lt;/a&gt; patches.&lt;br/&gt;
Tested on an client (X5675, 3.07GHz, 48GB memory, QDR Infiniband), RHEL6.2, lustre-2.1.2-RC1 (both servers and client). &lt;/p&gt;

&lt;p&gt;Confirmed the patches fix the performance regression problem. Here is test results.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;===== without patch (original 2.1.2-RC1) =====
# mpirun -np 4 /root/IOR -b 8g -t 1m -F -C -w -e -k -vv -o /lustre/file

... snip ...

Max Write: 1401.19 MiB/sec (1469.26 MB/sec)

# mpirun -np 4 /root/IOR -b 8g -t 1m -F -C -r -e -vv -o /lustre/file

... snip ...

Max Read:  1510.69 MiB/sec (1584.07 MB/sec)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;===== with LU-1408 patch =====
# mpirun -np 4 /root/IOR -b 8g -t 1m -F -C -w -e -k -vv -o /lustre/file

... snip ...

Max Write: 2578.54 MiB/sec (2703.80 MB/sec)

# pdsh -a &quot;sync; echo 3 &amp;gt; /proc/sys/vm/drop_caches&quot;

# mpirun -np 4 /root/IOR -b 8g -t 1m -F -C -r -e -vv -o /lustre/file

... snip ...

Max Read:  2663.67 MiB/sec (2793.06 MB/sec)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="39450" author="green" created="Mon, 28 May 2012 05:52:07 +0000"  >&lt;p&gt;Thanks for confirming the results.&lt;/p&gt;

&lt;p&gt;Can you please tell me which patches did you tests? Just the one in gerrit?&lt;/p&gt;</comment>
                            <comment id="39451" author="ihara" created="Mon, 28 May 2012 06:02:02 +0000"  >&lt;p&gt;I did test patch set 2 on &lt;a href=&quot;http://review.whamcloud.com/#change,2889&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2889&lt;/a&gt;&lt;br/&gt;
After my testing is done and verified, I wondered if I could set &quot;verified&quot; flag as one of manual tester, but I couldn&apos;t find how.&lt;br/&gt;
Any advises how we can do that? &lt;/p&gt;</comment>
                            <comment id="39500" author="hongchao.zhang" created="Tue, 29 May 2012 07:18:46 +0000"  >&lt;p&gt;Hi Ihara&lt;/p&gt;

&lt;p&gt;could you please help to test the patch set 3 on &lt;a href=&quot;http://review.whamcloud.com/#change,2889&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2889&lt;/a&gt;, the patch set 2 disables&lt;br/&gt;
the stack check, which is enabled for non x86-64 architecture, thanks!&lt;/p&gt;</comment>
                            <comment id="39712" author="hongchao.zhang" created="Thu, 31 May 2012 06:49:55 +0000"  >&lt;p&gt;Hi Ihara,&lt;/p&gt;

&lt;p&gt;the patch is updated, and could you please help to test patch set 5 on &lt;a href=&quot;http://review.whamcloud.com/#change,2889?&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2889?&lt;/a&gt; Thanks&lt;/p&gt;</comment>
                            <comment id="40003" author="hongchao.zhang" created="Tue, 5 Jun 2012 05:07:51 +0000"  >&lt;p&gt;Hi Ihara,&lt;/p&gt;

&lt;p&gt;Have you tested the patch set 5 at &lt;a href=&quot;http://review.whamcloud.com/#change,2889&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2889&lt;/a&gt;, which is a little different from patch set 2, thanks!&lt;/p&gt;</comment>
                            <comment id="40004" author="ihara" created="Tue, 5 Jun 2012 05:14:52 +0000"  >&lt;p&gt;Hongchao, sorry for delay... I will test patch soon. keep you updates here once my testing is done.&lt;br/&gt;
any advices I can add &quot;manual test&quot; flag on &lt;a href=&quot;http://review.whamcloud.com/#change,2889&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,2889&lt;/a&gt;, after test is done and if result is OK?&lt;/p&gt;</comment>
                            <comment id="40007" author="hongchao.zhang" created="Tue, 5 Jun 2012 05:35:49 +0000"  >&lt;p&gt;Hi Ihara, I have add you as one of the reviewers of the patch, you can add review feedback according to the test result, thanks!&lt;/p&gt;</comment>
                            <comment id="40015" author="ihara" created="Tue, 5 Jun 2012 08:35:47 +0000"  >&lt;p&gt;Hi Hongchao,&lt;/p&gt;

&lt;p&gt;Just tested the patch with 2.1.2-RC1. (since &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-969&quot; title=&quot;2.1 client stack overruns&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-969&quot;&gt;&lt;del&gt;LU-969&lt;/del&gt;&lt;/a&gt; patches were droped on 2.1.2-RC2), the latest patch seems to be OK.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;2.1.2-RC0 without any patches
# mpirun -np 4 /work/tools/bin/IOR -b 8g -t 1m -F -C -w -e -k -vv -o /lustre/file

Max Write: 2537.56 MiB/sec (2660.83 MB/sec)

# pdsh -a &quot;sync; echo 3 &amp;gt; /proc/sys/vm/drop_caches&quot;

# mpirun -np 4 /work/tools/bin/IOR -b 8g -t 1m -F -C -r -k -vv -o /lustre/file

Max Read:  2848.48 MiB/sec (2986.85 MB/sec)

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;2.1.2-RC1 + patches
# mpirun -np 4 /work/tools/bin/IOR -b 8g -t 1m -F -C -w -e -k -vv -o /lustre/file

Max Write: 2468.03 MiB/sec (2587.92 MB/sec)

# pdsh -a &quot;sync; echo 3 &amp;gt; /proc/sys/vm/drop_caches&quot;

# mpirun -np 4 /work/tools/bin/IOR -b 8g -t 1m -F -C -r -k -vv -o /lustre/file

Max Read:  2881.06 MiB/sec (3021.01 MB/sec)

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="42057" author="pjones" created="Fri, 20 Jul 2012 08:53:40 +0000"  >&lt;p&gt;Landed for 2.3&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="14455">LU-1413</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="12054">LU-744</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="12818">LU-969</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="11413" name="single-client-perforamnce-LU1408-rev2.xlsx" size="48947" author="ihara" created="Sun, 20 May 2012 12:41:38 +0000"/>
                            <attachment id="11412" name="single-client-perforamnce-LU1408.xlsx" size="49007" author="ihara" created="Sun, 20 May 2012 09:24:19 +0000"/>
                            <attachment id="11360" name="test-script.sh" size="591" author="ihara" created="Tue, 15 May 2012 09:47:30 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzv6kf:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>4598</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>