<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:45:17 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4722] IO Errors during the failover - SLES 11 SP2 - Lustre 2.4.2</title>
                <link>https://jira.whamcloud.com/browse/LU-4722</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;We have applied the patch provided in teh &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3645&quot; title=&quot;Interop 2.1.5 &amp;lt;--&amp;gt; 2.4 Write operations during failover errors out instead of stalling&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3645&quot;&gt;&lt;del&gt;LU-3645&lt;/del&gt;&lt;/a&gt;. And still the customer complains that the issue is can be reproduced.&lt;/p&gt;

&lt;p&gt;Attaching the latest set of logs.&lt;/p&gt;

&lt;p&gt;The issue re-occured on 18th Feb. &lt;/p&gt;</description>
                <environment>SLES 11 SP2 &lt;br/&gt;
Lustre 2.4.2</environment>
        <key id="23491">LU-4722</key>
            <summary>IO Errors during the failover - SLES 11 SP2 - Lustre 2.4.2</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="hongchao.zhang">Hongchao Zhang</assignee>
                                    <reporter username="rganesan@ddn.com">Rajeshwaran Ganesan</reporter>
                        <labels>
                            <label>mn4</label>
                    </labels>
                <created>Thu, 6 Mar 2014 16:04:50 +0000</created>
                <updated>Fri, 12 Dec 2014 13:10:26 +0000</updated>
                            <resolved>Fri, 12 Dec 2014 13:10:26 +0000</resolved>
                                    <version>Lustre 2.4.2</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>4</watches>
                                                                            <comments>
                            <comment id="78601" author="pjones" created="Thu, 6 Mar 2014 17:04:51 +0000"  >&lt;p&gt;Hongchao&lt;/p&gt;

&lt;p&gt;Could you please look into this one?&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="78702" author="hongchao.zhang" created="Fri, 7 Mar 2014 14:33:53 +0000"  >&lt;p&gt;the EIO(-5) found in the logs,&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;pfscn1/logs/kern:Jul 23 17:41:27 mds1 kernel: : LustreError: 3107:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xd72d sub-object on OST idx 0/1: rc = -5
pfscn1/logs/kern:Jul 23 17:41:27 mds1 kernel: : LustreError: 2574:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x5719 sub-object on OST idx 0/1: rc = -5
pfscn1/logs/kern:Jul 23 17:41:27 mds1 kernel: : LustreError: 3647:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x1 sub-object on OST idx 0/1: rc = -5
pfscn1/logs/kern:Aug  1 15:15:35 mds1 kernel: : LustreError: 12170:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xba34 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 15 14:37:35 mds2 kernel: : LustreError: 26395:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x49c2 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 15 14:37:35 mds2 kernel: : LustreError: 10162:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x1896d sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 15 14:37:35 mds2 kernel: : LustreError: 5067:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x2720 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 22 15:14:44 mds2 kernel: : LustreError: 2522:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xcdce sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 22 15:14:44 mds2 kernel: : LustreError: 2523:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xb96 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 22 15:14:44 mds2 kernel: : LustreError: 2738:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x4d74 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 22 15:14:44 mds2 kernel: : LustreError: 21918:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xb61 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 30 15:38:44 mds2 kernel: : LustreError: 17462:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x8bb8 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Jul 31 13:11:33 mds2 kernel: : LustreError: 17452:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x10e14 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Aug 27 15:01:01 mds2 kernel: : LustreError: 20790:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xe06d sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Aug 27 15:01:51 mds2 kernel: : LustreError: 21441:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xe064 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Sep  4 14:40:51 mds2 kernel: : LustreError: 15954:0:(lov_request.c:640:create_done()) qos_remedy_create failed -5
pfscn2/logs/kern:Sep  4 14:40:51 mds2 kernel: : LustreError: 15954:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x9a8/0x0 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Sep 25 15:17:39 pfscn2 kernel: : LustreError: 7308:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x3288 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:51:51 pfscn2 kernel: : LustreError: 18998:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x11052 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:51:51 pfscn2 kernel: : LustreError: 29944:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x10f5a sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:51:51 pfscn2 kernel: : LustreError: 4755:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x18c7 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:52:42 pfscn2 kernel: : LustreError: 18998:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x11053 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:52:42 pfscn2 kernel: : LustreError: 19004:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x10f5b sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:52:42 pfscn2 kernel: : LustreError: 4755:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x18c5 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:53:32 pfscn2 kernel: : LustreError: 19004:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x18c6 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:53:32 pfscn2 kernel: : LustreError: 18998:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x11054 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:53:32 pfscn2 kernel: : LustreError: 29944:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x10f5d sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:54:22 pfscn2 kernel: : LustreError: 29944:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x11055 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:54:22 pfscn2 kernel: : LustreError: 18998:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x10f5e sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct  8 16:54:22 pfscn2 kernel: : LustreError: 19004:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x18c8 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct 21 11:15:02 pfscn2 kernel: : LustreError: 21381:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xfe76 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct 23 10:48:39 pfscn2 kernel: : LustreError: 10250:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x10949 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Oct 23 10:54:28 pfscn2 kernel: : LustreError: 24178:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0xd948 sub-object on OST idx 0/1: rc = -5
pfscn2/logs/kern:Nov  4 21:33:56 pfscn2 kernel: : LustreError: 12037:0:(lov_request.c:579:lov_update_create_set()) error creating fid 0x1377d sub-object on OST idx 0/1: rc = -5
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;


&lt;p&gt;no IO error was found around 18th Feb.&lt;br/&gt;
Does it occur when failing over OST, just like &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3645&quot; title=&quot;Interop 2.1.5 &amp;lt;--&amp;gt; 2.4 Write operations during failover errors out instead of stalling&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3645&quot;&gt;&lt;del&gt;LU-3645&lt;/del&gt;&lt;/a&gt;?&lt;/p&gt;</comment>
                            <comment id="79239" author="rganesan@ddn.com" created="Thu, 13 Mar 2014 14:49:41 +0000"  >&lt;p&gt;Basically customer is running an high intensive IO application, they are doing some testing with SLES 11 and Lustre 2.4.2. They are testing one of the test case, reboot one of the OSS and check the failover. this is not working properly. &lt;/p&gt;


&lt;p&gt;attaching their test scanerio for your reference.&lt;/p&gt;</comment>
                            <comment id="79820" author="rganesan@ddn.com" created="Thu, 20 Mar 2014 07:59:27 +0000"  >&lt;p&gt;We would like to know why the failover is not working properly on the SUSE clients.  Do you need any other logs? &lt;/p&gt;</comment>
                            <comment id="79825" author="hongchao.zhang" created="Thu, 20 Mar 2014 10:52:18 +0000"  >&lt;p&gt;Hi, &lt;br/&gt;
the logs of the OSS (both the primary OSS and the failover OSS) will be useful to track the issue.&lt;br/&gt;
btw, did only SUSE clients have the problem and they encountered -EIO after failing over OSS? if so, the logs of the SUSE clients will be much useful.&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;</comment>
                            <comment id="79955" author="rganesan@ddn.com" created="Thu, 20 Mar 2014 23:24:15 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;We have tried new set of failover and we got the same results. So far we have only SUSE clients and we are seeing the issues. &lt;/p&gt;

&lt;p&gt;All the new log files are added into the ftp.whamcloud.com under the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4722&quot; title=&quot;IO Errors during the failover - SLES 11 SP2 - Lustre 2.4.2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4722&quot;&gt;&lt;del&gt;LU-4722&lt;/del&gt;&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;Attaching the Test Scenario for your reference. (check_SLES11SP3_20140320-1.txt)&lt;/p&gt;</comment>
                            <comment id="80210" author="rganesan@ddn.com" created="Tue, 25 Mar 2014 12:02:16 +0000"  >&lt;p&gt;&lt;br/&gt;
Do you have any update on this issue? Have you had a chance to look at the logs. &lt;/p&gt;</comment>
                            <comment id="80238" author="rganesan@ddn.com" created="Tue, 25 Mar 2014 17:08:52 +0000"  >&lt;p&gt;We have tried the same test on the RHEL client as well and result is same. The issue is easily easily reproduce able on both RHEL/SUSE linux. so, it is not specific to SUSE Linux OS. It must be fixed. &lt;/p&gt;</comment>
                            <comment id="80302" author="hongchao.zhang" created="Wed, 26 Mar 2014 16:08:19 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;yes, we have checked the logs, and it should be not the duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-3645&quot; title=&quot;Interop 2.1.5 &amp;lt;--&amp;gt; 2.4 Write operations during failover errors out instead of stalling&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-3645&quot;&gt;&lt;del&gt;LU-3645&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;this issue is related to the eviction of clients by OST,&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00010000:02000400:7.0:1395331134.233862:0:4638:0:(ldlm_lib.c:1581:target_start_recovery_timer()) pfscdat2-OST0000: Will be in recovery for at least 2:30, or until 1 client reconnects
00010000:02000400:7.0:1395331134.233874:0:4638:0:(ldlm_lib.c:1077:target_handle_connect()) pfscdat2-OST0000: Denying connection for new client 01e5a9c0-2c56-a938-4eea-7658375f5c04 (at 172.26.20.2@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:30
00010000:00080000:18.0F:1395331134.557763:0:4632:0:(ldlm_lib.c:1033:target_handle_connect()) pfscdat2-OST0000: connection from 69c34fd9-73e4-94d1-e43e-5c26d4b1ac9f@172.26.20.5@o2ib recovering/t0 exp (null) cur 1395331134 last 0
00010000:02000400:18.0:1395331134.557774:0:4632:0:(ldlm_lib.c:1077:target_handle_connect()) pfscdat2-OST0000: Denying connection for new client 69c34fd9-73e4-94d1-e43e-5c26d4b1ac9f (at 172.26.20.5@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:29
00010000:00080000:21.0F:1395331137.127888:0:4723:0:(ldlm_lib.c:1033:target_handle_connect()) pfscdat2-OST0000: connection from 0067c3cd-6643-b8bb-721b-08b1b92dccde@172.26.4.1@o2ib recovering/t0 exp (null) cur 1395331137 last 0
00010000:02000400:21.0:1395331137.127899:0:4723:0:(ldlm_lib.c:1077:target_handle_connect()) pfscdat2-OST0000: Denying connection for new client 0067c3cd-6643-b8bb-721b-08b1b92dccde (at 172.26.4.1@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:27
00010000:00080000:7.0:1395331137.164613:0:4638:0:(ldlm_lib.c:1033:target_handle_connect()) pfscdat2-OST0000: connection from 798993fc-cca8-3e6f-6b5d-5ce89dd9836b@172.26.20.1@o2ib recovering/t146029313516 exp (null) cur 1395331137 last 0
00010000:02000400:7.0:1395331137.164620:0:4638:0:(ldlm_lib.c:1077:target_handle_connect()) pfscdat2-OST0000: Denying connection for new client 798993fc-cca8-3e6f-6b5d-5ce89dd9836b (at 172.26.20.1@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:27
00010000:00080000:7.0:1395331137.212815:0:4638:0:(ldlm_lib.c:1033:target_handle_connect()) pfscdat2-OST0000: connection from 49ae061e-87a8-feb4-0a3c-96e0e66336ca@172.26.20.6@o2ib recovering/t0 exp (null) cur 1395331137 last 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;there is only one client found by pfscdat2-OST0000 to be needed to recover, after the recovery completed, the other clients were evicted by the OST.&lt;/p&gt;

&lt;p&gt;I am trying to write an debug patch to collect some debug info to track the issue now.&lt;/p&gt;</comment>
                            <comment id="80369" author="hongchao.zhang" created="Thu, 27 Mar 2014 13:04:09 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;Do you use both pfscn3 and pfscn4 to run the &quot;pfscdat2-OST0000&quot; service and these two nodes use the same device &quot;/dev/mapper/ost_pfscdat2_0&quot;?&lt;/p&gt;

&lt;p&gt;according to the logs, pfscn4 only found one client (the connection from MDT) saved in the OST disk (ost_pfscdat2_0), &lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00010000:02000400:7.0:1395331134.233862:0:4638:0:(ldlm_lib.c:1581:target_start_recovery_timer()) pfscdat2-OST0000: Will be in recovery for at least 2:30, or until 1 client reconnects
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00010000:00080000:23.0:1395331183.996474:0:4723:0:(ldlm_lib.c:748:target_handle_reconnect()) connect export for UUID &apos;pfscdat2-MDT0000-mdtlov_UUID&apos; at ffff880ffbe7a400, cookie 0x6b5630daed2ba3e
00010000:00080000:23.0:1395331183.996481:0:4723:0:(ldlm_lib.c:1033:target_handle_connect()) pfscdat2-OST0000: connection from pfscdat2-MDT0000-mdtlov_UUID@172.26.17.2@o2ib recovering/t150323924993 exp ffff880ffbe7a400 cur 1395331183 last 1395331132
00002000:00080000:23.0:1395331183.996492:0:4723:0:(ofd_obd.c:145:ofd_parse_connect_data()) pfscdat2-OST0000: Received MDS connection for group 0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;there are some write operations from iccn2(172.26.20.2), but not from iccn1, (only two write operations from iccn3).&lt;/p&gt;

&lt;p&gt;could you please help to get the debug log (lctl dk) before rebooting the OSS (pfscn4)?&lt;/p&gt;</comment>
                            <comment id="80444" author="rganesan@ddn.com" created="Fri, 28 Mar 2014 09:43:47 +0000"  >&lt;p&gt;Hello, &lt;/p&gt;

&lt;p&gt;You commented that &quot; I am trying to write an debug patch to collect some debug info to track the issue now.&quot;  May I know when the debug patch is available&quot; &lt;/p&gt;

&lt;p&gt;I can install the debug patch and get the debug logs and also lctl dk output.&lt;/p&gt;</comment>
                            <comment id="80484" author="hongchao.zhang" created="Fri, 28 Mar 2014 16:41:14 +0000"  >&lt;p&gt;Hi,&lt;/p&gt;

&lt;p&gt;the debug patch is at &lt;a href=&quot;http://review.whamcloud.com/#/c/9845/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/9845/&lt;/a&gt;, which prints more logs related to client allocation, deletion and initialization,&lt;br/&gt;
and the &quot;ha&quot; and &quot;inode&quot; should be added to the &quot;debug&quot; by &quot;lctl set_param debug=&apos;+ha +inode&apos;&quot;.&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;</comment>
                            <comment id="80622" author="rganesan@ddn.com" created="Mon, 31 Mar 2014 15:42:21 +0000"  >&lt;p&gt;Hello - Could you please provide us source RPM. &lt;/p&gt;

&lt;p&gt;Does it need to be applied on all the MDS/OSSes or only required for the Lustre clients, &lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;</comment>
                            <comment id="80954" author="rganesan@ddn.com" created="Thu, 3 Apr 2014 16:11:01 +0000"  >&lt;p&gt;Hello  - Could you please provide me the RPM with the patch and also, please let me know whether its a client patch or server patch.&lt;/p&gt;


&lt;p&gt;Thanks,&lt;br/&gt;
Rajesh&lt;/p&gt;</comment>
                            <comment id="81031" author="hongchao.zhang" created="Fri, 4 Apr 2014 07:26:49 +0000"  >&lt;p&gt;Hi Rajesh,&lt;/p&gt;

&lt;p&gt;you can download the RPMs from &lt;a href=&quot;http://build.whamcloud.com/job/lustre-reviews/22788/arch=x86_64,build_type=server,distro=el6,ib_stack=inkernel/artifact/artifacts/RPMS/x86_64/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-reviews/22788/arch=x86_64,build_type=server,distro=el6,ib_stack=inkernel/artifact/artifacts/RPMS/x86_64/&lt;/a&gt;&lt;br/&gt;
and it&apos;s only for server side.&lt;/p&gt;

&lt;p&gt;Regards,&lt;br/&gt;
Hongchao&lt;/p&gt;</comment>
                            <comment id="81258" author="rganesan@ddn.com" created="Wed, 9 Apr 2014 09:55:38 +0000"  >&lt;p&gt;This is an EXAScaler setup and we use MOFED. Could you please provide the RPM with MOFED. &lt;/p&gt;</comment>
                            <comment id="81604" author="rganesan@ddn.com" created="Tue, 15 Apr 2014 13:31:30 +0000"  >&lt;p&gt;the patch have been applied, we have re-tried the test, results are same. attaching teh logs for your reference. &lt;/p&gt;</comment>
                            <comment id="81816" author="hongchao.zhang" created="Thu, 17 Apr 2014 09:17:15 +0000"  >&lt;p&gt;Hi Rajesh,&lt;/p&gt;

&lt;p&gt;the uploaded logs seems to not be compressed correctly&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;hongchao@eric:/scratch/ftp/uploads/LU-4722$ file 2014-04-14-client_lctl_dk_20140414.tgz 
2014-04-14-client_lctl_dk_20140414.tgz: HTML document text
hongchao@eric:/scratch/ftp/uploads/LU-4722$ file client_messages_20140414.tgz 
client_messages_20140414.tgz: HTML document text
hongchao@eric:/scratch/ftp/uploads/LU-4722$ file server_lctl_dk_20140414.tgz 
server_lctl_dk_20140414.tgz: HTML document text
hongchao@eric:/scratch/ftp/uploads/LU-4722$ 
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;could you please check it? Thanks&lt;/p&gt;</comment>
                            <comment id="81820" author="hongchao.zhang" created="Thu, 17 Apr 2014 11:33:54 +0000"  >
&lt;p&gt;In pfscn3,&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Apr 14 16:20:04 pfscn3 kernel: : LDISKFS-fs (dm-9): mounted filesystem with ordered data mode. quota=on. Opts:
Apr 14 16:20:05 pfscn3 kernel: : Lustre: pfscdat2-OST0000: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-450
Apr 14 16:20:14 pfscn3 kernel: : Lustre: pfscdat2-OST0000: Will be in recovery for at least 2:30, or until 27 clients reconnect
Apr 14 16:22:44 pfscn3 kernel: : Lustre: pfscdat2-OST0000: recovery is timed out, evict stale exports
Apr 14 16:22:44 pfscn3 kernel: : Lustre: pfscdat2-OST0000: disconnecting 26 stale clients
Apr 14 16:22:44 pfscn3 kernel: : Lustre: pfscdat2-OST0000: Recovery over after 2:30, of 27 clients 1 recovered and 26 were evicted.
Apr 14 16:22:44 pfscn3 kernel: : Lustre: pfscdat2-OST0000: deleting orphan objects from 0x0:29711772 to 0x0:29714801
Apr 14 16:25:28 pfscn3 kernel: : Lustre: Failing over pfscdat2-OST0000
Apr 14 16:25:29 pfscn3 kernel: : Lustre: server umount pfscdat2-OST0000 complete
Apr 14 16:25:29 pfscn3 kernel: : LustreError: 137-5: pfscdat2-OST0000_UUID: not available for connect from 172.26.17.2@o2ib (no target)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;In pfscn4&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Apr 11 08:40:02 pfscn4 kernel: : LDISKFS-fs (dm-11): mounted filesystem with ordered data mode. quota=on. Opts:
Apr 11 08:40:02 pfscn4 kernel: : LustreError: 137-5: pfscdat2-OST0000_UUID: not available for connect from 172.26.16.28@o2ib (no target)
Apr 11 08:40:07 pfscn4 kernel: : Lustre: 3897:0:(client.c:1869:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent Apr 11 08:41:17 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Recovery over after 0:36, of 36 clients 36 recovered and 0 were evicted.
Apr 11 08:41:20 pfscn4 kernel: : Lustre: pfscdat2-OST0000: deleting orphan objects from 0x0:29380160 to 0x0:29380273

...

Apr 14 16:19:52 pfscn4 kernel: : Lustre: Failing over pfscdat2-OST0000
Apr 14 16:19:53 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Not available for connect from 172.26.20.7@o2ib (stopping)
Apr 14 16:19:53 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Not available for connect from 172.26.20.8@o2ib (stopping)
Apr 14 16:19:54 pfscn4 kernel: : LustreError: 137-5: pfscdat2-OST0000_UUID: not available for connect from 172.26.20.3@o2ib (no target)
Apr 14 16:19:57 pfscn4 kernel: : LustreError: 137-5: pfscdat2-OST0000_UUID: not available for connect from 172.26.17.2@o2ib (no target)
Apr 14 16:19:57 pfscn4 kernel: : LustreError: Skipped 1 previous similar message
Apr 14 16:19:58 pfscn4 kernel: : Lustre: server umount pfscdat2-OST0000 complete

...

Apr 14 16:25:43 pfscn4 kernel: : LDISKFS-fs (dm-11): mounted filesystem with ordered data mode. quota=on. Opts:
Apr 14 16:25:44 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Imperative Recovery enabled, recovery window shrunk from 300-900 down to 150-450
Apr 14 16:25:47 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Will be in recovery for at least 2:30, or until 1 client reconnects
Apr 14 16:25:47 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Denying connection for new client 90192127-1d80-d056-258f-193df5a6691b (at 172.26.4.4@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:30
Apr 14 16:25:49 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Denying connection for new client 9b4ef354-d4a2-79a9-196f-2666496727d6 (at 172.26.20.9@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:28
Apr 14 16:25:49 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Denying connection for new client b1ee0c39-c7c4-6f09-d8ec-5bf4d696e919 (at 172.26.4.1@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:27
Apr 14 16:25:49 pfscn4 kernel: : Lustre: Skipped 2 previous similar messages
Apr 14 16:25:51 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Denying connection for new client b1ee0c39-c7c4-6f09-d8ec-5bf4d696e919 (at 172.26.4.1@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:25
Apr 14 16:25:54 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Denying connection for new client 42dd99f1-05b3-2c90-ca73-e27b00e04746 (at 172.26.4.8@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:23
Apr 14 16:25:54 pfscn4 kernel: : Lustre: Skipped 10 previous similar messages
Apr 14 16:26:14 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Denying connection for new client b1ee0c39-c7c4-6f09-d8ec-5bf4d696e919 (at 172.26.4.1@o2ib), waiting for all 1 known clients (0 recovered, 0 in progress, and 0 evicted) to recover in 2:02
Apr 14 16:26:14 pfscn4 kernel: : Lustre: Skipped 1 previous similar message
Apr 14 16:26:20 pfscn4 kernel: : Lustre: pfscdat2-OST0000: Recovery over after 0:33, of 1 clients 1 recovered and 0 were evicted.
Apr 14 16:26:20 pfscn4 kernel: : Lustre: pfscdat2-OST0000: deleting orphan objects from 0x0:29711772 to 0x0:29714833
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;


&lt;p&gt;In pfscn3&lt;br/&gt;
the device &quot;dm-9&quot; was mounted at 16:20:04 as pfscdat2-OST0000, during recovery, Lustre indeed found there were 27 clients (26 normal clients,&lt;br/&gt;
1 client from MDT), but it seems these 26 normal clients didn&apos;t recover with pfscn3 (the eviction condition after recovery timeout is either the client&lt;br/&gt;
didn&apos;t need recovery or there was no queued replay request). then these clients were deleted and pfscdat2-OST0000 was unmounted at 16:25:29.&lt;/p&gt;

&lt;p&gt;In pfscn4&lt;br/&gt;
the device &quot;dm-11&quot; was mounted at 16:25:43 as pfscdat2-OST0000, but it didn&apos;t contain client records, then these clients thought it were evicted.&lt;/p&gt;

&lt;p&gt;then the problem could be why these clients didn&apos;t connect to pfscn3 to recover?&lt;/p&gt;</comment>
                            <comment id="81915" author="hongchao.zhang" created="Fri, 18 Apr 2014 04:35:54 +0000"  >&lt;p&gt;Hi Rajesh,&lt;/p&gt;

&lt;p&gt;Could you please check the content in &quot;/proc/fs/lustre/osc/pfscdat2-OST0000-osc-XXXX/import&quot; at one of your client nodes to verify whether&lt;br/&gt;
the failover_nids contains &quot;172.26.4.3@o2ib&quot; and &quot;172.26.4.4@o2ib&quot;?&lt;/p&gt;

&lt;p&gt;Thanks!&lt;/p&gt;</comment>
                            <comment id="82467" author="rganesan@ddn.com" created="Fri, 25 Apr 2014 12:37:17 +0000"  >&lt;p&gt;We did the erase config and added the failover IP using tunefs, &lt;/p&gt;

&lt;p&gt;1. We would like to know why some of the OSTs are having the failover IP multiple times, is there any known bug in 2.4.1.&lt;br/&gt;
2. Having multiple time will  it cause any negative impact?&lt;/p&gt;

&lt;p&gt;/proc/fs/lustre/osc/pfs2dat2-OST0012-osc-ffff881033429400/import: &lt;br/&gt;
failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;172.26.8.15@o2ib, 172.26.8.15@o2ib, 172.26.8.14@o2ib&amp;#93;&lt;/span&gt;&lt;/p&gt;


&lt;p&gt;/proc/fs/lustre/osc/pfs2dat2-OST0013-osc-ffff881033429400/import: &lt;br/&gt;
failover_nids: &lt;span class=&quot;error&quot;&gt;&amp;#91;172.26.8.15@o2ib, 172.26.8.14@o2ib&amp;#93;&lt;/span&gt;&lt;/p&gt;</comment>
                            <comment id="82739" author="hongchao.zhang" created="Tue, 29 Apr 2014 14:13:43 +0000"  >&lt;p&gt;what is the command line to create the failover nid? &lt;/p&gt;

&lt;p&gt;Could you please dump the config file of your system,&lt;br/&gt;
at MGS node:&lt;br/&gt;
lctl&amp;gt;dl                        (will show the device list, it will show the device number at the first column)&lt;br/&gt;
lctl&amp;gt;device #MGS     (MGS index, say, 1)&lt;br/&gt;
lctl&amp;gt;dump_cfg pfs2dat2-client       (it will dump the config to syslog)&lt;/p&gt;

&lt;p&gt;normally one nid (Node Id) will have only one UUID to represent it (one UUID could have more nid)&lt;br/&gt;
Having duplicated failover nid could cause some recovery problem, for it will use these nids one by one and it is possible to miss the recovery window.&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;</comment>
                            <comment id="82983" author="rganesan@ddn.com" created="Thu, 1 May 2014 12:47:43 +0000"  >
&lt;p&gt;tunefs.lustre --erase-params --mgsnode=172.26.8.12@o2ib &lt;br/&gt;
--mgsnode=172.26.8.13@o2ib --servicenode=172.26.8.14@o2ib &lt;br/&gt;
--servicenode=172.26.8.15@o2ib /dev/mapper/ost_pfs2dat2_0&lt;br/&gt;
tunefs.lustre --erase-params --mgsnode=172.26.8.12@o2ib &lt;br/&gt;
--mgsnode=172.26.8.13@o2ib --servicenode=172.26.8.14@o2ib &lt;br/&gt;
--servicenode=172.26.8.15@o2ib /dev/mapper/ost_pfs2dat2_1&lt;/p&gt;

&lt;p&gt;2. I have uploaded the files in the FTP server&lt;/p&gt;

&lt;p&gt;3. only the OSC on the MDT is affected, having the duplicate entries.  Anyway, we should find the reason for the wrong behaviour&lt;br/&gt;
of the servers.&lt;/p&gt;</comment>
                            <comment id="83134" author="hongchao.zhang" created="Sun, 4 May 2014 11:14:34 +0000"  >&lt;p&gt;the configs seems fine.&lt;br/&gt;
Do you also regenerate the config of pfscdat2? it adds &quot;172.26.17.4@o2ib&quot; and &quot;172.26.17.3@o2ib&quot; as target node, and it was &quot;172.26.4.4@o2ib&quot; and &quot;172.26.4.3@o2ib&quot; previously.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;cmd=cf010 marker=10(0x1)pfscdat2-OST0000 &apos;add osc&apos;
cmd=cf005 nid=172.26.17.4@o2ib(0x50000ac1a1104) 0:(null)  1:172.26.17.4@o2ib
cmd=cf001 0:pfscdat2-OST0000-osc  1:osc  2:pfscdat2-clilov_UUID
cmd=cf003 0:pfscdat2-OST0000-osc  1:pfscdat2-OST0000_UUID  2:172.26.17.4@o2ib
cmd=cf005 nid=172.26.17.4@o2ib(0x50000ac1a1104) 0:(null)  1:172.26.17.4@o2ib
cmd=cf00b 0:pfscdat2-OST0000-osc  1:172.26.17.4@o2ib
cmd=cf005 nid=172.26.17.3@o2ib(0x50000ac1a1103) 0:(null)  1:172.26.17.3@o2ib
cmd=cf00b 0:pfscdat2-OST0000-osc  1:172.26.17.3@o2ib 
cmd=cf00d 0:pfscdat2-clilov  1:pfscdat2-OST0000_UUID  2:0  3:1
cmd=cf010 marker=10(0x2)pfscdat2-OST0000 &apos;add osc&apos;
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Does the issue occur again in the new generated config?&lt;/p&gt;

&lt;p&gt;btw, since only the OSC(OSP) on the MDT is affected, could you please dump the config of MDT (the config-uuid-name is fsname-MDT0000, say, pfscdat2-MDT0000).&lt;br/&gt;
Thanks!&lt;/p&gt;</comment>
                            <comment id="83173" author="hongchao.zhang" created="Mon, 5 May 2014 11:00:12 +0000"  >&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;/proc/fs/lustre/osc/pfs2dat2-OST0012-osc-ffff881033429400/import:
failover_nids: [172.26.8.15@o2ib, 172.26.8.15@o2ib, 172.26.8.14@o2ib]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Do you mount Lustre client at MDT? normally, the OSC(is OSP actually, and there is a symlink in /proc/fs/lustre/osc/&lt;br/&gt;
for each entry in /proc/fs/lustre/osp/)  name for MDT is (fsname)-OSTXXXX-osc-MDT0000,&lt;br/&gt;
the name for client is (fsname)&lt;del&gt;OSTXXXX-osc&lt;/del&gt;(address of superblock).&lt;/p&gt;</comment>
                            <comment id="83281" author="rganesan@ddn.com" created="Tue, 6 May 2014 11:49:48 +0000"  >&lt;p&gt;There seems to be some confusion about our systems. The system with&lt;br/&gt;
prefix pfsc and IP addresses 172.26.4.x is our test system and the&lt;br/&gt;
system with prefix pfs2 and IP addresses 172.26.17.x is our production&lt;br/&gt;
system. We had seen the issue on both systems and therefore provided&lt;br/&gt;
logs from both systems. The configuration of both systems should be&lt;br/&gt;
very similar and also the config was newly generated on both systems.&lt;/p&gt;

&lt;p&gt;After the config was newly generated the remaining issue is that&lt;br/&gt;
duplicate IP addresses appear as failover_nids on the servers only.&lt;br/&gt;
This appears for /proc/fs/lustre/osp/*/import on the MDS but&lt;br/&gt;
astonishingly only on pfsc and not on pfs2. It also appears for&lt;br/&gt;
/proc/fs/lustre/osc/*/import if we mount the file systems on servers&lt;br/&gt;
but astonishingly only for some OSTs. This appears if we mount the file&lt;br/&gt;
system on an OSS, on an MDS which is currently MDT for another file&lt;br/&gt;
system or on an currently unused (failover) MDS.&lt;/p&gt;


&lt;p&gt;and also, uploading the logs into ftp site&lt;/p&gt;</comment>
                            <comment id="83301" author="hongchao.zhang" created="Tue, 6 May 2014 15:02:08 +0000"  >&lt;p&gt;there is no error in these configs.&lt;/p&gt;

&lt;p&gt;Could you please collect the debug logs(lctl dk &amp;gt;XXX.log) at the problematic node just after mounting the client (make sure the &quot;ha&quot; is contained in &quot;/proc/sys/lnet/debug&quot;)?&lt;br/&gt;
Thanks very much!&lt;/p&gt;</comment>
                            <comment id="83486" author="rganesan@ddn.com" created="Thu, 8 May 2014 09:36:31 +0000"  >&lt;p&gt;Hello Hongchao,&lt;/p&gt;

&lt;p&gt;I have uploaded the requested log files into ftp.whamcloud.com:/uploads/&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4722&quot; title=&quot;IO Errors during the failover - SLES 11 SP2 - Lustre 2.4.2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4722&quot;&gt;&lt;del&gt;LU-4722&lt;/del&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2014-05-08-SR30502_pfs2n17.llog.gz&lt;/p&gt;

&lt;p&gt;Thanks,&lt;br/&gt;
Rajesh&lt;/p&gt;</comment>
                            <comment id="83513" author="hongchao.zhang" created="Thu, 8 May 2014 15:01:56 +0000"  >&lt;p&gt;there is a bug in obd_str2uuid,&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt; &lt;span class=&quot;code-keyword&quot;&gt;static&lt;/span&gt; inline void obd_str2uuid(struct obd_uuid *uuid, &lt;span class=&quot;code-keyword&quot;&gt;const&lt;/span&gt; &lt;span class=&quot;code-object&quot;&gt;char&lt;/span&gt; *tmp)
 {
        strncpy((&lt;span class=&quot;code-object&quot;&gt;char&lt;/span&gt; *)uuid-&amp;gt;uuid, tmp, sizeof(*uuid));
        uuid-&amp;gt;uuid[sizeof(*uuid) - 1] = &lt;span class=&quot;code-quote&quot;&gt;&apos;\0&apos;&lt;/span&gt;;
 }
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;it take &quot;tmp&quot; also as a implicit &quot;obd_uuid&quot; type, but it isn&apos;t in all cases, such as in &quot;class_add_uuid&quot;,  the &quot;tmp&quot; is&lt;br/&gt;
&quot;lustre_cfg_string(lcfg, 1)&quot;,  and obd_str2uuid will copy some undefined data beyond the &quot;tmp&quot; to &quot;uuid&quot; and could cause two same&lt;br/&gt;
&quot;uuid&quot; in config were thought to be different.&lt;/p&gt;

&lt;p&gt;the patch against b2_4 is tracked at &lt;a href=&quot;http://review.whamcloud.com/#/c/10269/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/10269/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hi Rajesh,&lt;br/&gt;
Could you please try the patch in your site?&lt;br/&gt;
Thanks!&lt;/p&gt;</comment>
                            <comment id="84241" author="hongchao.zhang" created="Fri, 16 May 2014 13:54:15 +0000"  >&lt;p&gt;Hi Rajesh,&lt;/p&gt;

&lt;p&gt;What is the result of the test?&lt;/p&gt;

&lt;p&gt;Thanks.&lt;/p&gt;</comment>
                            <comment id="84245" author="rganesan@ddn.com" created="Fri, 16 May 2014 13:59:02 +0000"  >&lt;p&gt;We are in the process of applying the patch. I will get back to you with the results. &lt;/p&gt;</comment>
                            <comment id="100222" author="pjones" created="Fri, 28 Nov 2014 13:49:21 +0000"  >&lt;p&gt;Any update Rajesh?&lt;/p&gt;</comment>
                            <comment id="100231" author="rganesan@ddn.com" created="Fri, 28 Nov 2014 16:06:34 +0000"  >&lt;p&gt;We can close this LU.  Customer had upgraded the Server and Clients and they don&apos;t see this issue. &lt;/p&gt;</comment>
                            <comment id="100242" author="pjones" created="Fri, 28 Nov 2014 19:48:05 +0000"  >&lt;p&gt;Rajesh&lt;/p&gt;

&lt;p&gt;To be clear, do you mean upgraded to a newer Lustre version or upgraded to use the patch supplied?&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="100808" author="pjones" created="Fri, 5 Dec 2014 11:12:18 +0000"  >&lt;p&gt;Rajesh?&lt;/p&gt;</comment>
                            <comment id="100809" author="rganesan@ddn.com" created="Fri, 5 Dec 2014 11:20:44 +0000"  >&lt;p&gt;We have upgraded both server side and client side.&lt;/p&gt;

&lt;p&gt;1. On the server side customer upgraded into 2.4.3 with the Patch &lt;/p&gt;


&lt;p&gt;And now they don&apos;t see the issue, and it can be closed. &lt;/p&gt;</comment>
                            <comment id="100810" author="pjones" created="Fri, 5 Dec 2014 11:30:05 +0000"  >&lt;p&gt;Thanks Rajesh. My concern is that this patch has not yet been landed to newer versions of Lustre so if the customer were to upgrade it might mean that this issue reoccurs for them.&lt;/p&gt;

&lt;p&gt;Hongchao&lt;/p&gt;

&lt;p&gt;This fix was for SLES11 SP2 clients. Is it still necessary for SLES11 SP3 clients which is what is supported on master and b2_5? If so, please could you push this fix firstly against master to get it finalized and landed. If this issue is specific to SLES11 SP2 only then I agree that it is ok to mark the ticket as Resolved.&lt;/p&gt;

&lt;p&gt;Thanks&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="101432" author="hongchao.zhang" created="Fri, 12 Dec 2014 12:48:29 +0000"  >&lt;p&gt;Hi, this patch isn&apos;t needed for SLES11 SP3.&lt;/p&gt;</comment>
                            <comment id="101436" author="pjones" created="Fri, 12 Dec 2014 13:10:26 +0000"  >&lt;p&gt;Thanks for confirming Hongchao. Ok so then we do not need to land it to more current releases and I will close the ticket.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="14232" name="2014-02-19-client_messages_20140219(1).tgz" size="800981" author="rganesan@ddn.com" created="Thu, 6 Mar 2014 16:09:21 +0000"/>
                            <attachment id="14234" name="2014-02-19-es_lustre_showall_2014-02-19_161417(1).tar.bz2" size="310" author="rganesan@ddn.com" created="Thu, 6 Mar 2014 16:09:21 +0000"/>
                            <attachment id="14233" name="2014-02-19-es_lustre_showall_2014-02-19_161417.tar.bz2" size="304" author="rganesan@ddn.com" created="Thu, 6 Mar 2014 16:09:21 +0000"/>
                            <attachment id="14235" name="2014-02-19-server_lctl_dk_20140218.tgz" size="272" author="rganesan@ddn.com" created="Thu, 6 Mar 2014 16:09:21 +0000"/>
                            <attachment id="14287" name="check_SLES11SP3_20140210.txt" size="16818" author="rganesan@ddn.com" created="Thu, 13 Mar 2014 14:49:54 +0000"/>
                            <attachment id="14520" name="check_SLES11SP3_20140320-1.txt" size="11667" author="rganesan@ddn.com" created="Thu, 20 Mar 2014 23:24:34 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwgx3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>12978</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>