<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:22:45 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2145] The llog_origin_handle_create operation failed with -2</title>
                <link>https://jira.whamcloud.com/browse/LU-2145</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;When running IOR&apos;s writing to our Orion test system, I am seeing many Lustre error messages of the form:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;Jan  4 16:24:52 zwicky142 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
Jan  4 16:24:52 zwicky146 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
Jan  4 16:24:53 zwicky150 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
Jan  4 16:24:53 zwicky125 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
Jan  4 16:24:53 zwicky123 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
Jan  4 16:24:53 zwicky151 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
Jan  4 16:24:53 zwicky115 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
Jan  4 16:24:54 zwicky147 kernel: LustreError: 11-0: an error occurred &lt;span class=&quot;code-keyword&quot;&gt;while&lt;/span&gt; communicating with 10.1.1.211@o2ib9. The llog_origin_handle_create operation failed with -2
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;These are all messages from the clients, with 10.1.1.211 being the address of our MDS.&lt;/p&gt;

&lt;p&gt;The clients are running:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;$ rpm -qa | grep lustre-modules
lustre-modules-2.1.0-15chaos_2.6.32_220.1chaos.ch5.x86_64.x86_64
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;And the servers:&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;# zwicky1 /root &amp;gt; rpm -qa | grep lustre-orion-modules
lustre-orion-modules-2.2.49.50-2chaos_2.6.32_220.2.1.1chaos.ch5.x86_64.x86_64
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

</description>
                <environment></environment>
        <key id="12800">LU-2145</key>
            <summary>The llog_origin_handle_create operation failed with -2</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="tappro">Mikhail Pershin</assignee>
                                    <reporter username="prakash">Prakash Surya</reporter>
                        <labels>
                            <label>sequoia</label>
                    </labels>
                <created>Wed, 4 Jan 2012 19:33:01 +0000</created>
                <updated>Mon, 16 Sep 2013 10:17:01 +0000</updated>
                            <resolved>Mon, 16 Sep 2013 10:17:01 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                    <version>Lustre 2.5.0</version>
                                    <fixVersion>Lustre 2.5.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="25834" author="morrone" created="Wed, 4 Jan 2012 19:58:22 +0000"  >&lt;p&gt;Blocker for ORI on Sequoia.&lt;/p&gt;</comment>
                            <comment id="25843" author="pjones" created="Wed, 4 Jan 2012 22:55:21 +0000"  >&lt;p&gt;Should this ticket have been opened under the ORI project rather than the LU project?&lt;/p&gt;</comment>
                            <comment id="25891" author="prakash" created="Thu, 5 Jan 2012 12:07:20 +0000"  >&lt;p&gt;Yeah, sorry, it probably should have been an ORI ticket. It was too late to fix by the time I realized it was an LU.&lt;/p&gt;</comment>
                            <comment id="25895" author="pjones" created="Thu, 5 Jan 2012 12:38:29 +0000"  >&lt;p&gt;No problem. I have moved it to the Orion project.&lt;/p&gt;</comment>
                            <comment id="25930" author="tappro" created="Thu, 5 Jan 2012 16:08:21 +0000"  >&lt;p&gt;are these messages seen upon client mount or during IOR itself? This can be harmless messages, related to sptlrpc llog. Is there some log (syslog or console) from any client? &lt;/p&gt;

&lt;p&gt;This bug is set as blocker, are these messages affecting IOR execution somehow?&lt;/p&gt;</comment>
                            <comment id="25931" author="prakash" created="Thu, 5 Jan 2012 16:44:35 +0000"  >&lt;p&gt;I haven&apos;t checked during mount time, but they are definitely seen during IOR execution itself.&lt;/p&gt;

&lt;p&gt;We&apos;ll need to get confirmation from Chris, but I think it can be demoted from a blocker since they don&apos;t appear to be reaching the application and causing failures.&lt;/p&gt;

&lt;p&gt;As far as I can tell, they are relatively harmless, although it does muddy the syslog messages quite a bit. If these messages can be safely ignored, they shouldn&apos;t make it to syslog at all IMO.&lt;/p&gt;

&lt;p&gt;I can upload the client syslog info if it&apos;s helpful..?&lt;/p&gt;</comment>
                            <comment id="25988" author="tappro" created="Fri, 6 Jan 2012 02:45:56 +0000"  >&lt;p&gt;Yes, please, upload some logs, I agree also that these messages are too noisy and can be avoided if harmless&lt;/p&gt;</comment>
                            <comment id="26012" author="prakash" created="Fri, 6 Jan 2012 11:51:26 +0000"  >&lt;p&gt;Here&apos;s one of our Lustre syslog files.&lt;/p&gt;

&lt;p&gt;zwicky&lt;span class=&quot;error&quot;&gt;&amp;#91;101-164&amp;#93;&lt;/span&gt; are compute nodes.&lt;br/&gt;
zwicky&lt;span class=&quot;error&quot;&gt;&amp;#91;1-2&amp;#93;&lt;/span&gt;,zwicky-mds1 are the Orion servers.&lt;/p&gt;

&lt;p&gt;Let me know if any other information or logs would be helpful.&lt;/p&gt;</comment>
                            <comment id="38168" author="behlendorf" created="Fri, 4 May 2012 12:33:11 +0000"  >&lt;p&gt;With the various console cleanup patches applies things are a little clearer. The error is for the MGS in fact not the MDT, since it&apos;s my understanding there&apos;s no need for the llog here ENOENT isn&apos;t unexpected. What&apos;s unclear to me is why we&apos;re making this request to the MGS in the first place. Below an except from the MGS lustre log. &lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt; LustreError: 11-0: MGC10.1.1.211@o2ib9: Communicating with 10.1.1.211@o2ib9, operation llog_origin_handle_create failed with -2. &lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt; If someone could propose a patch to resolve this that would be great since we&apos;re getting a ton of noise if the console because of it.&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
00000100:00100000:7.0:1336148328.439828:0:5585:0:(service.c:1706:ptlrpc_server_handle_request()) Handling RPC pname:cluuid+ref:pid:xid:nid:opc ll_mgs_00:8b011063-f502-34dc-d164-c95bb23ad469+7:6944:x1400718304314232:12345-192.
20000000:00000001:7.0:1336148328.439830:0:5585:0:(mgs_handler.c:602:mgs_handle()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; entered
20000000:01000000:7.0:1336148328.439831:0:5585:0:(mgs_handler.c:694:mgs_handle()) @@@ llog_init  req@ffff880302ff9850 x1400718304314232/t0(0) o501-&amp;gt;8b011063-f502-34dc-d164-c95bb23ad469@192.168.65.108@o2ib6:0/0 lens 264/0 e 0
00000040:00000001:7.0:1336148328.439835:0:5585:0:(llog_server.c:78:llog_origin_handle_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; entered
00000040:00000040:7.0:1336148328.439837:0:5585:0:(llog_server.c:91:llog_origin_handle_open()) MGS: opening log lcz-sptlrpc
00000040:00000040:7.0:1336148328.439838:0:5585:0:(lustre_log.h:529:llog_ctxt_get()) GETting ctxt ffff8803143d8780 : &lt;span class=&quot;code-keyword&quot;&gt;new&lt;/span&gt; refcount 2
00000040:00000001:7.0:1336148328.439839:0:5585:0:(llog.c:691:llog_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; entered
00000040:00000001:7.0:1336148328.439840:0:5585:0:(llog.c:70:llog_alloc_handle()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; entered
00000040:00000010:7.0:1336148328.439841:0:5585:0:(llog.c:72:llog_alloc_handle()) kmalloced &lt;span class=&quot;code-quote&quot;&gt;&apos;loghandle&apos;&lt;/span&gt;: 184 at ffff8805f800e680.
00000040:00000001:7.0:1336148328.439841:0:5585:0:(llog.c:80:llog_alloc_handle()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; leaving (rc=18446612157949863552 : -131915759688064 : ffff8805f800e680)
00000040:00000001:7.0:1336148328.439843:0:5585:0:(llog_osd.c:925:llog_osd_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; entered
00000004:00000001:7.0:1336148328.439845:0:5585:0:(osd_handler.c:1231:osd_object_read_lock()) obj=ffff880303190bf8 owner=(&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;) r_locks=0 w_locks=0
00000004:00000001:7.0:1336148328.439846:0:5585:0:(osd_handler.c:2349:osd_index_lookup()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; entered
00000004:00000001:7.0:1336148328.439859:0:5585:0:(osd_handler.c:2358:osd_index_lookup()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; leaving (rc=18446744073709551614 : -2 : fffffffffffffffe)
00000004:00000001:7.0:1336148328.439860:0:5585:0:(osd_handler.c:1262:osd_object_read_unlock()) obj=ffff880303190bf8 owner=(&lt;span class=&quot;code-keyword&quot;&gt;null&lt;/span&gt;) r_locks=1 w_locks=0
00000040:00000001:7.0:1336148328.439861:0:5585:0:(llog_osd.c:951:llog_osd_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; leaving via out (rc=18446744073709551614 : -2 : 0xfffffffffffffffe)
00000040:00000001:7.0:1336148328.439862:0:5585:0:(llog_osd.c:988:llog_osd_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; leaving (rc=18446744073709551614 : -2 : fffffffffffffffe)
00000040:00000010:7.0:1336148328.439863:0:5585:0:(llog.c:100:llog_free_handle()) kfreed &lt;span class=&quot;code-quote&quot;&gt;&apos;loghandle&apos;&lt;/span&gt;: 184 at ffff8805f800e680.
00000040:00000001:7.0:1336148328.439864:0:5585:0:(llog.c:717:llog_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; leaving (rc=18446744073709551614 : -2 : fffffffffffffffe)
00000040:00000001:7.0:1336148328.439865:0:5585:0:(llog_server.c:104:llog_origin_handle_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; leaving via out_pop (rc=18446744073709551614 : -2 : 0xfffffffffffffffe)
00000040:00000040:7.0:1336148328.439866:0:5585:0:(lustre_log.h:539:llog_ctxt_put()) PUTting ctxt ffff8803143d8780 : &lt;span class=&quot;code-keyword&quot;&gt;new&lt;/span&gt; refcount 1
00000040:00000001:7.0:1336148328.439867:0:5585:0:(llog_server.c:117:llog_origin_handle_open()) &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; leaving (rc=18446744073709551614 : -2 : fffffffffffffffe)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="38170" author="adilger" created="Fri, 4 May 2012 12:42:20 +0000"  >&lt;p&gt;The problem here is that this RPC handler function is returning an error code of -2 (i.e. as if it had some error handling the RPC itself), instead of the function returning 0 and setting a rq_status=2 (i.e. the RPC was handled correctly, but it produced an error).&lt;/p&gt;</comment>
                            <comment id="38177" author="tappro" created="Fri, 4 May 2012 13:33:17 +0000"  >&lt;p&gt;This should be done in the same manner as MDT handler, using err_serious() primitive to distinguish serious errors returning error code and processing errors to be returned in rq_status. I expect that shouldn&apos;t be hard.&lt;/p&gt;</comment>
                            <comment id="38782" author="behlendorf" created="Mon, 14 May 2012 19:07:37 +0000"  >&lt;p&gt;We&apos;re also seeing bursts of the following warning presumably from similar reasons to that described above.     Because these errors are quite a bit of noise to the logs there a good chance that this week I&apos;ll put a patch together to address them as Mike suggests using err_serious().  It looks as is I can just use the MDT as an example and add the needed infrastructure to the MGS.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;May 14 13:52:43 grove608 kernel: LustreError: 42024:0:(mgc_request.c:250:do_config_log_add()) failed processing sptlrpc log: -2
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="38949" author="prakash" created="Wed, 16 May 2012 15:57:35 +0000"  >&lt;p&gt;I&apos;m looking at making a patch for this issue, but I&apos;m a little confused as to when err_serious() should be used and when it shouldn&apos;t. Can anybody provide me with some background info as to when and why err_serious should be used? I&apos;m looking at the MDT handler code, but can&apos;t spot a pattern to it&apos;s usage there.&lt;/p&gt;</comment>
                            <comment id="38955" author="adilger" created="Wed, 16 May 2012 18:05:30 +0000"  >&lt;p&gt;In the &quot;good old days&quot; (before err_serious existed) the rule was that if there was an error processing the RPC itself (e.g. corruption in the request, badly-formed RPC message buffers, bad opcode, etc) then the request handler would return a non-zero error (that eventually is stored in rq_status/pb_status) and rq_type = PTL_RPC_MSG_ERR is returned (via ptlrpc_send_error()) and this would result in a (hopefully rare) console message.&lt;/p&gt;

&lt;p&gt;If there was an error while executing the request (e.g. permission error on a file, ENOMEM/ENOSPC/... in the filesystem, etc) then the operation error is stored in rq_status and returns 0 to the RPC handler code, which set the rq_type is PTL_RPC_MSG_REPLY (via ptlrpc_send_reply()), with nothing printed on the console.  In some cases, code got this wrong and incorrectly caused console messages to be printed (which should be considered a bug to be fixed).&lt;/p&gt;

&lt;p&gt;I don&apos;t know the details of how error_serious() factors into this, but maybe this gives you some background.  Hopefully Alex or Mike can chime in on the details.&lt;/p&gt;</comment>
                            <comment id="38977" author="tappro" created="Thu, 17 May 2012 06:53:43 +0000"  >&lt;p&gt;Right, err_serious() has the same means, see mdt_req_handle(), it is common handler which calls proper operation handler and gets &apos;rc&apos; back and there is no way to know the type of &apos;rc&apos;, check comments there:&lt;/p&gt;

&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;        &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (likely(rc == 0)) {
                /*
                 * &lt;span class=&quot;code-object&quot;&gt;Process&lt;/span&gt; request, there can be two types of rc:
                 * 1) errors with msg unpack/pack, other failures outside the
                 * operation itself. This is counted as serious errors;
                 * 2) errors during fs operation, should be placed in rq_status
                 * only
                 */
                rc = h-&amp;gt;mh_act(info);
                &lt;span class=&quot;code-keyword&quot;&gt;if&lt;/span&gt; (rc == 0 &amp;amp;&amp;amp;
                    !req-&amp;gt;rq_no_reply &amp;amp;&amp;amp; req-&amp;gt;rq_reply_state == NULL) {
                        DEBUG_REQ(D_ERROR, req, &lt;span class=&quot;code-quote&quot;&gt;&quot;MDT \&quot;&lt;/span&gt;handler\&lt;span class=&quot;code-quote&quot;&gt;&quot; %s did not &quot;&lt;/span&gt;
                                  &lt;span class=&quot;code-quote&quot;&gt;&quot;pack reply and returned 0 error\n&quot;&lt;/span&gt;,
                                  h-&amp;gt;mh_name);
                        LBUG();
                }
                serious = is_serious(rc);
                rc = clear_serious(rc);
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Such technique can be used for mgs_handler as well, if it calls function which does not pure processing but also pack/unpack work and other sanity checks of RPC consistency.&lt;/p&gt;</comment>
                            <comment id="38995" author="prakash" created="Thu, 17 May 2012 11:51:02 +0000"  >&lt;p&gt;Thanks! That should help me get started.&lt;/p&gt;</comment>
                            <comment id="40161" author="behlendorf" created="Wed, 6 Jun 2012 19:10:44 +0000"  >&lt;p&gt;Mike the following patch fully implements the MDT style ptlrpc handlers for the MGS.  In the process of implementing that support the console warning which prompted me to file this issue are fixed.&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/3055&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/3055&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="46372" author="tappro" created="Thu, 11 Oct 2012 07:53:27 +0000"  >&lt;p&gt;Can I move this ticket to the Lustre project or someone needs it with historical &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2145&quot; title=&quot;The llog_origin_handle_create operation failed with -2&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2145&quot;&gt;&lt;del&gt;ORI-455&lt;/del&gt;&lt;/a&gt; ID?&lt;/p&gt;</comment>
                            <comment id="46373" author="pjones" created="Thu, 11 Oct 2012 08:49:48 +0000"  >&lt;p&gt;It&apos;s ok Mike - anyone trying to use the original ORI number will find it resolves to this ticket in its new location &lt;img class=&quot;emoticon&quot; src=&quot;https://jira.whamcloud.com/images/icons/emoticons/smile.png&quot; height=&quot;16&quot; width=&quot;16&quot; align=&quot;absmiddle&quot; alt=&quot;&quot; border=&quot;0&quot;/&gt;&lt;/p&gt;</comment>
                            <comment id="46909" author="ian" created="Thu, 25 Oct 2012 11:50:51 +0000"  >&lt;p&gt;Patches &lt;a href=&quot;http://review.whamcloud.com/#change,4258&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,4258&lt;/a&gt; and &lt;a href=&quot;http://review.whamcloud.com/#change,4273&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,4273&lt;/a&gt; both landed. Can it be closed now?&lt;/p&gt;</comment>
                            <comment id="46928" author="morrone" created="Thu, 25 Oct 2012 17:04:40 +0000"  >&lt;p&gt;I&apos;ve pulled the first one into our tree and will make sure the message is really gone.&lt;/p&gt;</comment>
                            <comment id="47079" author="tappro" created="Tue, 30 Oct 2012 08:09:59 +0000"  >&lt;p&gt;sorry, I&apos;ve missed last discussion here. No, it is far from being closed, these two patches are just preparation for more patches, they will be pushed in couple days.&lt;/p&gt;</comment>
                            <comment id="47707" author="tappro" created="Mon, 12 Nov 2012 18:49:40 +0000"  >&lt;p&gt;It takes more time than expected. I am trying to do that in correct way, first move ptlrpc services start/stop to the target and then handling request processing in common way. Estimated date for first part completion is next week.&lt;/p&gt;</comment>
                            <comment id="47950" author="adilger" created="Fri, 16 Nov 2012 14:35:07 +0000"  >&lt;p&gt;Mike, is the original problem in this bug fixed, and you are currently working on unified target code cleanup?  If that is the case, I&apos;d like to reduce the severity of this bug from being a 2.4.0 blocker.&lt;/p&gt;</comment>
                            <comment id="48452" author="morrone" created="Tue, 27 Nov 2012 19:59:55 +0000"  >&lt;p&gt;How is it going, Mike?&lt;/p&gt;</comment>
                            <comment id="48538" author="tappro" created="Thu, 29 Nov 2012 13:24:01 +0000"  >&lt;p&gt;Chris, still in progress, adapting Brian patch to the master, trying to keep it small and don&apos;t touch MDS side as it is critical for DNE landing. I expect to push patch to the gerrit in few days.&lt;/p&gt;</comment>
                            <comment id="48746" author="tappro" created="Tue, 4 Dec 2012 14:54:52 +0000"  >&lt;p&gt;move mgs ptlrpc services into unified server &lt;a href=&quot;http://review.whamcloud.com/#change,4715&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#change,4715&lt;/a&gt;&lt;br/&gt;
It doesn&apos;t solve original problem, just prepare things, next patch will do all MGS request handlers in unified way and fix this issue.&lt;/p&gt;</comment>
                            <comment id="49200" author="tappro" created="Thu, 13 Dec 2012 14:33:05 +0000"  >
&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/4826&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4826&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;combined patch to fix initial problem. All service handling is kept in MGS for now, but unified request handler is used to handle MGS request.&lt;/p&gt;</comment>
                            <comment id="53615" author="jlevi" created="Fri, 8 Mar 2013 14:22:25 +0000"  >&lt;p&gt;Please confirm that there is no defect outstanding and this is just the Unified Target patch being landed.&lt;/p&gt;</comment>
                            <comment id="53662" author="tappro" created="Sun, 10 Mar 2013 23:00:10 +0000"  >&lt;p&gt;this patch solves initial problem with spamming error messages. And it is related to UT at the same time.&lt;/p&gt;</comment>
                            <comment id="55273" author="simmonsja" created="Tue, 2 Apr 2013 12:12:26 +0000"  >&lt;p&gt;This patch; &lt;a href=&quot;http://review.whamcloud.com/4826&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4826&lt;/a&gt;; seems to have gone into limbo.&lt;/p&gt;</comment>
                            <comment id="55280" author="pjones" created="Tue, 2 Apr 2013 13:27:37 +0000"  >&lt;p&gt;James&lt;/p&gt;

&lt;p&gt;Indeed. The senior engineers involved are entirely focused on the remaining issues blocking 2.4. This work is expected to be brought to completion for 2.5&lt;/p&gt;

&lt;p&gt;Peter&lt;/p&gt;</comment>
                            <comment id="57396" author="morrone" created="Tue, 30 Apr 2013 23:04:06 +0000"  >&lt;p&gt;&lt;em&gt;Something&lt;/em&gt; needs to happen before 2.4.0 to stop the console error message spam.&lt;/p&gt;</comment>
                            <comment id="57452" author="adilger" created="Wed, 1 May 2013 17:35:49 +0000"  >&lt;p&gt;Chris, what are the current spamming messages.  The llog code has undergone some significant restructuring already, and I can&apos;t find either llog_origin_handle_create() or &quot;operation failed&quot; in the current code.&lt;/p&gt;

&lt;p&gt;I&apos;m happy to land a patch to quiet the spammy messages, but the unified target patch is 1500 lines, and at this point we are down to the last few blockers, and trying to limit the size/complexity of landed patches.&lt;/p&gt;</comment>
                            <comment id="57464" author="morrone" created="Wed, 1 May 2013 18:04:03 +0000"  >&lt;p&gt;Andreas, I think the error comes from ptlrpc_check_status().  On master the message has changed a bit so a that a simple grep for &quot;operation failed&quot; will miss it, and &quot;llog_origin_handle_create&quot; is a %s that is filled in by ll_opcode2str(opc).&lt;/p&gt;

&lt;p&gt;But as far as I know the situation still remains that will trigger ptlrpc_check_status() printing an error message.  According to the most recent comment from Mike in this ticket, the original problem for which this ticket was opened is only fixed in the larger UT patch that has not yet landed.&lt;/p&gt;

&lt;p&gt;I too am fine with a patch that just quiets the console message, as long as nothing bad is really happening behind the scenes that will bite us down the road.  Which is why I posted that &lt;em&gt;something&lt;/em&gt; needs to happen.  I wasn&apos;t particularly advocating for UT this close to the release date.&lt;/p&gt;</comment>
                            <comment id="66710" author="tappro" created="Mon, 16 Sep 2013 10:17:01 +0000"  >&lt;p&gt;patch was landed&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="10183">LU-2611</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="10720" name="lustre.log-20120105.gz" size="228525" author="prakash" created="Fri, 6 Jan 2012 11:51:26 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                            <customfield id="customfield_10070" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Project</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10031"><![CDATA[Orion]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzuw1b:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>2852</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>