<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:50:53 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-5368] errors in/from ldlm_run_ast_work() ignored</title>
                <link>https://jira.whamcloud.com/browse/LU-5368</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;In ldlm_run_ast_work() errors from ptlrpc_set_wait() are not returned to the caller. In ldlm_process_inodebits_lock() error returns (other than -ERESTART) from ldlm_run_ast_work() are ignored.&lt;/p&gt;

&lt;p&gt;Oleg and I discussed this a bit:&lt;/p&gt;

&lt;p&gt;&lt;span class=&quot;error&quot;&gt;&amp;#91;14:20:47&amp;#93;&lt;/span&gt; John Hammond: In ldlm_run_ast_work() we ignore errors from ptlrpc. Is this intentional, unintentional, or other?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:21:32&amp;#93;&lt;/span&gt; John Hammond: Also the callers of ldlm_run_ast_work() often do not propagate its errors.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:21:46&amp;#93;&lt;/span&gt; Oleg Drokin: the idea is that we cannot do anything about it.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:21:53&amp;#93;&lt;/span&gt; Oleg Drokin: there was some patch somewhere t od oresends&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:22:06&amp;#93;&lt;/span&gt; John Hammond: Yes. Looking at that now.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:22:20&amp;#93;&lt;/span&gt; John Hammond: You mean &lt;a href=&quot;http://review.whamcloud.com/#/c/9335/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/9335/&lt;/a&gt; right?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:22:45&amp;#93;&lt;/span&gt; Oleg Drokin: yes, that would be part of this&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:24:59&amp;#93;&lt;/span&gt; John Hammond: My thought was: If there is an error in ptlrpc then currently the handler just gets stuck in ldlm_completion_ast(). Wouldn&apos;t it be better to return an error back to the client in this situation?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:48:33&amp;#93;&lt;/span&gt; Oleg Drokin: it gets stuck?&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:49:31&amp;#93;&lt;/span&gt; John Hammond: Sure. Waiting for the lock to be granted.&lt;br/&gt;
&lt;span class=&quot;error&quot;&gt;&amp;#91;14:53:07&amp;#93;&lt;/span&gt; Oleg Drokin: we probably should just call failed_ast on the spot for that particular lock to evict entire client. returning an error does not tell us much because there might be more than one lock blocking granting ofthis one and such&lt;/p&gt;</description>
                <environment></environment>
        <key id="25649">LU-5368</key>
            <summary>errors in/from ldlm_run_ast_work() ignored</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="jhammond">John Hammond</reporter>
                        <labels>
                            <label>ldlm</label>
                    </labels>
                <created>Fri, 18 Jul 2014 19:11:43 +0000</created>
                <updated>Tue, 20 Oct 2020 14:32:32 +0000</updated>
                                                                                <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="216301" author="tappro" created="Thu, 14 Dec 2017 15:32:04 +0000"  >&lt;p&gt;it is not quite clear should we do anything about that or not. In case of error lock will be timed out after all and client - evicted, in that sense there is nothing to do. What can be changed here - If ptlrpc_set_wait() fails then we can repeat sending attempt again and again until timeout will happen or send will succeed OR we can evict client on such failure immediately. Probably it is worth to implement both actions depending on error code&lt;/p&gt;</comment>
                            <comment id="242763" author="mhanafi" created="Tue, 26 Feb 2019 01:35:42 +0000"  >&lt;p&gt;In &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-11644&quot; title=&quot;LNet: Service thread inactive for 300  causes client evictions &quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-11644&quot;&gt;LU-11644&lt;/a&gt; we are getting task traces like this could this be related it?&lt;/p&gt;
&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt;
an 28 00:10:58 nbp2-oss17 kernel: [205777.962494] LNet: Skipped 3 previous similar messages
Jan 28 00:10:58 nbp2-oss17 kernel: [205777.977865] Pid: 19032, comm: ll_ost00_146 3.10.0-693.21.1.el7.20180508.x86_64.lustre2105 #1 SMP Mon Aug 27 23:04:41 UTC 2018
Jan 28 00:10:58 nbp2-oss17 kernel: [205777.977867] Call Trace:
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.977876]  [&amp;lt;ffffffffa0bcd0b0&amp;gt;] ptlrpc_set_wait+0x4c0/0x920 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982409]  [&amp;lt;ffffffffa0b8ae53&amp;gt;] ldlm_run_ast_work+0xd3/0x3a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982438]  [&amp;lt;ffffffffa0baba8b&amp;gt;] ldlm_glimpse_locks+0x3b/0x100 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982448]  [&amp;lt;ffffffffa10f98a4&amp;gt;] ofd_intent_policy+0x444/0xa40 [ofd]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982474]  [&amp;lt;ffffffffa0b8a2cd&amp;gt;] ldlm_lock_enqueue+0x38d/0x980 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982505]  [&amp;lt;ffffffffa0bb3b23&amp;gt;] ldlm_handle_enqueue0+0x9d3/0x16a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982547]  [&amp;lt;ffffffffa0c39232&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982583]  [&amp;lt;ffffffffa0c3ce9a&amp;gt;] tgt_request_handle+0x92a/0x1370 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982615]  [&amp;lt;ffffffffa0be548b&amp;gt;] ptlrpc_server_handle_request+0x23b/0xaa0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982647]  [&amp;lt;ffffffffa0be9472&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982651]  [&amp;lt;ffffffff810b1131&amp;gt;] kthread+0xd1/0xe0
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982653]  [&amp;lt;ffffffff816a14f7&amp;gt;] ret_from_fork+0x77/0xb0
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982670]  [&amp;lt;ffffffffffffffff&amp;gt;] 0xffffffffffffffff
Jan 28 00:11:00 nbp2-oss17 kernel: [205777.982672] LustreError: dumping log to /tmp/lustre-log.1548663058.19032
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.215909] Pid: 19019, comm: ll_ost00_138 3.10.0-693.21.1.el7.20180508.x86_64.lustre2105 #1 SMP Mon Aug 27 23:04:41 UTC 2018
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.215909] Call Trace:
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.215974]  [&amp;lt;ffffffffa0bcd0b0&amp;gt;] ptlrpc_set_wait+0x4c0/0x920 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216000]  [&amp;lt;ffffffffa0b8ae53&amp;gt;] ldlm_run_ast_work+0xd3/0x3a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216028]  [&amp;lt;ffffffffa0baba8b&amp;gt;] ldlm_glimpse_locks+0x3b/0x100 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216039]  [&amp;lt;ffffffffa10f98a4&amp;gt;] ofd_intent_policy+0x444/0xa40 [ofd]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216064]  [&amp;lt;ffffffffa0b8a2cd&amp;gt;] ldlm_lock_enqueue+0x38d/0x980 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216093]  [&amp;lt;ffffffffa0bb3b23&amp;gt;] ldlm_handle_enqueue0+0x9d3/0x16a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216135]  [&amp;lt;ffffffffa0c39232&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216171]  [&amp;lt;ffffffffa0c3ce9a&amp;gt;] tgt_request_handle+0x92a/0x1370 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216204]  [&amp;lt;ffffffffa0be548b&amp;gt;] ptlrpc_server_handle_request+0x23b/0xaa0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216236]  [&amp;lt;ffffffffa0be9472&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216239]  [&amp;lt;ffffffff810b1131&amp;gt;] kthread+0xd1/0xe0
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216242]  [&amp;lt;ffffffff816a14f7&amp;gt;] ret_from_fork+0x77/0xb0
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216259]  [&amp;lt;ffffffffffffffff&amp;gt;] 0xffffffffffffffff
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216262] Pid: 18757, comm: ll_ost00_035 3.10.0-693.21.1.el7.20180508.x86_64.lustre2105 #1 SMP Mon Aug 27 23:04:41 UTC 2018
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216262] Call Trace:
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216297]  [&amp;lt;ffffffffa0bcd0b0&amp;gt;] ptlrpc_set_wait+0x4c0/0x920 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216322]  [&amp;lt;ffffffffa0b8ae53&amp;gt;] ldlm_run_ast_work+0xd3/0x3a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216350]  [&amp;lt;ffffffffa0baba8b&amp;gt;] ldlm_glimpse_locks+0x3b/0x100 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216356]  [&amp;lt;ffffffffa10f98a4&amp;gt;] ofd_intent_policy+0x444/0xa40 [ofd]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216382]  [&amp;lt;ffffffffa0b8a2cd&amp;gt;] ldlm_lock_enqueue+0x38d/0x980 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216410]  [&amp;lt;ffffffffa0bb3b23&amp;gt;] ldlm_handle_enqueue0+0x9d3/0x16a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216445]  [&amp;lt;ffffffffa0c39232&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216480]  [&amp;lt;ffffffffa0c3ce9a&amp;gt;] tgt_request_handle+0x92a/0x1370 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216513]  [&amp;lt;ffffffffa0be548b&amp;gt;] ptlrpc_server_handle_request+0x23b/0xaa0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216539]  [&amp;lt;ffffffffa0be9472&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216541]  [&amp;lt;ffffffff810b1131&amp;gt;] kthread+0xd1/0xe0
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216542]  [&amp;lt;ffffffff816a14f7&amp;gt;] ret_from_fork+0x77/0xb0
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216546]  [&amp;lt;ffffffffffffffff&amp;gt;] 0xffffffffffffffff
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216548] Pid: 19228, comm: ll_ost00_238 3.10.0-693.21.1.el7.20180508.x86_64.lustre2105 #1 SMP Mon Aug 27 23:04:41 UTC 2018
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216549] Call Trace:
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216578]  [&amp;lt;ffffffffa0bcd0b0&amp;gt;] ptlrpc_set_wait+0x4c0/0x920 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216598]  [&amp;lt;ffffffffa0b8ae53&amp;gt;] ldlm_run_ast_work+0xd3/0x3a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216620]  [&amp;lt;ffffffffa0baba8b&amp;gt;] ldlm_glimpse_locks+0x3b/0x100 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216625]  [&amp;lt;ffffffffa10f98a4&amp;gt;] ofd_intent_policy+0x444/0xa40 [ofd]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216644]  [&amp;lt;ffffffffa0b8a2cd&amp;gt;] ldlm_lock_enqueue+0x38d/0x980 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216667]  [&amp;lt;ffffffffa0bb3b23&amp;gt;] ldlm_handle_enqueue0+0x9d3/0x16a0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216697]  [&amp;lt;ffffffffa0c39232&amp;gt;] tgt_enqueue+0x62/0x210 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216727]  [&amp;lt;ffffffffa0c3ce9a&amp;gt;] tgt_request_handle+0x92a/0x1370 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216752]  [&amp;lt;ffffffffa0be548b&amp;gt;] ptlrpc_server_handle_request+0x23b/0xaa0 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216777]  [&amp;lt;ffffffffa0be9472&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216779]  [&amp;lt;ffffffff810b1131&amp;gt;] kthread+0xd1/0xe0
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216780]  [&amp;lt;ffffffff816a14f7&amp;gt;] ret_from_fork+0x77/0xb0
Jan 28 00:11:00 nbp2-oss17 kernel: [205778.216784]  [&amp;lt;ffffffffffffffff&amp;gt;] 0xffffffffffffffff

&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="60919">LU-13984</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwrrz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>14972</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>