<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:57:19 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-6112] osc_ldlm_completion_ast() wrong error path handler</title>
                <link>https://jira.whamcloud.com/browse/LU-6112</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;A system about to enter acceptance is experiencing unexpected client evictions following the power down of an unrelated client. During cleanup from the eviction, the client reports:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&amp;gt; 2014-11-02T01:36:06.015880-05:00 c1-0c0s2n3 LustreError: 11-0: snx11023-OST0006-osc-ffff880850703800: Communicating with 10.149.4.6@o2ib, operation obd_ping failed with -107.
&amp;gt; 2014-11-02T01:36:06.015921-05:00 c1-0c0s2n3 Lustre: snx11023-OST0006-osc-ffff880850703800: Connection to snx11023-OST0006 (at 10.149.4.6@o2ib) was lost; in progress operations using this service will wait for recovery to complete
&amp;gt; 2014-11-02T01:36:06.103816-05:00 c1-0c0s2n3 LustreError: 167-0: snx11023-OST0006-osc-ffff880850703800: This client was evicted by snx11023-OST0006; in progress operations using this service will fail.
&amp;gt; 2014-11-02T01:36:06.133063-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:806:ldlm_resource_complain()) snx11023-OST0006-osc-ffff880850703800: namespace resource [0xb428979:0x0:0x0].0 (ffff8805c02f0a80) refcount nonzero (1) after lock cleanup; forcing cleanup.
&amp;gt; 2014-11-02T01:36:06.160248-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:1445:ldlm_resource_dump()) --- Resource: [0xb428979:0x0:0x0].0 (ffff8805c02f0a80) refcount = 2
&amp;gt; 2014-11-02T01:36:06.160266-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:1448:ldlm_resource_dump()) Granted locks (in reverse order):
&amp;gt; 2014-11-02T01:36:06.210835-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:1451:ldlm_resource_dump()) ### ### ns: snx11023-OST0006-osc-ffff880850703800 lock: ffff8806e69fd2c0/0x3e4e9540570d857e lrc: 2/0,1 mode: PW/PW res: [0xb428979:0x0:0x0].0 rrc: 2 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;18446744073709551615) flags: 0x506400020000 nid: local remote: 0xc6a482c115c30a13 expref: -99 pid: 23818 timeout: 0 lvb_type: 1
&amp;gt; 2014-11-02T01:36:06.250360-05:00 c1-0c0s2n3 Lustre: snx11023-OST0006-osc-ffff880850703800: Connection restored to snx11023-OST0006 (at 10.149.4.6@o2ib)
}&amp;gt; 2014-11-02T01:36:06.015880-05:00 c1-0c0s2n3 LustreError: 11-0: snx11023-OST0006-osc-ffff880850703800: Communicating with 10.149.4.6@o2ib, operation obd_ping failed with -107.
&amp;gt; 2014-11-02T01:36:06.015921-05:00 c1-0c0s2n3 Lustre: snx11023-OST0006-osc-ffff880850703800: Connection to snx11023-OST0006 (at 10.149.4.6@o2ib) was lost; in progress operations using this service will wait for recovery to complete
&amp;gt; 2014-11-02T01:36:06.103816-05:00 c1-0c0s2n3 LustreError: 167-0: snx11023-OST0006-osc-ffff880850703800: This client was evicted by snx11023-OST0006; in progress operations using this service will fail.
&amp;gt; 2014-11-02T01:36:06.133063-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:806:ldlm_resource_complain()) snx11023-OST0006-osc-ffff880850703800: namespace resource [0xb428979:0x0:0x0].0 (ffff8805c02f0a80) refcount nonzero (1) after lock cleanup; forcing cleanup.
&amp;gt; 2014-11-02T01:36:06.160248-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:1445:ldlm_resource_dump()) --- Resource: [0xb428979:0x0:0x0].0 (ffff8805c02f0a80) refcount = 2
&amp;gt; 2014-11-02T01:36:06.160266-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:1448:ldlm_resource_dump()) Granted locks (in reverse order):
&amp;gt; 2014-11-02T01:36:06.210835-05:00 c1-0c0s2n3 LustreError: 23831:0:(ldlm_resource.c:1451:ldlm_resource_dump()) ### ### ns: snx11023-OST0006-osc-ffff880850703800 lock: ffff8806e69fd2c0/0x3e4e9540570d857e lrc: 2/0,1 mode: PW/PW res: [0xb428979:0x0:0x0].0 rrc: 2 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;18446744073709551615) flags: 0x506400020000 nid: local remote: 0xc6a482c115c30a13 expref: -99 pid: 23818 timeout: 0 lvb_type: 1
&amp;gt; 2014-11-02T01:36:06.250360-05:00 c1-0c0s2n3 Lustre: snx11023-OST0006-osc-ffff880850703800: Connection restored to snx11023-OST0006 (at 10.149.4.6@o2ib)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The server eviction is:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;&amp;gt; Nov  2 01:35:50 snx11023n005 kernel: LustreError: 0:0:(ldlm_lockd.c:363:waiting_locks_callback()) ### lock callback timer expired after 180s: evicting client at 203@gni1  ns: filter-snx11023-OST0006_UUID lock: ffff8802f3f5c480/0xc6a482c115c30a13 lrc: 3/0,0 mode: PW/PW res: 188909945/0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;18446744073709551615) flags: 0x4000000000020 remote: 0x3e4e9540570d857e expref: 4 pid: 78080 timeout 4526536428
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;From different logs&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000020:00000001:5.0:1418783943.976679:0:8277:0:(cl_lock.c:1242:cl_enqueue_try()) Process leaving (rc=1 : 1 : 1)
00000020:00000001:5.0:1418783943.976680:0:8277:0:(cl_lock.c:962:cl_lock_state_wait()) Process entered
00000020:00010000:5.0:1418783943.976680:0:8277:0:(cl_lock.c:151:cl_lock_trace0()) state wait lock: ffff8807de9e1738@(3 ffff880833626800 1 1 0 1 1 0)(ffff880ff77259c8/0/1) at cl_lock_state_wait():968
00000020:00000001:5.0:1418783943.976682:0:8277:0:(cl_lock.c:151:cl_lock_trace0()) put mutex: ffff8807de9e1738@(3 ffff880833626800 1 1 0 1 1 0)(ffff880ff77259c8/0/1) at cl_lock_mutex_put():755
00000020:00000001:5.0:1418783943.983353:0:8277:0:(cl_lock.c:151:cl_lock_trace0()) got mutex: ffff8807de9e1738@(3 ffff880833626800 1 1 0 1 1 0)(ffff880ff77259c8/0/1) at cl_lock_mutex_tail():663
00000020:00000001:5.0:1418783943.983355:0:8277:0:(cl_lock.c:999:cl_lock_state_wait()) Process leaving (rc=18446744073709551104 : -512 : fffffffffffffe00)
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;While Lustre is waiting lock grant, process got signal. cl_lock_state_wait() failed with ERESTARTSYS, and this cause detaches osc_lock. And after that result is - one waiters, one eviction, and refcount nonzero.&lt;/p&gt;</description>
                <environment></environment>
        <key id="28153">LU-6112</key>
            <summary>osc_ldlm_completion_ast() wrong error path handler</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="2">Won&apos;t Fix</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="aboyko">Alexander Boyko</reporter>
                        <labels>
                            <label>patch</label>
                    </labels>
                <created>Mon, 12 Jan 2015 10:03:01 +0000</created>
                <updated>Wed, 17 Jun 2015 18:30:09 +0000</updated>
                            <resolved>Wed, 17 Jun 2015 18:30:08 +0000</resolved>
                                    <version>Lustre 2.5.1</version>
                    <version>Lustre 2.5.2</version>
                    <version>Lustre 2.5.3</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="103176" author="gerrit" created="Mon, 12 Jan 2015 10:09:05 +0000"  >&lt;p&gt;Alexander Boyko (alexander.boyko@seagate.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/13346&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/13346&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-6112&quot; title=&quot;osc_ldlm_completion_ast() wrong error path handler&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-6112&quot;&gt;&lt;del&gt;LU-6112&lt;/del&gt;&lt;/a&gt; ldlm: wakeup ldlm waiters when error happened&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: b2_5&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 46b01a78d5c532d4cf53ae40d36e9898f6ff3806&lt;/p&gt;</comment>
                            <comment id="103177" author="aboyko" created="Mon, 12 Jan 2015 10:11:14 +0000"  >&lt;p&gt;patch &lt;a href=&quot;http://review.whamcloud.com/13346&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/13346&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="103783" author="spitzcor" created="Fri, 16 Jan 2015 19:45:13 +0000"  >&lt;p&gt;We don&apos;t need a master patch for this bug then, right?&lt;/p&gt;</comment>
                            <comment id="108391" author="aboyko" created="Mon, 2 Mar 2015 13:05:17 +0000"  >&lt;p&gt;yes&lt;/p&gt;</comment>
                            <comment id="118845" author="amk" created="Wed, 17 Jun 2015 18:18:50 +0000"  >&lt;p&gt;This bug can be closed since the patch is not needed in master. The patch does resolve the issue in 2.5.&lt;/p&gt;</comment>
                            <comment id="118856" author="pjones" created="Wed, 17 Jun 2015 18:30:09 +0000"  >&lt;p&gt;Thanks Anne - it&apos;s good to reduced clutter where we can&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzx3sf:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>17020</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>