<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:12:08 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7812] sanity test_120e: 1 blocking RPC occured</title>
                <link>https://jira.whamcloud.com/browse/LU-7812</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for wangshilong &amp;lt;wshilong@ddn.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/6ce8d47a-db1b-11e5-877a-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/6ce8d47a-db1b-11e5-877a-5254006e85c2&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_120e failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;1 blocking RPC occured.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Please provide additional information about the failure here.&lt;/p&gt;

&lt;p&gt;Info required for matching: sanity 120e&lt;/p&gt;</description>
                <environment></environment>
        <key id="34951">LU-7812</key>
            <summary>sanity test_120e: 1 blocking RPC occured</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="green">Oleg Drokin</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Thu, 25 Feb 2016 03:02:59 +0000</created>
                <updated>Tue, 14 Mar 2017 21:17:05 +0000</updated>
                            <resolved>Tue, 17 Jan 2017 18:04:00 +0000</resolved>
                                                    <fixVersion>Lustre 2.10.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>8</watches>
                                                                            <comments>
                            <comment id="144559" author="bogl" created="Thu, 3 Mar 2016 20:59:43 +0000"  >&lt;p&gt;another seen on master:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a6f909f2-e17c-11e5-937c-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a6f909f2-e17c-11e5-937c-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="144969" author="gerrit" created="Wed, 9 Mar 2016 03:17:16 +0000"  >&lt;p&gt;Andreas Dilger (andreas.dilger@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/18826&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/18826&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7812&quot; title=&quot;sanity test_120e: 1 blocking RPC occured&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7812&quot;&gt;&lt;del&gt;LU-7812&lt;/del&gt;&lt;/a&gt; tests: print out more stats for test failures&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 85bf5b4bcf327276be4b1764b482e739cb452746&lt;/p&gt;</comment>
                            <comment id="146572" author="gerrit" created="Wed, 23 Mar 2016 06:03:30 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/18826/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/18826/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7812&quot; title=&quot;sanity test_120e: 1 blocking RPC occured&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7812&quot;&gt;&lt;del&gt;LU-7812&lt;/del&gt;&lt;/a&gt; tests: print more stats for sanity 120e failure&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 7cc486b0a9e5000f8a4a130ce0ba7288cde585d7&lt;/p&gt;</comment>
                            <comment id="150516" author="adilger" created="Thu, 28 Apr 2016 23:39:36 +0000"  >&lt;p&gt;Looking at the debug logs of a recent failure (&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/26e8e22a-0cd1-11e6-9e5d-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/26e8e22a-0cd1-11e6-9e5d-5254006e85c2&lt;/a&gt;), it appears that an &lt;tt&gt;LDLM_BL_CALLBACK&lt;/tt&gt; RPC is arriving at the client, even though the client is supposed to have dropped all of its locks:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00000100:00100000:0.0:1461790257.944465:0:4445:0:(service.c:1922:ptlrpc_server_handle_req_in()) got req x1532794272717024
00000100:00100000:0.0:1461790257.944494:0:4445:0:(service.c:2070:ptlrpc_server_handle_request()) Handling RPC pname:cluuid+ref:pid:xid:nid:opc ldlm_cb00_001:LOV_OSC_UUID+4:1261:x1532794272717024:12345-10.9.4.226@tcp:104
00000100:00100000:0.0:1461790257.944583:0:4445:0:(service.c:2120:ptlrpc_server_handle_request()) Handled RPC pname:cluuid+ref:pid:xid:nid:opc ldlm_cb00_001:LOV_OSC_UUID+4:1261:x1532794272717024:12345-10.9.4.226@tcp:104 Request procesed in 88us (150us total) trans 0 rc 0/0
00010000:00010000:0.0:1461790257.944606:0:12644:0:(ldlm_lockd.c:1764:ldlm_handle_bl_callback()) ### client blocking AST callback handler ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 2/0,0 mode: PR/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x420000000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790257.944627:0:12644:0:(ldlm_lockd.c:1777:ldlm_handle_bl_callback()) Lock ffff880063943d80 already unused, calling callback (ffffffffa0af2c30)
00010000:00010000:0.0:1461790257.944629:0:12644:0:(ldlm_request.c:1113:ldlm_cli_cancel_local()) ### client-side cancel ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 3/0,0 mode: PR/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x428400000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790257.944687:0:12644:0:(ldlm_request.c:1175:ldlm_cancel_pack()) ### packing ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 2/0,0 mode: --/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x4c69400000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790257.944694:0:12644:0:(ldlm_lockd.c:1786:ldlm_handle_bl_callback()) ### client blocking callback handler END ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 1/0,0 mode: --/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x4c69400000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790257.944698:0:12644:0:(ldlm_lock.c:200:ldlm_lock_put()) ### final lock_put on destroyed lock, freeing it. ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 0/0,0 mode: --/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x4c69400000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;It appears that the lock is the &lt;tt&gt;LCK_PR&lt;/tt&gt; lock gotten from &lt;tt&gt;dd if=$DIR/$tdir/f1 of=/dev/null&lt;/tt&gt;, since the &lt;tt&gt;LCK_PW&lt;/tt&gt; lock on the &lt;tt&gt;[0x931:0x0:0x0]&lt;/tt&gt; resource was dropped after the first &lt;tt&gt;cancel_lru_locks osc&lt;/tt&gt; that also caused the dirty data to be written to the OST:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00010000:00010000:0.0:1461790256.415781:0:18613:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0000-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.415805:0:18613:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0001-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.415824:0:18613:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0002-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.415843:0:18613:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0003-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.415851:0:12645:0:(ldlm_request.c:1113:ldlm_cli_cancel_local()) ### client-side cancel ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943580/0xc744c7e3377cb731 lrc: 3/0,0 mode: PW/PW res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x28400000000 nid: local remote: 0x38471ca6aa1b9ff4 expref: -99 pid: 18609 timeout: 0 lvb_type: 1
00000008:00100000:0.0:1461790256.415914:0:12645:0:(osc_request.c:1315:osc_brw_prep_request()) brw rpc ffff8800597b4080 - object 0x0:2353 offset 0&amp;lt;&amp;gt;512
00000100:00100000:0.0:1461790256.415932:0:12729:0:(client.c:1589:ptlrpc_send_new_req()) Sending RPC pname:cluuid:pid:xid:nid:opc ptlrpcd_00_00:1765eedc-c6c6-d3ef-2f08-6efd98a5aa0c:12729:1532794264066032:10.9.4.226@tcp:4
00000008:00010000:0.0:1461790256.418958:0:12729:0:(osc_cache.c:857:osc_extent_finish()) ### extent: ffff88005ae1f3d8
 ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943580/0xc744c7e3377cb731 lrc: 3/0,0 mode: PW/PW res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x29400000000 nid: local remote: 0x38471ca6aa1b9ff4 expref: -99 pid: 18609 timeout: 0 lvb_type: 1
00000008:00010000:0.0:1461790256.418973:0:12729:0:(osc_cache.c:489:__osc_extent_remove()) ### extent: ffff88005ae1f3d8
 ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943580/0xc744c7e3377cb731 lrc: 3/0,0 mode: PW/PW res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x29400000000 nid: local remote: 0x38471ca6aa1b9ff4 expref: -99 pid: 18609 timeout: 0 lvb_type: 1
00000100:00100000:0.0:1461790256.418984:0:12729:0:(client.c:1997:ptlrpc_check_set()) Completed RPC pname:cluuid:pid:xid:nid:opc ptlrpcd_00_00:1765eedc-c6c6-d3ef-2f08-6efd98a5aa0c:12729:1532794264066032:10.9.4.226@tcp:4
00000100:00100000:0.0:1461790256.419023:0:12645:0:(client.c:1589:ptlrpc_send_new_req()) Sending RPC pname:cluuid:pid:xid:nid:opc ldlm_bl_26:1765eedc-c6c6-d3ef-2f08-6efd98a5aa0c:12645:1532794264066048:10.9.4.226@tcp:103
00000100:00100000:0.0:1461790256.420085:0:12645:0:(client.c:1997:ptlrpc_check_set()) Completed RPC pname:cluuid:pid:xid:nid:opc ldlm_bl_26:1765eedc-c6c6-d3ef-2f08-6efd98a5aa0c:12645:1532794264066048:10.9.4.226@tcp:103
00010000:00010000:0.0:1461790256.420092:0:12645:0:(ldlm_lock.c:200:ldlm_lock_put()) ### final lock_put on destroyed lock, freeing it. ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943580/0xc744c7e3377cb731 lrc: 0/0,0 mode: --/PW res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x4869400000000 nid: local remote: 0x38471ca6aa1b9ff4 expref: -99 pid: 18609 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790256.420142:0:18613:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0004-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.420163:0:18613:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0005-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.420181:0:18613:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0006-osc-ffff8800591dbc00
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The &lt;tt&gt;LCK_PR&lt;/tt&gt; lock is fetched during the read and added to the LRU and then later matched for the &lt;tt&gt;stat $DIR/$tdir/f1&lt;/tt&gt; to get the file size and returned to the LRU:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00010000:00010000:0.0:1461790256.432444:0:18616:0:(ldlm_lock.c:876:ldlm_lock_decref_internal()) ### add lock into lru list ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 3/0,0 mode: PR/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x10020000000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
:
:
00000001:00010000:0.0:1461790256.439644:0:18617:0:(glimpse.c:96:cl_glimpse_lock()) Glimpsing inode [0x200001b71:0xabe:0x0]
00000008:00010000:0.0:1461790256.439654:0:18617:0:(osc_lock.c:1201:osc_lock_init()) ### lock ffff880051d6e470, osc lock ffff880064b45070, flags 0x1002
00010000:00010000:0.0:1461790256.439660:0:18617:0:(ldlm_lock.c:749:ldlm_lock_addref_internal_nolock()) ### ldlm_lock_addref(PR) ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 2/1,0 mode: PR/PR res: [0x931:0x0:0x0].0x0 rrc: 2 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x20000000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790256.439666:0:18617:0:(ldlm_lock.c:1474:ldlm_lock_match()) ### matched (0 18446744073709551615) ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 2/1,0 mode: PR/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x20000000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790256.439677:0:18617:0:(ldlm_lock.c:803:ldlm_lock_decref_internal_nolock()) ### ldlm_lock_decref(PR) ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 5/1,0 mode: PR/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x10020000000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00010000:00010000:0.0:1461790256.439689:0:18617:0:(ldlm_lock.c:876:ldlm_lock_decref_internal()) ### add lock into lru list ns: lustre-OST0003-osc-ffff8800591dbc00 lock: ffff880063943d80/0xc744c7e3377cb74d lrc: 4/0,0 mode: PR/PR res: [0x931:0x0:0x0].0x0 rrc: 1 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x10020000000000 nid: local remote: 0x38471ca6aa1b9ffb expref: -99 pid: 18616 timeout: 0 lvb_type: 1
00000080:00200000:0.0:1461790256.439697:0:18617:0:(file.c:971:ll_merge_attr()) [0x200001b71:0xabe:0x0] updating i_size 512
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;However, the lock seems to have been skipped during the immediately following &lt;tt&gt;cancel_lru_locks osc&lt;/tt&gt; call:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00010000:00010000:0.0:1461790256.442518:0:18618:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0000-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.442558:0:18618:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0001-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.442612:0:18618:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0002-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.442644:0:18618:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0003-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.442675:0:18618:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0004-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.442706:0:18618:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0005-osc-ffff8800591dbc00
00010000:00010000:0.0:1461790256.442735:0:18618:0:(ldlm_resource.c:322:lprocfs_lru_size_seq_write()) dropping all unused locks from namespace lustre-OST0006-osc-ffff8800591dbc00
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;and the lock is only called back from the OST when the file unlink triggers the object destroy on the OST, as shown above.&lt;/p&gt;

&lt;p&gt;We need to find out why &lt;tt&gt;cancel_lru_locks&lt;/tt&gt; didn&apos;t actually drop this lock.  I don&apos;t think it is necessarily more efficient to have the client explicitly cancel the OST locks compared to just waiting for the OST blocking callback, since the client could be doing something else in the meantime that would allow the OST locks to be piggy-backed onto another OST RPC via ELC.  It might make sense to move the OST locks to the end of the LRU so that they are sure to be cancelled with the next RPC to this OST.&lt;/p&gt;</comment>
                            <comment id="150517" author="adilger" created="Thu, 28 Apr 2016 23:40:17 +0000"  >&lt;p&gt;&lt;a href=&quot;https://jira.whamcloud.com/browse/LU-4206&quot; title=&quot;Sanity test_120e fails with 1 blocking RPC occured.&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-4206&quot;&gt;&lt;del&gt;LU-4206&lt;/del&gt;&lt;/a&gt; only worked around the test failure, it didn&apos;t really fix it.&lt;/p&gt;</comment>
                            <comment id="161629" author="bogl" created="Thu, 11 Aug 2016 16:02:00 +0000"  >&lt;p&gt;another on master, sles11sp4 client and server:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/1d257222-5fdc-11e6-906c-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/1d257222-5fdc-11e6-906c-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="175279" author="bogl" created="Mon, 28 Nov 2016 21:19:54 +0000"  >&lt;p&gt;more on master, sles11sp4 client and server:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/fe524e90-b5af-11e6-a223-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/fe524e90-b5af-11e6-a223-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/48ea6166-b5c1-11e6-a223-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/48ea6166-b5c1-11e6-a223-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I think this may be a 100% fail on sles11sp4.&lt;br/&gt;
raising this ticket to Blocker&lt;/p&gt;</comment>
                            <comment id="175639" author="bogl" created="Wed, 30 Nov 2016 13:45:48 +0000"  >&lt;p&gt;more on master, sles12sp1&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/ad4ab7c4-b6d8-11e6-a559-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/ad4ab7c4-b6d8-11e6-a559-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/c864a7d0-b729-11e6-a559-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/c864a7d0-b729-11e6-a559-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/a0975c86-b759-11e6-a559-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/a0975c86-b759-11e6-a559-5254006e85c2&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;this may be a 100% fail on sles12 too&lt;/p&gt;</comment>
                            <comment id="178508" author="bogl" created="Tue, 20 Dec 2016 12:33:47 +0000"  >&lt;p&gt;more on master:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/59612658-c668-11e6-8cb7-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/59612658-c668-11e6-8cb7-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/7457d9e4-c897-11e6-8a5b-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/7457d9e4-c897-11e6-8a5b-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="179257" author="bogl" created="Fri, 30 Dec 2016 20:33:46 +0000"  >&lt;p&gt;more on master:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/114e1984-cece-11e6-af6a-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/114e1984-cece-11e6-af6a-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/edf384e2-d0b2-11e6-bbdd-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/edf384e2-d0b2-11e6-bbdd-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/93fba206-d163-11e6-bbdd-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/93fba206-d163-11e6-bbdd-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/9c3a97e2-d69f-11e6-b630-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/9c3a97e2-d69f-11e6-b630-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/187943b6-d6bf-11e6-bb30-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/187943b6-d6bf-11e6-bb30-5254006e85c2&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/8a1d30ea-d773-11e6-923b-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/8a1d30ea-d773-11e6-923b-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="180355" author="jay" created="Tue, 10 Jan 2017 22:19:53 +0000"  >&lt;blockquote&gt;
&lt;p&gt;We need to find out why cancel_lru_locks didn&apos;t actually drop this lock. ...&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;&lt;tt&gt;cancel_lru_locks osc&lt;/tt&gt; cancels unused locks which are the locks with zero &lt;tt&gt;l_readers&lt;/tt&gt; and &lt;tt&gt;l_writers&lt;/tt&gt;. There exists a race window that &lt;tt&gt;stat $DIR/$tdir $DIR/$tdir/f1 &amp;gt; /dev/null&lt;/tt&gt; has completed but ldlm callback thread is still held the glimpse locks when &lt;tt&gt;cancel_lru_locks osc&lt;/tt&gt; is being called, therefore the glimpse locks are not supposed to be &apos;unused&apos; and miss the cancellation.&lt;/p&gt;

&lt;p&gt;THis problem can be fixed by adding some delay between &lt;tt&gt;cancel_lru_locks osc&lt;/tt&gt; and &lt;tt&gt;stat $DIR/$tdir $DIR/$tdir/f1 &amp;gt; /dev/null&lt;/tt&gt;.&lt;/p&gt;</comment>
                            <comment id="180379" author="gerrit" created="Wed, 11 Jan 2017 02:16:29 +0000"  >&lt;p&gt;Jinshan Xiong (jinshan.xiong@intel.com) uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/24811&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/24811&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7812&quot; title=&quot;sanity test_120e: 1 blocking RPC occured&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7812&quot;&gt;&lt;del&gt;LU-7812&lt;/del&gt;&lt;/a&gt; tests: add a race condition for sanity:120e&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: f4a1c85cb6353e057ea562a1094781b1924a9050&lt;/p&gt;</comment>
                            <comment id="180443" author="bogl" created="Wed, 11 Jan 2017 14:54:26 +0000"  >&lt;p&gt;There is a very similar looking fail in test_120f.  see &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7889&quot; title=&quot;sanity test_120f: 1 blocking RPC occured&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7889&quot;&gt;&lt;del&gt;LU-7889&lt;/del&gt;&lt;/a&gt;.  Does it need a similar test fix?&lt;/p&gt;</comment>
                            <comment id="180464" author="jay" created="Wed, 11 Jan 2017 16:42:07 +0000"  >&lt;p&gt;Yes, I think so. We can merge these two tickets.&lt;/p&gt;</comment>
                            <comment id="180999" author="gerrit" created="Tue, 17 Jan 2017 17:31:37 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;https://review.whamcloud.com/24811/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/24811/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7812&quot; title=&quot;sanity test_120e: 1 blocking RPC occured&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7812&quot;&gt;&lt;del&gt;LU-7812&lt;/del&gt;&lt;/a&gt; tests: address race condition for sanity:120&lt;/p&gt;
{e,f}
&lt;p&gt;Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: e240fb5099af8e62c532d314317095800ebb6864&lt;/p&gt;</comment>
                            <comment id="181002" author="pjones" created="Tue, 17 Jan 2017 18:04:00 +0000"  >&lt;p&gt;Landed for 2.10&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="21844">LU-4206</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="35472">LU-7889</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzy2nb:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>