<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:34:06 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-3460] recovery-small test_51 timeout: lqe_iter_cb(): Inuse quota entry</title>
                <link>https://jira.whamcloud.com/browse/LU-3460</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for Andreas Dilger &amp;lt;andreas.dilger@intel.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;http://maloo.whamcloud.com/test_sets/f96a0ba0-d2f2-11e2-a349-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://maloo.whamcloud.com/test_sets/f96a0ba0-d2f2-11e2-a349-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The sub-test test_51 failed with the following error in the MDS log:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;lqe_iter_cb(): Inuse quota entry&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Info required for matching: recovery-small 51&lt;/p&gt;</description>
                <environment></environment>
        <key id="19388">LU-3460</key>
            <summary>recovery-small test_51 timeout: lqe_iter_cb(): Inuse quota entry</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="niu">Niu Yawei</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Wed, 12 Jun 2013 16:26:54 +0000</created>
                <updated>Wed, 6 Nov 2013 15:26:57 +0000</updated>
                            <resolved>Thu, 11 Jul 2013 05:45:59 +0000</resolved>
                                    <version>Lustre 2.4.1</version>
                                    <fixVersion>Lustre 2.5.0</fixVersion>
                    <fixVersion>Lustre 2.4.2</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="60768" author="jlevi" created="Mon, 17 Jun 2013 16:07:36 +0000"  >&lt;p&gt;Niu, Johann,&lt;br/&gt;
Could one of you have a look at this ticket and please comment?&lt;br/&gt;
Thank you!&lt;/p&gt;</comment>
                            <comment id="60785" author="niu" created="Tue, 18 Jun 2013 02:54:06 +0000"  >&lt;div class=&quot;code panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;codeContent panelContent&quot;&gt;
&lt;pre class=&quot;code-java&quot;&gt; /usr/lib64/lustre/tests; LUSTRE=&lt;span class=&quot;code-quote&quot;&gt;&quot;/usr/lib64/lustre&quot;&lt;/span&gt; sh -c &lt;span class=&quot;code-quote&quot;&gt;&quot;umount -d /mnt/mds1&quot;&lt;/span&gt;);echo XXRETCODE:$?&apos;
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:810:ldlm_resource_complain()) Namespace lustre-MDT0000-lwp-MDT0000 resource refcount nonzero (1) after lock cleanup; forcing cleanup.
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:810:ldlm_resource_complain()) Skipped 1 previous similar message
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:816:ldlm_resource_complain()) Resource: ffff88006a8fec00 (8589934598/65536/0/0) (rc: 1)
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:816:ldlm_resource_complain()) Skipped 1 previous similar message
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:1457:ldlm_resource_dump()) --- Resource: ffff88006a8fec00 (8589934598/65536/0/0) (rc: 2)
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:1460:ldlm_resource_dump()) Granted locks (in reverse order):
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:1463:ldlm_resource_dump()) ### ### ns: lustre-MDT0000-lwp-MDT0000 lock: ffff88006a8ff340/0x21e5b86f50e944e2 lrc: 2/1,0 mode: CR/CR res: 8589934598/65536 rrc: 2 type: PLN flags: 0x1106400000000 nid: local remote: 0x21e5b86f50e9450c expref: -99 pid: 28646 timeout: 0lvb_type: 2
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:1463:ldlm_resource_dump()) Skipped 1 previous similar message
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:1457:ldlm_resource_dump()) --- Resource: ffff88006a8feac0 (8589934598/16842752/0/0) (rc: 2)
Jun 11 06:22:36 client-27vm7 kernel: LustreError: 28977:0:(ldlm_resource.c:1460:ldlm_resource_dump()) Granted locks (in reverse order):
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Looks somebody is still holding the reference on id lock, so the lqe can&apos;t be put on cleanup (id lock holding lqe reference), but I don&apos;t see why the id locks are not freed yet.&lt;/p&gt;</comment>
                            <comment id="60971" author="niu" created="Fri, 21 Jun 2013 08:00:06 +0000"  >&lt;p&gt;This is probably related to &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2067&quot; title=&quot;ldlm_resource_complain()) Namespace MGC resource refcount nonzero after lock cleanup&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2067&quot;&gt;&lt;del&gt;LU-2067&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;It is possible that lock reader/writer isn&apos;t dropped to zero when ldlm_namespace_cleanup() is called, imagine following scenario:&lt;/p&gt;

&lt;ul class=&quot;alternate&quot; type=&quot;square&quot;&gt;
	&lt;li&gt;ldlm_cli_enqueue() is called to create the lock, and increased lock reader/writer;&lt;/li&gt;
	&lt;li&gt;before the enqueue request is added in imp_sending_list or imp_delayed_list, shutdown happened;&lt;/li&gt;
	&lt;li&gt;shutdown procedure aborted inflight RPCs, but the enqueue request can&apos;t be aborted since it&apos;s neither on sending list nor delayed list;&lt;/li&gt;
	&lt;li&gt;shutdown procedure moving on to obd_import_event(IMP_EVENT_ACTIVE)-&amp;gt;ldlm_namespace_cleanup() to cleanup all locks;&lt;/li&gt;
	&lt;li&gt;ldlm_namespace_cleanup() found that the lock just created still has 1 reader/writer, because the interpret callback for this lock enqueue hasn&apos;t been called yet (where the reader/writer is dropped;&lt;/li&gt;
&lt;/ul&gt;


&lt;p&gt;That&apos;s why we can see the warnning message from ldlm_namespace_cleanup(), though the lock will be released finally.&lt;/p&gt;

&lt;p&gt;The problem of lqe refecount is that: even if the lock will be cancelled at the end, the lqe could be possibly re-added in adjust list (see qsd_id_blocking_ast()-&amp;gt;qsd_adjust()), then the lqe refcount will never be dropped since the quota writeback thread has been stopped. I&apos;ll cook a patch to fix this problem.&lt;/p&gt;</comment>
                            <comment id="60972" author="niu" created="Fri, 21 Jun 2013 08:12:22 +0000"  >&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/6731&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/6731&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="61464" author="sarah" created="Thu, 27 Jun 2013 21:51:49 +0000"  >&lt;p&gt;another instance on zfs&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/cb382a8e-ddcc-11e2-85a3-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/cb382a8e-ddcc-11e2-85a3-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="62074" author="niu" created="Thu, 11 Jul 2013 05:45:59 +0000"  >&lt;p&gt;patch landed.&lt;/p&gt;</comment>
                            <comment id="70584" author="yujian" created="Mon, 4 Nov 2013 03:30:14 +0000"  >&lt;p&gt;Lustre build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_4/47/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_4/47/&lt;/a&gt;&lt;br/&gt;
Distro/Arch: RHEL6.4/x86_64&lt;/p&gt;

&lt;p&gt;FSTYPE=zfs&lt;br/&gt;
MDSCOUNT=1&lt;br/&gt;
OSTCOUNT=2&lt;/p&gt;

&lt;p&gt;recovery-small test 51 hung as follows:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Stopping /mnt/mds1 (opts:) on fat-intel-2vm3
CMD: fat-intel-2vm3 umount -d /mnt/mds1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Console log on MDS:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;15:12:27:Lustre: DEBUG MARKER: umount -d /mnt/mds1
15:12:27:LustreError: 26988:0:(ldlm_resource.c:804:ldlm_resource_complain()) lustre-MDT0000-lwp-MDT0000: namespace resource [0x200000006:0x1010000:0x0].0 (ffff88007017ea80) refcount nonzero (1) after lock cleanup; forcing cleanup.
15:12:27:LustreError: 26988:0:(ldlm_resource.c:1415:ldlm_resource_dump()) --- Resource: [0x200000006:0x1010000:0x0].0 (ffff88007017ea80) refcount = 2
15:12:27:LustreError: 26988:0:(ldlm_resource.c:1418:ldlm_resource_dump()) Granted locks (in reverse order):
15:12:27:LustreError: 26988:0:(ldlm_resource.c:1421:ldlm_resource_dump()) ### ### ns: lustre-MDT0000-lwp-MDT0000 lock: ffff88006ef0bb40/0xdd1be0eccfc5313 lrc: 2/1,0 mode: CR/CR res: [0x200000006:0x1010000:0x0].0 rrc: 2 type: PLN flags: 0x1106400000000 nid: local remote: 0xdd1be0eccfc5359 expref: -99 pid: 26657 timeout: 0 lvb_type: 2
15:12:27:LustreError: 26988:0:(ldlm_resource.c:804:ldlm_resource_complain()) lustre-MDT0000-lwp-MDT0000: namespace resource [0x200000006:0x10000:0x0].0 (ffff88007017ebc0) refcount nonzero (1) after lock cleanup; forcing cleanup.
15:12:27:LustreError: 26988:0:(ldlm_resource.c:1415:ldlm_resource_dump()) --- Resource: [0x200000006:0x10000:0x0].0 (ffff88007017ebc0) refcount = 2
15:12:27:LustreError: 26988:0:(ldlm_resource.c:1418:ldlm_resource_dump()) Granted locks (in reverse order):
15:12:27:LustreError: 26988:0:(ldlm_resource.c:1421:ldlm_resource_dump()) ### ### ns: lustre-MDT0000-lwp-MDT0000 lock: ffff88007017f240/0xdd1be0eccfc531a lrc: 2/1,0 mode: CR/CR res: [0x200000006:0x10000:0x0].0 rrc: 2 type: PLN flags: 0x1106400000000 nid: local remote: 0xdd1be0eccfc5360 expref: -99 pid: 26656 timeout: 0 lvb_type: 2
15:12:28:LustreError: 26988:0:(ldlm_lib.c:2137:target_stop_recovery_thread()) lustre-MDT0000: Aborting recovery
15:12:28:Lustre: 26659:0:(ldlm_lib.c:1801:target_recovery_overseer()) recovery is aborted, evict exports in recovery
15:12:28:Lustre: 26659:0:(ldlm_lib.c:1801:target_recovery_overseer()) Skipped 2 previous similar messages
15:12:28:LustreError: 26650:0:(osp_precreate.c:737:osp_precreate_cleanup_orphans()) lustre-OST0000-osc-MDT0000: cannot cleanup orphans: rc = -5
15:12:28:LustreError: 26988:0:(lquota_entry.c:147:lqe_iter_cb()) $$$ Inuse quota entry qsd:lustre-MDT0000 qtype:usr id:60000 enforced:1 granted:0 pending:0 waiting:0 req:0 usage:0 qunit:0 qtune:0 edquot:0
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Maloo report: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/2edc7780-4451-11e3-8472-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/2edc7780-4451-11e3-8472-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;replay-single test 61b also hit the same failure:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/74517e7e-4450-11e3-8472-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/74517e7e-4450-11e3-8472-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;replay-dual test 16 also hit the same failure:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/9bf98132-4451-11e3-8472-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/9bf98132-4451-11e3-8472-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The patch is needed on Lustre b2_4 branch.&lt;/p&gt;</comment>
                            <comment id="70657" author="bogl" created="Mon, 4 Nov 2013 21:11:47 +0000"  >&lt;p&gt;backport to b2_4: &lt;a href=&quot;http://review.whamcloud.com/8169&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/8169&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvt5j:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>8656</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>