<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:07:43 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-7302] sanity-lfsck test_23b hangs on START_NAMESPACE</title>
                <link>https://jira.whamcloud.com/browse/LU-7302</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;
stdout.log
Total allocated inode limit: 0, total allocated block limit: 0

== sanity-lfsck test 23b: LFSCK can repair dangling name entry (2) == 01:58:28 (1443319108)
#####
The objectA has multiple hard links, one of them corresponding
to the name entry_B. But there is something wrong for the name
entry_B and cause entry_B to references non-exist object_C.
In the first-stage scanning, the LFSCK will think the entry_B
as dangling, and re-create the lost object_C. When the LFSCK
comes to the second-stage scanning, it will find that the
former re-creating object_C is not proper, and will try to
replace the object_C with the real object_A.
Inject failure stub on MDT0 to simulate dangling name entry
fail_loc=0x1621
fail_loc=0
&apos;ls&apos; should fail because of dangling name entry
Trigger namespace LFSCK to find out dangling name entry


stderr.log
pdsh@fre0203: fre0201: ssh exited with exit code 1
pdsh@fre0203: fre0202: ssh exited with exit code 1
Using TIMEOUT=20
excepting tests: 



&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;In the ticket &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7217&quot; title=&quot;sanity-lfsck test_23b hangs on START_NAMESPACE&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7217&quot;&gt;&lt;del&gt;LU-7217&lt;/del&gt;&lt;/a&gt; similar timeout is seen during test_23c due to &lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;#define OBD_FAIL_LFSCK_DELAY3           0x1602
        do_facet $SINGLEMDS $LCTL set_param fail_val=10 fail_loc=0x1602
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;in test_23b I see following, &lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;4178 00000080:00200000:1.0:1443429106.393336:0:14812:0:(dcache.c:424:ll_revalidate_nd()) VFS Op:name=d23b.sanity-lfsck, flags=0
4179 00000080:00200000:1.0:1443429106.393338:0:14812:0:(file.c:3217:__ll_inode_revalidate()) VFS Op:inode=[0x200000401:0x1:0x0](ffff88011bc48140),name=d23b.sanity-lfsck
4180 00000002:00010000:1.0:1443429106.393339:0:14812:0:(mdc_locks.c:1114:mdc_intent_lock()) (name: ,[0x200000401:0x1:0x0]) in obj [0x200000401:0x1:0x0], intent: getattr flags 00
4181 00010000:00010000:1.0:1443429106.393341:0:14812:0:(ldlm_lock.c:1493:ldlm_lock_match()) ### not matched ns ffff88011a7faa00 type 13 mode 30 res 8589935617/1 (0 0)
4182 00010000:00010000:1.0:1443429106.393350:0:14812:0:(ldlm_lock.c:746:ldlm_lock_addref_internal_nolock()) ### ldlm_lock_addref(CR) ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb     9fa48c lrc: 3/1,0 mode: --/CR res: [0x200000401:0x1:0x0].0 bits 0x0 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x0 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4183 00010000:00010000:1.0:1443429106.393352:0:14812:0:(ldlm_request.c:925:ldlm_cli_enqueue()) ### client-side enqueue START, flags 0x1000
4184  ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb9fa48c lrc: 3/1,0 mode: --/CR res: [0x200000401:0x1:0x0].0 bits 0x2 rrc: 2 type: IBT flags: 0x0 nid: local remote: 0x0 expref:      -99 pid: 14812 timeout: 0 lvb_type: 0
4185 00010000:00010000:1.0:1443429106.393355:0:14812:0:(ldlm_request.c:980:ldlm_cli_enqueue()) ### sending request ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb9fa48c lrc: 3/1,0      mode: --/CR res: [0x200000401:0x1:0x0].0 bits 0x2 rrc: 2 type: IBT flags: 0x0 nid: local remote: 0x0 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4186 00000100:00100000:1.0:1443429106.393360:0:14812:0:(client.c:1575:ptlrpc_send_new_req()) Sending RPC pname:cluuid:pid:xid:nid:opc rm:d5ce60dd-b58f-953f-4b1f-b15d8912a206:14812:1513543112333784:192.168.101     .33@tcp:101
4187 00000100:00100000:1.0:1443429106.393384:0:14812:0:(client.c:2255:ptlrpc_set_wait()) set ffff88011ae58680 going to sleep for 6 seconds
4188 00000100:00100000:1.0:1443429106.396347:0:14812:0:(client.c:1970:ptlrpc_check_set()) Completed RPC pname:cluuid:pid:xid:nid:opc rm:d5ce60dd-b58f-953f-4b1f-b15d8912a206:14812:1513543112333784:192.168.101.     33@tcp:101
4189 00010000:00010000:1.0:1443429106.396351:0:14812:0:(ldlm_request.c:639:ldlm_cli_enqueue_fini()) ### server returned different mode PR ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c     2a0fb9fa48c lrc: 4/1,0 mode: --/CR res: [0x200000401:0x1:0x0].0 bits 0x2 rrc: 2 type: IBT flags: 0x0 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4190 00010000:00010000:1.0:1443429106.396354:0:14812:0:(ldlm_request.c:665:ldlm_cli_enqueue_fini()) ### client-side enqueue, new policy data ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d     58c2a0fb9fa48c lrc: 4/1,0 mode: --/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x0 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4191 00010000:00010000:1.0:1443429106.396356:0:14812:0:(ldlm_lock.c:1047:ldlm_granted_list_add_lock()) ### About to add lock: ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb9fa48c      lrc: 4/1,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4192 00010000:00010000:1.0:1443429106.396359:0:14812:0:(ldlm_request.c:710:ldlm_cli_enqueue_fini()) ### client-side enqueue END ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb9fa48     c lrc: 4/1,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x0 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4193 00010000:00010000:1.0:1443429106.396362:0:14812:0:(ldlm_lock.c:746:ldlm_lock_addref_internal_nolock()) ### ldlm_lock_addref(PR) ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb     9fa48c lrc: 5/2,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4194 00010000:00010000:1.0:1443429106.396364:0:14812:0:(ldlm_lock.c:800:ldlm_lock_decref_internal_nolock()) ### ldlm_lock_decref(CR) ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb     9fa48c lrc: 5/2,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4195 00010000:00010000:1.0:1443429106.396366:0:14812:0:(ldlm_lock.c:891:ldlm_lock_decref_internal()) ### do not add lock into lru list ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0     fb9fa48c lrc: 4/1,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4196 00000002:00100000:1.0:1443429106.396369:0:14812:0:(mdc_locks.c:609:mdc_finish_enqueue()) @@@ op: 8 disposition: b, status: 0  req@ffff88011bf403c0 x1513543112333784/t0(0) o101-&amp;gt;lustre-MDT0000-mdc-ffff880     1198a9400@192.168.101.33@tcp:12/10 lens 696/536 e 0 to 0 dl 1443429113 ref 1 fl Complete:R/0/0 rc 0/0
4197 00000002:00010000:1.0:1443429106.396373:0:14812:0:(mdc_locks.c:979:mdc_finish_intent_lock()) ### matching against this ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb9fa48c lr     c: 3/1,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x0 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4198 00010000:00010000:1.0:1443429106.396383:0:14812:0:(ldlm_lock.c:1493:ldlm_lock_match()) ### not matched ns ffff88011a7faa00 type 13 mode 4 res 8589935617/1 (0 0)
4199 00000080:00200000:1.0:1443429106.396390:0:14812:0:(llite_lib.c:1952:ll_update_inode()) inode=[0x200000401:0x1:0x0], updating i_size 4096
4200 00000080:00010000:1.0:1443429106.396392:0:14812:0:(dcache.c:346:ll_lookup_finish_locks()) setting l_data to inode [0x200000401:0x1:0x0](ffff88011bc48140)
4201 00000080:00010000:1.0:1443429106.396393:0:14812:0:(llite_internal.h:1394:ll_set_lock_data()) setting l_data to inode [0x200000401:0x1:0x0](ffff88011bc48140) for lock 0x2d58c2a0fb9fa48c
4202 00000080:00010000:1.0:1443429106.396395:0:14812:0:(dcache.c:249:ll_intent_drop_lock()) releasing lock with cookie 0x2d58c2a0fb9fa48c from it ffff88011bf77d28
4203 00010000:00010000:1.0:1443429106.396395:0:14812:0:(ldlm_lock.c:800:ldlm_lock_decref_internal_nolock()) ### ldlm_lock_decref(PR) ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb     9fa48c lrc: 3/1,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4204 00010000:00010000:1.0:1443429106.396397:0:14812:0:(ldlm_lock.c:874:ldlm_lock_decref_internal()) ### add lock into lru list ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd64ac0/0x2d58c2a0fb9fa48     c lrc: 2/0,0 mode: PR/PR res: [0x200000401:0x1:0x0].0 bits 0x13 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129b042 expref: -99 pid: 14812 timeout: 0 lvb_type: 0
4205 00000080:00200000:1.0:1443429106.396417:0:14812:0:(file.c:3217:__ll_inode_revalidate()) VFS Op:inode=[0x200000007:0x1:0x0](ffff88011bc49c40),name=/
4206 00000002:00010000:1.0:1443429106.396419:0:14812:0:(mdc_locks.c:1114:mdc_intent_lock()) (name: ,[0x200000007:0x1:0x0]) in obj [0x200000007:0x1:0x0], intent: lookup flags 00
4207 00010000:00010000:1.0:1443429106.396421:0:14812:0:(ldlm_lock.c:746:ldlm_lock_addref_internal_nolock()) ### ldlm_lock_addref(PR) ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd644c0/0x2d58c2a0fb     9fa415 lrc: 2/1,0 mode: PR/PR res: [0x200000007:0x1:0x0].0 bits 0x11 rrc: 3 type: IBT flags: 0x0 nid: local remote: 0x1c3b041d4129ae97 expref: -99 pid: 11235 timeout: 0 lvb_type: 0
4208 00010000:00010000:1.0:1443429106.396423:0:14812:0:(ldlm_lock.c:1472:ldlm_lock_match()) ### matched (0 0) ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd644c0/0x2d58c2a0fb9fa415 lrc: 2/1,0 mode:      PR/PR res: [0x200000007:0x1:0x0].0 bits 0x11 rrc: 2 type: IBT flags: 0x0 nid: local remote: 0x1c3b041d4129ae97 expref: -99 pid: 11235 timeout: 0 lvb_type: 0
4209 00000080:00010000:1.0:1443429106.396426:0:14812:0:(dcache.c:346:ll_lookup_finish_locks()) setting l_data to inode [0x200000007:0x1:0x0](ffff88011bc49c40)
4210 00000080:00010000:1.0:1443429106.396427:0:14812:0:(llite_internal.h:1394:ll_set_lock_data()) setting l_data to inode [0x200000007:0x1:0x0](ffff88011bc49c40) for lock 0x2d58c2a0fb9fa415
4211 00000080:00010000:1.0:1443429106.396428:0:14812:0:(dcache.c:249:ll_intent_drop_lock()) releasing lock with cookie 0x2d58c2a0fb9fa415 from it ffff88011bf77b68
4212 00010000:00010000:1.0:1443429106.396428:0:14812:0:(ldlm_lock.c:800:ldlm_lock_decref_internal_nolock()) ### ldlm_lock_decref(PR) ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd644c0/0x2d58c2a0fb     9fa415 lrc: 3/1,0 mode: PR/PR res: [0x200000007:0x1:0x0].0 bits 0x11 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129ae97 expref: -99 pid: 11235 timeout: 0 lvb_type: 0
4213 00010000:00010000:1.0:1443429106.396431:0:14812:0:(ldlm_lock.c:874:ldlm_lock_decref_internal()) ### add lock into lru list ns: lustre-MDT0000-mdc-ffff8801198a9400 lock: ffff88011bd644c0/0x2d58c2a0fb9fa41     5 lrc: 2/0,0 mode: PR/PR res: [0x200000007:0x1:0x0].0 bits 0x11 rrc: 2 type: IBT flags: 0x10000000000000 nid: local remote: 0x1c3b041d4129ae97 expref: -99 pid: 11235 timeout: 0 lvb_type: 0
4214 00000080:00200000:1.0:1443429106.396435:0:14812:0:(file.c:3510:ll_inode_permission()) VFS Op:inode=[0x200000007:0x1:0x0](ffff88011bc49c40), inode mode 41ed mask 1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment>Configuration : 4 node setup : 1 MDS/ 1 OSS/ 2 clients&lt;br/&gt;
Release&lt;br/&gt;
Server 2.7.60&lt;br/&gt;
Client 2.7.60</environment>
        <key id="32656">LU-7302</key>
            <summary>sanity-lfsck test_23b hangs on START_NAMESPACE</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="yong.fan">nasf</assignee>
                                    <reporter username="parinay">parinay v kondekar</reporter>
                        <labels>
                    </labels>
                <created>Thu, 15 Oct 2015 12:22:19 +0000</created>
                <updated>Wed, 25 Jan 2017 19:01:56 +0000</updated>
                            <resolved>Tue, 14 Jun 2016 03:54:53 +0000</resolved>
                                                    <fixVersion>Lustre 2.9.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>7</watches>
                                                                            <comments>
                            <comment id="130541" author="jgmitter" created="Thu, 15 Oct 2015 17:39:28 +0000"  >&lt;p&gt;Hi Fan Yong,&lt;br/&gt;
Can you have a look to see if this is a failure?&lt;br/&gt;
Thanks.&lt;br/&gt;
Joe&lt;/p&gt;</comment>
                            <comment id="131360" author="yong.fan" created="Fri, 23 Oct 2015 14:09:37 +0000"  >&lt;p&gt;The debug log does not contains valid lfsck information, the lfsck entry function lfsck_start() has not been run. So the system should has been blocked before that. If the issues can be reproduced, please enable &quot;-1&quot; level debug on the MDS/OSS to collect more information. The &quot;dmesg&quot; is also valuable sometimes.&lt;/p&gt;</comment>
                            <comment id="131504" author="parinay" created="Mon, 26 Oct 2015 09:43:19 +0000"  >&lt;p&gt;Latest lctl.dk&lt;/p&gt;</comment>
                            <comment id="132290" author="yong.fan" created="Mon, 2 Nov 2015 04:38:59 +0000"  >&lt;p&gt;The latest lctl.dk shows that the test_23b was NOT hung as original exported. But the test result still failed because no dangling name entry was found, instead, the dangling name entry was removed as invalid one.&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00100000:10000000:0.0:1444529448.836945:0:31689:0:(lfsck_namespace.c:2155:lfsck_namespace_repair_dirent()) lustre-MDT0000-osd: namespace LFSCK assistant found bad name entry for: parent [0x200000401:0x2:0x0], child [0x200000402:0x1:0x0], name foo, type in name entry 100000, type claimed by child 0. repair it by removing with new name2 foo: rc = 1
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;That is unexpected. Unfortunately, the -1 level debug log was not enabled (on the MDS) during the test. So we cannot know what caused such unexpected result.&lt;/p&gt;

&lt;p&gt;Parinay, would you please to reproduce the failure with -1 level debug enabled on the MDS? Thanks!&lt;br/&gt;
For example:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;PTLDEBUG=-1 sh sanity-lfsck.sh
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="138774" author="parinay" created="Wed, 13 Jan 2016 07:10:29 +0000"  >&lt;p&gt;Fan Yong,&lt;br/&gt;
Kirtan, tried, to reproduce this with PTLDEBUG=-1 on single node setup. I am attaching the same here. We are trying to reproduce it on 4 node setup.&lt;/p&gt;

&lt;p&gt;Thanks.&lt;/p&gt;</comment>
                            <comment id="140180" author="yong.fan" created="Wed, 27 Jan 2016 09:53:48 +0000"  >&lt;p&gt;There are two reasons for the failure:&lt;/p&gt;

&lt;p&gt;1) OI scrub has been triggered unexpectedly when lookup an remove inode. Such issue has been fixed via the patch &lt;a href=&quot;http://review.whamcloud.com/16439&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/16439&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00100000:10000000:0.0:1452490115.979072:0:4842:0:(osd_scrub.c:653:osd_scrub_prep()) lustre-MDT0000: OI scrub prep, flags = 0x52
00100000:10000000:0.0:1452490115.979078:0:4842:0:(osd_scrub.c:1488:osd_scrub_main()) lustre-MDT0000: OI scrub start, flags = 0x52, pos = 12
00080000:12000000:0.0:1452490115.979079:0:4137:0:(osd_handler.c:3948:osd_consistency_check()) lustre-MDT0000: trigger OI scrub by RPC for [0x200000401:0x4:0x0], rc = 0 [2]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;2) There is some improper logic inside OI scrub, and caused the osd_otable_it_next() and OI scrub thread to wait for each other.&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;00100000:00000001:0.0:1452490116.019213:0:4853:0:(osd_scrub.c:2784:osd_otable_it_load()) Process entered
00100000:00000001:0.0:1452490116.019214:0:4853:0:(osd_scrub.c:2686:osd_otable_it_next()) Process entered
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;The hung point is at:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;static int osd_otable_it_next(const struct lu_env *env, struct dt_it *di)
{
...
        if (it-&amp;gt;ooi_cache.ooc_pos_preload &amp;gt;= scrub-&amp;gt;os_pos_current)
                l_wait_event(thread-&amp;gt;t_ctl_waitq,
                             osd_otable_it_wakeup(scrub, it),
                             &amp;amp;lwi);
...
}
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;With the 1st issue resolved, test_23b trouble will not be triggered again, but the 2nd issue still may cause other problem. I will make patch to fix that.&lt;/p&gt;</comment>
                            <comment id="140186" author="gerrit" created="Wed, 27 Jan 2016 11:50:07 +0000"  >&lt;p&gt;Fan Yong (fan.yong@intel.com) uploaded a new patch: &lt;a href=&quot;http://review.whamcloud.com/18175&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/18175&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7302&quot; title=&quot;sanity-lfsck test_23b hangs on START_NAMESPACE&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7302&quot;&gt;&lt;del&gt;LU-7302&lt;/del&gt;&lt;/a&gt; scrub: join the running OI scrub properly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: 654582bb0821abc53f85c4f755295baec6fb7a28&lt;/p&gt;</comment>
                            <comment id="155605" author="gerrit" created="Tue, 14 Jun 2016 03:50:49 +0000"  >&lt;p&gt;Oleg Drokin (oleg.drokin@intel.com) merged in patch &lt;a href=&quot;http://review.whamcloud.com/18175/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/18175/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-7302&quot; title=&quot;sanity-lfsck test_23b hangs on START_NAMESPACE&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-7302&quot;&gt;&lt;del&gt;LU-7302&lt;/del&gt;&lt;/a&gt; scrub: join the running OI scrub properly&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: 93205e724bd7a07ef56636d1042a430777b89405&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                                        </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                            <attachment id="19303" name="23b.lctl.tgz" size="580244" author="parinay" created="Thu, 15 Oct 2015 12:22:19 +0000"/>
                            <attachment id="19439" name="26th.oct.23b.lctl.tgz" size="2026954" author="parinay" created="Mon, 26 Oct 2015 09:43:19 +0000"/>
                            <attachment id="20100" name="MRP_2979_ptldebug5.txt" size="1946204" author="parinay" created="Wed, 13 Jan 2016 07:10:29 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzxqj3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>