<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:25:19 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2451] recovery-small test_24b: BUG: soft lockup - CPU#0 stuck for 67s! [ll_imp_inval:4791]</title>
                <link>https://jira.whamcloud.com/browse/LU-2451</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;The recovery-small test 24b hung due to the following soft lockup issue on client:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;18:45:08:Lustre: DEBUG MARKER: == recovery-small test 24b: test dirty page discard due to client eviction == 18:45:02 (1354934702)
18:45:08:Lustre: DEBUG MARKER: cancel_lru_locks osc start
18:45:08:Lustre: DEBUG MARKER: cancel_lru_locks osc stop
18:45:08:LustreError: 11-0: an error occurred while communicating with 10.10.4.151@tcp. The ost_write operation failed with -107
18:45:08:LustreError: 167-0: This client was evicted by lustre-OST0000; in progress operations using this service will fail.
18:45:09:Lustre: 20511:0:(llite_lib.c:2285:ll_dirty_page_discard_warn()) dirty page discard: 10.10.4.150@tcp:/lustre/fid: [0x200000401:0x42:0x0]//d0.recovery-small/d24/f24b-1 may get corrupted (rc -4)
18:45:09:LustreError: 20511:0:(client.c:1060:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff88007b678c00 x1420751262452328/t0(0) o4-&amp;gt;lustre-OST0000-osc-ffff88007b271800@10.10.4.151@tcp:6/4 lens 456/416 e 0 to 0 dl 0 ref 2 fl Rpc:/0/ffffffff rc 0/-1
18:46:20:BUG: soft lockup - CPU#0 stuck for 67s! [ll_imp_inval:4791]
18:46:20:Modules linked in: lustre(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) lquota(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) libcfs(U) nfs fscache nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode virtio_balloon 8139too 8139cp mii i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk virtio_pci virtio_ring virtio pata_acpi ata_generic ata_piix dm_mirror dm_region_hash dm_log dm_mod [last unloaded: libcfs]
18:46:22:CPU 0 
18:46:22:Modules linked in: lustre(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) lquota(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) libcfs(U) nfs fscache nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode virtio_balloon 8139too 8139cp mii i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk virtio_pci virtio_ring virtio pata_acpi ata_generic ata_piix dm_mirror dm_region_hash dm_log dm_mod [last unloaded: libcfs]
18:46:22:
18:46:22:Pid: 4791, comm: ll_imp_inval Not tainted 2.6.32-279.14.1.el6.x86_64 #1 Red Hat KVM
18:46:22:RIP: 0010:[&amp;lt;ffffffff8150098e&amp;gt;]  [&amp;lt;ffffffff8150098e&amp;gt;] _spin_lock+0x1e/0x30
18:46:22:RSP: 0018:ffff88007b60de00  EFLAGS: 00000206
18:46:22:RAX: 0000000000000001 RBX: ffff88007b60de00 RCX: 0000000000000000
18:46:22:RDX: 0000000000000000 RSI: ffffffffa04dfce6 RDI: ffff88007c616734
18:46:22:RBP: ffffffff8100bc0e R08: 00000000ffffff0a R09: 0000000000000000
18:46:23:R10: 000000000000000f R11: 000000000000000f R12: 0000000000000010
18:46:23:R13: ffffffffa0848ef1 R14: ffff88007b60ddc0 R15: 0000000000000000
18:46:24:FS:  00007f65fc14a700(0000) GS:ffff880002200000(0000) knlGS:0000000000000000
18:46:24:CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
18:46:24:CR2: 000000000208b3c0 CR3: 000000003796c000 CR4: 00000000000006f0
18:46:24:DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
18:46:24:DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
18:46:24:Process ll_imp_inval (pid: 4791, threadinfo ffff88007b60c000, task ffff88007b90b500)
18:46:24:Stack:
18:46:24: ffff88007b60de60 ffffffffa090dead 0000000000000010 ffff880000808003
18:46:24:&amp;lt;d&amp;gt; ffff88007b60de30 000000018109252c ffff88007b60de60 ffff880076109800
18:46:24:&amp;lt;d&amp;gt; ffff88007c616138 ffffffff00000100 ebc0de0100000000 0000000000004e1a
18:46:24:Call Trace:
18:46:25: [&amp;lt;ffffffffa090dead&amp;gt;] ? osc_import_event+0x3ad/0x1470 [osc]
18:46:25: [&amp;lt;ffffffffa0687b01&amp;gt;] ? ptlrpc_invalidate_import+0x2d1/0x910 [ptlrpc]
18:46:25: [&amp;lt;ffffffff810602c0&amp;gt;] ? default_wake_function+0x0/0x20
18:46:25: [&amp;lt;ffffffff811abaef&amp;gt;] ? unshare_fs_struct+0x5f/0xb0
18:46:25: [&amp;lt;ffffffffa0688360&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
18:46:25: [&amp;lt;ffffffffa06883af&amp;gt;] ? ptlrpc_invalidate_import_thread+0x4f/0x2f0 [ptlrpc]
18:46:25: [&amp;lt;ffffffff8100c14a&amp;gt;] ? child_rip+0xa/0x20
18:46:25: [&amp;lt;ffffffffa0688360&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
18:46:25: [&amp;lt;ffffffffa0688360&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
18:46:25: [&amp;lt;ffffffff8100c140&amp;gt;] ? child_rip+0x0/0x20
18:46:26:Code: 00 00 00 01 74 05 e8 62 e0 d7 ff c9 c3 55 48 89 e5 0f 1f 44 00 00 b8 00 00 01 00 3e 0f c1 07 0f b7 d0 c1 e8 10 39 c2 74 0e f3 90 &amp;lt;0f&amp;gt; 1f 44 00 00 83 3f 00 75 f4 eb df c9 c3 0f 1f 40 00 55 48 89 
18:46:26:Call Trace:
18:46:26: [&amp;lt;ffffffffa090dead&amp;gt;] ? osc_import_event+0x3ad/0x1470 [osc]
18:46:26: [&amp;lt;ffffffffa0687b01&amp;gt;] ? ptlrpc_invalidate_import+0x2d1/0x910 [ptlrpc]
18:46:26: [&amp;lt;ffffffff810602c0&amp;gt;] ? default_wake_function+0x0/0x20
18:46:26: [&amp;lt;ffffffff811abaef&amp;gt;] ? unshare_fs_struct+0x5f/0xb0
18:46:26: [&amp;lt;ffffffffa0688360&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
18:46:26: [&amp;lt;ffffffffa06883af&amp;gt;] ? ptlrpc_invalidate_import_thread+0x4f/0x2f0 [ptlrpc]
18:46:26: [&amp;lt;ffffffff8100c14a&amp;gt;] ? child_rip+0xa/0x20
18:46:27: [&amp;lt;ffffffffa0688360&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
18:46:27: [&amp;lt;ffffffffa0688360&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
18:46:27: [&amp;lt;ffffffff8100c140&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Maloo report: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/1d70d272-41da-11e2-adcf-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/1d70d272-41da-11e2-adcf-52540035b04c&lt;/a&gt;&lt;/p&gt;</description>
                <environment>Lustre Branch: b2_1&lt;br/&gt;
Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_1/148&quot;&gt;http://build.whamcloud.com/job/lustre-b2_1/148&lt;/a&gt;&lt;br/&gt;
Distro/Arch: RHEL6.3/x86_64 (kernel version: 2.6.32-279.14.1.el6)&lt;br/&gt;
Network: TCP (1GigE)&lt;br/&gt;
</environment>
        <key id="16881">LU-2451</key>
            <summary>recovery-small test_24b: BUG: soft lockup - CPU#0 stuck for 67s! [ll_imp_inval:4791]</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="3">Duplicate</resolution>
                                        <assignee username="niu">Niu Yawei</assignee>
                                    <reporter username="yujian">Jian Yu</reporter>
                        <labels>
                    </labels>
                <created>Mon, 10 Dec 2012 09:43:41 +0000</created>
                <updated>Tue, 22 Dec 2015 02:53:19 +0000</updated>
                            <resolved>Tue, 22 Dec 2015 02:53:19 +0000</resolved>
                                    <version>Lustre 2.1.4</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="49033" author="niu" created="Tue, 11 Dec 2012 04:13:31 +0000"  >&lt;p&gt;Probaly someone was holding the cl_loi_list_lock for too long time and caused the soft lockup, unfortunately, there isn&apos;t any debuglog and full stack trace on client.&lt;/p&gt;

&lt;p&gt;The test_24b is a new added test, so it&apos;s hard to say if it was a standing problem or caused by recent commits.&lt;/p&gt;</comment>
                            <comment id="49049" author="jay" created="Tue, 11 Dec 2012 11:49:08 +0000"  >&lt;p&gt;liang has fixed a similar issue but not sure if it&apos;s landed. BTW, this may not be loi_list_lock because otherwise debug info would be printed.&lt;/p&gt;</comment>
                            <comment id="49102" author="niu" created="Wed, 12 Dec 2012 01:49:05 +0000"  >&lt;p&gt;I think you were mentioning the &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2263&quot; title=&quot;CPU Soft Lockups due to many threads spinning on import lock on Sequoia IO nodes&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2263&quot;&gt;&lt;del&gt;LU-2263&lt;/del&gt;&lt;/a&gt; &amp;amp; &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2327&quot; title=&quot;Clients stuck in mount&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2327&quot;&gt;&lt;del&gt;LU-2327&lt;/del&gt;&lt;/a&gt;, they are not fixed yet.&lt;/p&gt;

&lt;p&gt;Looks CLIENT_OBD_LIST_LOCK_DEBUG isn&apos;t enabled by default in b2_1, let&apos;s enable it to see if we can capture any useful log.&lt;/p&gt;</comment>
                            <comment id="49103" author="niu" created="Wed, 12 Dec 2012 02:00:17 +0000"  >&lt;p&gt;enable CLIENT_OBD_LIST_LOCK_DEBUG by default: &lt;a href=&quot;http://review.whamcloud.com/4806&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4806&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49176" author="niu" created="Wed, 12 Dec 2012 22:44:03 +0000"  >&lt;p&gt;This bug only happened once on b2_1, maybe we should lower the priority and see if it can be reproduced in the futhure (hopefully, with debug log and full stack trace).&lt;/p&gt;</comment>
                            <comment id="49315" author="yujian" created="Mon, 17 Dec 2012 11:58:13 +0000"  >&lt;p&gt;Lustre Tag: v2_1_4_RC1&lt;br/&gt;
Distro/Arch: RHEL6.3/x86_64&lt;/p&gt;

&lt;p&gt;The same issue occurred again:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/670ac2da-4829-11e2-a406-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/670ac2da-4829-11e2-a406-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49361" author="yujian" created="Tue, 18 Dec 2012 02:40:38 +0000"  >&lt;p&gt;RHEL6.3/x86_64 (2.3.0 Server + 2.1.4 Client):&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/bd3b0534-486f-11e2-8cdc-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/bd3b0534-486f-11e2-8cdc-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49426" author="jay" created="Wed, 19 Dec 2012 01:42:47 +0000"  >&lt;p&gt;will fix it in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="49427" author="jay" created="Wed, 19 Dec 2012 01:47:57 +0000"  >&lt;p&gt;it still makes sense to land the patch 4806 so it will be easier to find out this kind of issue&lt;/p&gt;</comment>
                            <comment id="49429" author="yujian" created="Wed, 19 Dec 2012 01:56:30 +0000"  >&lt;blockquote&gt;&lt;p&gt;enable CLIENT_OBD_LIST_LOCK_DEBUG by default: &lt;a href=&quot;http://review.whamcloud.com/4806&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4806&lt;/a&gt;&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Testing on patch Set 2 build &lt;a href=&quot;http://build.whamcloud.com/job/lustre-reviews/11215/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-reviews/11215/&lt;/a&gt; failed.&lt;/p&gt;

&lt;p&gt;After running recovery-small test 24a and 24b, removing the test directories failed and unmounting client hung.&lt;/p&gt;

&lt;p&gt;Console log on the client showed that:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: DEBUG MARKER: /usr/sbin/lctl mark == recovery-small test 24b: test dirty page discard due to client eviction == 22:23:00 \(1355898180\)
Lustre: DEBUG MARKER: == recovery-small test 24b: test dirty page discard due to client eviction == 22:23:00 (1355898180)
Lustre: DEBUG MARKER: cancel_lru_locks osc start
Lustre: DEBUG MARKER: cancel_lru_locks osc stop
LustreError: 11-0: an error occurred while communicating with 10.10.4.209@tcp. The ost_write operation failed with -107
Lustre: lustre-OST0000-osc-ffff8800740c6000: Connection to lustre-OST0000 (at 10.10.4.209@tcp) was lost; in progress operations using this service will wait for recovery to complete
LustreError: 167-0: This client was evicted by lustre-OST0000; in progress operations using this service will fail.
Lustre: 5975:0:(llite_lib.c:2287:ll_dirty_page_discard_warn()) dirty page discard: 10.10.4.208@tcp:/lustre/fid: [0x200000400:0x9:0x0]//d0.recovery-small/d24/f24b-1 may get corrupted (rc -4)
LustreError: 5975:0:(client.c:1060:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff8800745a1400 x1421761918796156/t0(0) o4-&amp;gt;lustre-OST0000-osc-ffff8800740c6000@10.10.4.209@tcp:6/4 lens 456/416 e 0 to 0 dl 0 ref 2 fl Rpc:/0/ffffffff rc 0/-1
Lustre: ll_imp_inval:11228: lock ffff8800745b8738 was acquired by &amp;lt;ptlrpcd-brw:5975:brw_interpret:2299&amp;gt; for 5 seconds.
Lustre: ====== for process holding the lock =====
Pid: 5975, comm: ptlrpcd-brw

Call Trace:
 [&amp;lt;ffffffff8106244a&amp;gt;] __cond_resched+0x2a/0x40
 [&amp;lt;ffffffff814fea60&amp;gt;] _cond_resched+0x30/0x40
 [&amp;lt;ffffffff811272e5&amp;gt;] __alloc_pages_nodemask+0x165/0x940
 [&amp;lt;ffffffffa03ad831&amp;gt;] ? libcfs_debug_vmsg2+0x4e1/0xb60 [libcfs]
 [&amp;lt;ffffffff8115c7ea&amp;gt;] alloc_pages_current+0xaa/0x110
 [&amp;lt;ffffffff8112572e&amp;gt;] __get_free_pages+0xe/0x50
 [&amp;lt;ffffffffa093b252&amp;gt;] ll_dirty_page_discard_warn+0x42/0x220 [lustre]
 [&amp;lt;ffffffffa03adef1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
 [&amp;lt;ffffffffa0938d6b&amp;gt;] ? ll_queue_done_writing+0xfb/0x300 [lustre]
 [&amp;lt;ffffffffa0978dec&amp;gt;] vvp_page_completion_common+0xcc/0xf0 [lustre]
 [&amp;lt;ffffffffa0978e5b&amp;gt;] vvp_page_completion_write_common+0x4b/0x60 [lustre]
 [&amp;lt;ffffffffa097982e&amp;gt;] vvp_page_completion_write+0x10e/0x230 [lustre]
 [&amp;lt;ffffffffa04f937f&amp;gt;] cl_page_completion+0xff/0x690 [obdclass]
 [&amp;lt;ffffffffa0790c92&amp;gt;] osc_completion+0x162/0x3d0 [osc]
 [&amp;lt;ffffffffa0774904&amp;gt;] osc_ap_completion+0x134/0x330 [osc]
 [&amp;lt;ffffffffa0787e71&amp;gt;] brw_interpret+0x6e1/0x12d0 [osc]
 [&amp;lt;ffffffffa0648a2b&amp;gt;] ? ptlrpc_unregister_bulk+0x1fb/0xae0 [ptlrpc]
 [&amp;lt;ffffffffa0641d4b&amp;gt;] ptlrpc_check_set+0x29b/0x1b00 [ptlrpc]
 [&amp;lt;ffffffffa03adef1&amp;gt;] ? libcfs_debug_msg+0x41/0x50 [libcfs]
 [&amp;lt;ffffffffa0672fd0&amp;gt;] ptlrpcd_check+0x1a0/0x230 [ptlrpc]
 [&amp;lt;ffffffffa067326e&amp;gt;] ptlrpcd+0x20e/0x370 [ptlrpc]
 [&amp;lt;ffffffff810602c0&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffffa0673060&amp;gt;] ? ptlrpcd+0x0/0x370 [ptlrpc]
 [&amp;lt;ffffffff8100c14a&amp;gt;] child_rip+0xa/0x20
 [&amp;lt;ffffffffa0673060&amp;gt;] ? ptlrpcd+0x0/0x370 [ptlrpc]
 [&amp;lt;ffffffffa0673060&amp;gt;] ? ptlrpcd+0x0/0x370 [ptlrpc]
 [&amp;lt;ffffffff8100c140&amp;gt;] ? child_rip+0x0/0x20
 
Lustre: ====== for current process =====
Pid: 11228, comm: ll_imp_inval

Call Trace:
 [&amp;lt;ffffffffa03a27f5&amp;gt;] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
 [&amp;lt;ffffffffa0784d4f&amp;gt;] osc_import_event+0x55f/0x1cc0 [osc]
 [&amp;lt;ffffffffa066dc49&amp;gt;] ptlrpc_invalidate_import+0x2a9/0x8d0 [ptlrpc]
 [&amp;lt;ffffffff810602c0&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffff811abaef&amp;gt;] ? unshare_fs_struct+0x5f/0xb0
 [&amp;lt;ffffffffa066e490&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
 [&amp;lt;ffffffffa066e4df&amp;gt;] ptlrpc_invalidate_import_thread+0x4f/0x2f0 [ptlrpc]
 [&amp;lt;ffffffff8100c14a&amp;gt;] child_rip+0xa/0x20
 [&amp;lt;ffffffffa066e490&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
 [&amp;lt;ffffffffa066e490&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
 [&amp;lt;ffffffff8100c140&amp;gt;] ? child_rip+0x0/0x20

Lustre: ====== end =======
Lustre: 5975:0:(llite_lib.c:2287:ll_dirty_page_discard_warn()) dirty page discard: 10.10.4.208@tcp:/lustre/fid: [0x200000400:0xa:0x0]//d0.recovery-small/d24/f24b-2 may get corrupted (rc -108)
LustreError: 5977:0:(client.c:1060:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff88007d243800 x1421761918796168/t0(0) o101-&amp;gt;lustre-OST0000-osc-ffff8800740c6000@10.10.4.209@tcp:28/4 lens 296/352 e 0 to 0 dl 0 ref 1 fl Rpc:/0/ffffffff rc 0/-1
Lustre: DEBUG MARKER: df /mnt/lustre; uname -n &amp;gt;&amp;gt; /mnt/lustre/recon
LustreError: 11297:0:(namei.c:1128:ll_objects_destroy()) obd destroy objid 0xb error -5
Lustre: DEBUG MARKER: lctl set_param -n fail_loc=0 2&amp;gt;/dev/null || true
Lustre: DEBUG MARKER: /usr/sbin/lctl mark == recovery-small recovery-small.sh test complete, duration 65 sec == 22:23:07 \(1355898187\)
Lustre: DEBUG MARKER: == recovery-small recovery-small.sh test complete, duration 65 sec == 22:23:07 (1355898187)
LustreError: 11636:0:(namei.c:1128:ll_objects_destroy()) obd destroy objid 0x5 error -5
Lustre: DEBUG MARKER: /usr/sbin/lctl mark  recovery-small : @@@@@@ FAIL: remove sub-test dirs failed
Lustre: DEBUG MARKER: recovery-small : @@@@@@ FAIL: remove sub-test dirs failed
Lustre: DEBUG MARKER: /usr/sbin/lctl dk &amp;gt; /home/yujian/test_logs/2012-12-18/221902/recovery-small..debug_log.$(hostname -s).1355898187.log;
         dmesg &amp;gt; /home/yujian/test_logs/2012-12-18/221902/recovery-small..dmesg.$(hostname -s).1355898187.log
Lustre: DEBUG MARKER: running=$(grep -c /mnt/lustre&apos; &apos; /proc/mounts);
if [ $running -ne 0 ] ; then
echo Stopping client $(hostname) /mnt/lustre opts:;
lsof -t /mnt/lustre || need_kill=no;
if [ x != x -a x$need_kill != xno ]; then
    pids=$(lsof -t /mnt/lustre | sort -u);

INFO: task ll_imp_inval:11228 blocked for more than 120 seconds.
&quot;echo 0 &amp;gt; /proc/sys/kernel/hung_task_timeout_secs&quot; disables this message.
ll_imp_inval  D 0000000000000000     0 11228      2 0x00000080
 ffff880073f91d20 0000000000000046 ffffffffa079ccdc 0000000000000000
 ffff880073f91d90 000000008106c04f ffffffffa07ab370 ffff88007d2f9449
 ffff88007433dab8 ffff880073f91fd8 000000000000fb88 ffff88007433dab8
Call Trace:
 [&amp;lt;ffffffff814fee92&amp;gt;] schedule_timeout+0x192/0x2e0
 [&amp;lt;ffffffff8107e1c0&amp;gt;] ? process_timeout+0x0/0x10
 [&amp;lt;ffffffffa03a35ff&amp;gt;] cfs_pause+0x1f/0x30 [libcfs]
 [&amp;lt;ffffffffa0784dad&amp;gt;] osc_import_event+0x5bd/0x1cc0 [osc]
 [&amp;lt;ffffffffa066dc49&amp;gt;] ptlrpc_invalidate_import+0x2a9/0x8d0 [ptlrpc]
 [&amp;lt;ffffffff810602c0&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffff811abaef&amp;gt;] ? unshare_fs_struct+0x5f/0xb0
 [&amp;lt;ffffffffa066e490&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
 [&amp;lt;ffffffffa066e4df&amp;gt;] ptlrpc_invalidate_import_thread+0x4f/0x2f0 [ptlrpc]
 [&amp;lt;ffffffff8100c14a&amp;gt;] child_rip+0xa/0x20
 [&amp;lt;ffffffffa066e490&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
 [&amp;lt;ffffffffa066e490&amp;gt;] ? ptlrpc_invalidate_import_thread+0x0/0x2f0 [ptlrpc]
 [&amp;lt;ffffffff8100c140&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</comment>
                            <comment id="49431" author="niu" created="Wed, 19 Dec 2012 02:37:41 +0000"  >&lt;p&gt;Hi, Yujian, did you try the fix by Jingshan in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;? I think this should be duplicated with &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="49434" author="yujian" created="Wed, 19 Dec 2012 03:26:01 +0000"  >&lt;p&gt;bq Hi, Yujian, did you try the fix by Jingshan in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;? I think this should be duplicated with &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Yes, I just finished verifying the patch of &lt;a href=&quot;http://review.whamcloud.com/4866&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4866&lt;/a&gt; for &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;. Both recovery-small test 24 and lfsck  passed:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/20b2cac6-49b4-11e2-bef8-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/20b2cac6-49b4-11e2-bef8-52540035b04c&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/24c38b8c-49b4-11e2-bef8-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/24c38b8c-49b4-11e2-bef8-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="49437" author="adilger" created="Wed, 19 Dec 2012 04:38:51 +0000"  >&lt;p&gt;Duplicate of &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-2505&quot; title=&quot;lfsck: BUG: soft lockup - CPU#0 stuck for 67s! [umount:22194]&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-2505&quot;&gt;&lt;del&gt;LU-2505&lt;/del&gt;&lt;/a&gt;.&lt;/p&gt;</comment>
                            <comment id="49456" author="jay" created="Wed, 19 Dec 2012 12:06:30 +0000"  >&lt;p&gt;reopen the ticket because I want to land patch 4806.&lt;/p&gt;</comment>
                            <comment id="128720" author="adilger" created="Tue, 29 Sep 2015 08:24:33 +0000"  >&lt;p&gt;Should &lt;a href=&quot;http://review.whamcloud.com/4806&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/4806&lt;/a&gt; be abandoned and this patch be closed?&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                            <outwardlinks description="duplicates">
                                        <issuelink>
            <issuekey id="16958">LU-2505</issuekey>
        </issuelink>
                            </outwardlinks>
                                                        </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvdhj:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5791</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>