<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:42:28 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-4410] sanityn test 40a: BUG: soft lockup - CPU#0 stuck for 67s! [ptlrpcd_0:2892]</title>
                <link>https://jira.whamcloud.com/browse/LU-4410</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;sanityn test 40a hung and hit the following failure on one client:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;21:36:52:Lustre: DEBUG MARKER: == sanityn test 40a: pdirops: create vs others ================ 21:34:49 (1387604089)
21:36:53:BUG: soft lockup - CPU#0 stuck for 67s! [ptlrpcd_0:2892]
21:36:53:Modules linked in: lustre(U) obdecho(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) sha512_generic sha256_generic libcfs(U) nfs fscache nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode 8139too 8139cp mii virtio_balloon i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk virtio_pci virtio_ring virtio pata_acpi ata_generic ata_piix dm_mirror dm_region_hash dm_log dm_mod [last unloaded: speedstep_lib]
21:36:53:CPU 0 
21:36:53:Modules linked in: lustre(U) obdecho(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U)
21:36:53:BUG: soft lockup - CPU#1 stuck for 67s! [ll_sa_4070:4079]
21:36:53:Modules linked in: lustre(U) obdecho(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) sha512_generic sha256_generic libcfs(U) nfs fscache nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode 8139too 8139cp mii virtio_balloon i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk virtio_pci virtio_ring virtio pata_acpi ata_generic ata_piix dm_mirror dm_region_hash dm_log dm_mod [last unloaded: speedstep_lib]
21:36:53:CPU 1 
21:36:53:Modules linked in: lustre(U) obdecho(U) mgc(U) lov(U) osc(U) mdc(U) lmv(U) fid(U) fld(U) ptlrpc(U) obdclass(U) lvfs(U) ksocklnd(U) lnet(U) sha512_generic sha256_generic libcfs(U) nfs fscache nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa ib_mad ib_core microcode 8139too 8139cp mii virtio_balloon i2c_piix4 i2c_core ext3 jbd mbcache virtio_blk virtio_pci virtio_ring virtio pata_acpi ata_generic ata_piix dm_mirror dm_region_hash dm_log dm_mod [last unloaded: speedstep_lib]
21:36:53:
21:36:53:Pid: 4079, comm: ll_sa_4070 Not tainted 2.6.32-358.23.2.el6.x86_64 #1 Red Hat KVM
21:36:53:RIP: 0010:[&amp;lt;ffffffff81510aae&amp;gt;]  [&amp;lt;ffffffff81510aae&amp;gt;] _spin_lock+0x1e/0x30
21:36:53:RSP: 0018:ffff88006c26bda0  EFLAGS: 00000206
21:36:53:RAX: 0000000000000002 RBX: ffff88006c26bda0 RCX: ffff88007cfd8800
21:36:54:RDX: 0000000000000000 RSI: ffff88006c25fec0 RDI: ffff88007a737ec0
21:36:54:RBP: ffffffff8100bb8e R08: ffff88007d860e68 R09: 00000000fffffffe
21:36:54:R10: 0000000000000000 R11: 0000000000000001 R12: ffff88006c26bd80
21:36:54:R13: ffff88006d6c9000 R14: 0000000000001000 R15: 0000000000000000
21:36:54:FS:  00007fb227702700(0000) GS:ffff880002300000(0000) knlGS:0000000000000000
21:36:54:CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
21:36:54:CR2: 00007f7bbff64000 CR3: 000000006c183000 CR4: 00000000000006e0
21:36:54:DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
21:36:54:DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
21:36:54:Process ll_sa_4070 (pid: 4079, threadinfo ffff88006c26a000, task ffff88006bd25500)
21:36:54:Stack:
21:36:54: ffff88006c26be10 ffffffffa0abb680 ffff88007a737bf8 ffff88006e9501c8
21:36:54:&amp;lt;d&amp;gt; 0000000000000000 ffff88007a737b00 ffff88007caa01c0 ffff88006bf57200
21:36:54:&amp;lt;d&amp;gt; ffff88006c26bdf0 ffff88007a7ba800 ffff88007a7ba970 ffff88007a737e80
21:36:54:Call Trace:
21:36:54: [&amp;lt;ffffffffa0abb680&amp;gt;] ? ll_post_statahead+0x50/0xa80 [lustre]
21:36:55: [&amp;lt;ffffffffa0abf8c8&amp;gt;] ? ll_statahead_thread+0x268/0xfa0 [lustre]
21:36:55: [&amp;lt;ffffffff81063990&amp;gt;] ? default_wake_function+0x0/0x20
21:36:55: [&amp;lt;ffffffffa0abf660&amp;gt;] ? ll_statahead_thread+0x0/0xfa0 [lustre]
21:36:55: [&amp;lt;ffffffff8100c0ca&amp;gt;] ? child_rip+0xa/0x20
21:36:55: [&amp;lt;ffffffffa0abf660&amp;gt;] ? ll_statahead_thread+0x0/0xfa0 [lustre]
21:36:55: [&amp;lt;ffffffffa0abf660&amp;gt;] ? ll_statahead_thread+0x0/0xfa0 [lustre]
21:36:55: [&amp;lt;ffffffff8100c0c0&amp;gt;] ? child_rip+0x0/0x20
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;Maloo report: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/7cca784a-6b4b-11e3-99ba-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/7cca784a-6b4b-11e3-99ba-52540035b04c&lt;/a&gt;&lt;/p&gt;</description>
                <environment>&lt;br/&gt;
Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_4/70/&quot;&gt;http://build.whamcloud.com/job/lustre-b2_4/70/&lt;/a&gt; (2.4.2 RC2)&lt;br/&gt;
Distro/Arch: RHEL6.4/x86_64&lt;br/&gt;
FSTYPE=zfs&lt;br/&gt;
</environment>
        <key id="22553">LU-4410</key>
            <summary>sanityn test 40a: BUG: soft lockup - CPU#0 stuck for 67s! [ptlrpcd_0:2892]</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="1" iconUrl="https://jira.whamcloud.com/images/icons/priorities/blocker.svg">Blocker</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="yujian">Jian Yu</reporter>
                        <labels>
                            <label>22pl</label>
                    </labels>
                <created>Mon, 23 Dec 2013 06:31:40 +0000</created>
                <updated>Sun, 10 Oct 2021 21:09:03 +0000</updated>
                            <resolved>Sun, 10 Oct 2021 21:09:03 +0000</resolved>
                                    <version>Lustre 2.6.0</version>
                    <version>Lustre 2.4.2</version>
                    <version>Lustre 2.5.2</version>
                    <version>Lustre 2.5.3</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>9</watches>
                                                                            <comments>
                            <comment id="74010" author="yujian" created="Mon, 23 Dec 2013 06:35:40 +0000"  >&lt;p&gt;Here is the query result on Maloo for sanityn test 40a timeout failure on Lustre b2_4 branch:&lt;br/&gt;
&lt;a href=&quot;http://tinyurl.com/ovmwgad&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://tinyurl.com/ovmwgad&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The failure did not occur on previous Lustre b2_4 builds.&lt;/p&gt;</comment>
                            <comment id="74012" author="yujian" created="Mon, 23 Dec 2013 07:56:31 +0000"  >&lt;p&gt;Just ran the same test with FSTYPE=zfs on Lustre 2.4.2 RC2 for 10 times. All runs passed. So, it seems like this is an occasional failure.&lt;/p&gt;</comment>
                            <comment id="85177" author="jamesanunez" created="Thu, 29 May 2014 21:13:26 +0000"  >&lt;p&gt;I hit this problem while testing a b2_5 patch and lustre-rsync_test test_6 timedout. Logs are at &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/0712c6e6-e762-11e3-b2f3-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/0712c6e6-e762-11e3-b2f3-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Note: This was run on ldiskfs, not ZFS&lt;/p&gt;</comment>
                            <comment id="85786" author="yujian" created="Thu, 5 Jun 2014 08:46:57 +0000"  >&lt;p&gt;More instance on Lustre b2_5 branch while running lustre-rsync-test test 6 with FSTYPE=ldiskfs:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/b8b57494-ec5c-11e3-98b3-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/b8b57494-ec5c-11e3-98b3-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="86077" author="yujian" created="Mon, 9 Jun 2014 11:28:11 +0000"  >&lt;p&gt;Hi Nasf,&lt;/p&gt;

&lt;p&gt;It looks like the failure is related to statahead. It originally occurred on Lustre b2_4 branch with ZFS on sanityn test 40a, and now frequently occurs on Lustre b2_5 branch with ldiskfs on lustre-rsync-test test 6. Could you please help take a look if these two test failures have the same root cause? Thanks.&lt;/p&gt;</comment>
                            <comment id="86093" author="yong.fan" created="Mon, 9 Jun 2014 14:21:04 +0000"  >&lt;p&gt;It seems that some thread was blocked with the ll_inode_info::lli_sa_lock held. There is some known bug for it. Here is the patch:&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;http://review.whamcloud.com/#/c/9665/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/9665/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Would you please to try the patch? Thanks!&lt;/p&gt;</comment>
                            <comment id="86183" author="yujian" created="Tue, 10 Jun 2014 01:49:37 +0000"  >&lt;blockquote&gt;&lt;p&gt;Would you please to try the patch?&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;Sure, I&apos;ll do this. Thank you!&lt;/p&gt;</comment>
                            <comment id="86319" author="yujian" created="Wed, 11 Jun 2014 14:35:15 +0000"  >&lt;p&gt;Here is the patch back-ported to Lustre b2_5 branch: &lt;a href=&quot;http://review.whamcloud.com/10674&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10674&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="86605" author="di.wang" created="Fri, 13 Jun 2014 19:27:45 +0000"  >&lt;p&gt;Hmm, I saw similar problem when I run my patch &lt;a href=&quot;http://review.whamcloud.com/#/c/10622/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/#/c/10622/&lt;/a&gt;  on master. &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/196f5da8-f2d5-11e3-b88b-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/196f5da8-f2d5-11e3-b88b-52540035b04c&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Is this also needed on master?&lt;/p&gt;</comment>
                            <comment id="86770" author="yujian" created="Tue, 17 Jun 2014 01:49:49 +0000"  >&lt;blockquote&gt;&lt;p&gt;Here is the patch back-ported to Lustre b2_5 branch: &lt;a href=&quot;http://review.whamcloud.com/10674&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/10674&lt;/a&gt;&lt;/p&gt;&lt;/blockquote&gt;
&lt;p&gt;The patch was reverted from Lustre b2_5 branch because we need wait until master version is fully ready.&lt;/p&gt;</comment>
                            <comment id="86900" author="yujian" created="Wed, 18 Jun 2014 10:26:00 +0000"  >&lt;p&gt;Another sanityn test 40a failure instance on Lustre b2_5 branch:&lt;br/&gt;
&lt;a href=&quot;https://maloo.whamcloud.com/test_sets/17d0d2b0-f5fa-11e3-9d30-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/17d0d2b0-f5fa-11e3-9d30-52540035b04c&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="89195" author="utopiabound" created="Wed, 16 Jul 2014 13:36:52 +0000"  >&lt;p&gt;Another lustre-rsync-test test_6 on master branch review-dne-part-1:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/5f96c5ec-094a-11e4-b76d-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/5f96c5ec-094a-11e4-b76d-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="90946" author="liwei" created="Wed, 6 Aug 2014 00:42:41 +0000"  >&lt;p&gt;lustre-rsync-test 6, master, zfs, single MDT: &lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/67240e86-1cf6-11e4-9a83-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/67240e86-1cf6-11e4-9a83-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="91029" author="yong.fan" created="Thu, 7 Aug 2014 01:05:41 +0000"  >&lt;p&gt;Another failure instance:&lt;/p&gt;

&lt;p&gt;&lt;a href=&quot;https://testing.hpdd.intel.com/test_sessions/6d60501e-1dbb-11e4-8fe8-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sessions/6d60501e-1dbb-11e4-8fe8-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="93291" author="yujian" created="Fri, 5 Sep 2014 00:51:41 +0000"  >&lt;p&gt;While verifying patch &lt;a href=&quot;http://review.whamcloud.com/11615&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/11615&lt;/a&gt; with FSTYPE=zfs on Lustre b2_5 branch, lustre-rsync-test hit the same failure:&lt;br/&gt;
&lt;a href=&quot;https://testing.hpdd.intel.com/test_sets/8cc59532-30ac-11e4-9e60-5254006e85c2&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.hpdd.intel.com/test_sets/8cc59532-30ac-11e4-9e60-5254006e85c2&lt;/a&gt;&lt;/p&gt;</comment>
                            <comment id="107755" author="yujian" created="Tue, 24 Feb 2015 08:54:39 +0000"  >&lt;p&gt;Here is the back-ported patch for Lustre b2_5 branch: &lt;a href=&quot;http://review.whamcloud.com/13846&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://review.whamcloud.com/13846&lt;/a&gt;&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                        <issuelink>
            <issuekey id="26477">LU-5605</issuekey>
        </issuelink>
            <issuelink>
            <issuekey id="28803">LU-6270</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzwbrz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>12104</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>