<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:23:02 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-2180] Test failure on test suite parallel-scale-nfsv4: BUG: scheduling while atomic: ldlm_elt/8548/0x00000200</title>
                <link>https://jira.whamcloud.com/browse/LU-2180</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;This issue was created by maloo for yujian &amp;lt;yujian@whamcloud.com&amp;gt;&lt;/p&gt;

&lt;p&gt;This issue relates to the following test suite run: &lt;a href=&quot;https://maloo.whamcloud.com/test_sets/44cf5b04-16a1-11e2-afe1-52540035b04c&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://maloo.whamcloud.com/test_sets/44cf5b04-16a1-11e2-afe1-52540035b04c&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Lustre Tag: v2_3_0_RC3&lt;br/&gt;
Lustre Build: &lt;a href=&quot;http://build.whamcloud.com/job/lustre-b2_3/36&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;http://build.whamcloud.com/job/lustre-b2_3/36&lt;/a&gt;&lt;br/&gt;
Distro/Arch: RHEL6.3/x86_64(server), FC15/x86_64(client)&lt;br/&gt;
Network: TCP&lt;br/&gt;
ENABLE_QUOTA=yes&lt;/p&gt;

&lt;p&gt;After running parallel-scale-nfsv4 test_iorfpp, unmounting OST hung, and the following error showed up in the OSS console log:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;Lustre: DEBUG MARKER: == parallel-scale-nfsv4 parallel-scale-nfs.sh test complete, duration 2732 sec ======================= 00:45:35 (1350287135)
LustreError: 138-a: lustre-OST0001: A client on nid 10.10.4.133@tcp was evicted due to a lock blocking callback time out: rc -107
LustreError: 138-a: lustre-OST0004: A client on nid 10.10.4.133@tcp was evicted due to a lock blocking callback time out: rc -107
LustreError: 12155:0:(client.c:1110:ptlrpc_import_delay_req()) @@@ IMP_CLOSED   req@ffff88041c891000 x1415872606374127/t0(0) o104-&amp;gt;lustre-OST0005@10.10.4.133@tcp:15/16 lens 296/224 e 0 to 0 dl 0 ref 1 fl Rpc:N/0/ffffffff rc 0/-1
LustreError: 12155:0:(ldlm_lockd.c:682:ldlm_handle_ast_error()) ### client (nid 10.10.4.133@tcp) returned 0 from blocking AST ns: filter-lustre-OST0005_UUID lock: ffff88011b8e9280/0x952f6899ef78e799 lrc: 4/0,0 mode: PW/PW res: 18768/0 rrc: 2 type: EXT [0-&amp;gt;18446744073709551615] (req 0-&amp;gt;4095) flags: 0x10020 nid: 10.10.4.133@tcp remote: 0xaee7ce80b9fe38c0 expref: 5 pid: 8553 timeout 4301494110
LustreError: Skipped 12 previous similar messages
BUG: scheduling while atomic: ldlm_elt/8548/0x00000200
Modules linked in: nfs fscache obdfilter(U) fsfilt_ldiskfs(U) ost(U) mgc(U) ldiskfs(U) jbd2 lustre(U) lquota(U) lov(U) osc(U) mdc(U) fid(U) fld(U) ksocklnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) sha512_generic sha256_generic libcfs(U) nfsd lockd nfs_acl auth_rpcgss exportfs autofs4 sunrpc cpufreq_ondemand powernow_k8 freq_table mperf ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 ib_sa mlx4_ib ib_mad ib_core mlx4_en mlx4_core igb dca microcode serio_raw k10temp amd64_edac_mod edac_core edac_mce_amd i2c_piix4 i2c_core sg shpchp ext3 jbd mbcache sd_mod crc_t10dif ata_generic pata_acpi pata_atiixp ahci dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]
Pid: 8548, comm: ldlm_elt Not tainted 2.6.32-279.5.1.el6_lustre.g8932d49.x86_64 #1
Call Trace:
 [&amp;lt;ffffffff81057df6&amp;gt;] ? __schedule_bug+0x66/0x70
 [&amp;lt;ffffffff814fe298&amp;gt;] ? thread_return+0x646/0x76e
 [&amp;lt;ffffffff8127a43c&amp;gt;] ? put_dec+0x10c/0x110
 [&amp;lt;ffffffff811ad70e&amp;gt;] ? bh_lru_install+0x16e/0x1a0
 [&amp;lt;ffffffff810923be&amp;gt;] ? prepare_to_wait+0x4e/0x80
 [&amp;lt;ffffffffa0b721cd&amp;gt;] ? do_get_write_access+0x29d/0x520 [jbd2]
 [&amp;lt;ffffffff81092110&amp;gt;] ? wake_bit_function+0x0/0x50
 [&amp;lt;ffffffffa0b725a1&amp;gt;] ? jbd2_journal_get_write_access+0x31/0x50 [jbd2]
 [&amp;lt;ffffffffa0b98478&amp;gt;] ? __ldiskfs_journal_get_write_access+0x38/0x80 [ldiskfs]
 [&amp;lt;ffffffffa0ba4473&amp;gt;] ? ldiskfs_reserve_inode_write+0x73/0xa0 [ldiskfs]
 [&amp;lt;ffffffffa0baa590&amp;gt;] ? ldiskfs_delete_inode+0x0/0x250 [ldiskfs]
 [&amp;lt;ffffffffa0ba44ec&amp;gt;] ? ldiskfs_mark_inode_dirty+0x4c/0x1f0 [ldiskfs]
 [&amp;lt;ffffffffa0b714f0&amp;gt;] ? jbd2_journal_start+0xd0/0x110 [jbd2]
 [&amp;lt;ffffffffa0baa590&amp;gt;] ? ldiskfs_delete_inode+0x0/0x250 [ldiskfs]
 [&amp;lt;ffffffffa0baa64c&amp;gt;] ? ldiskfs_delete_inode+0xbc/0x250 [ldiskfs]
 [&amp;lt;ffffffff8119631e&amp;gt;] ? generic_delete_inode+0xde/0x1d0
 [&amp;lt;ffffffff81196475&amp;gt;] ? generic_drop_inode+0x65/0x80
 [&amp;lt;ffffffff811952c2&amp;gt;] ? iput+0x62/0x70
 [&amp;lt;ffffffffa0cb1bb2&amp;gt;] ? filter_lvbo_free+0x22/0x100 [obdfilter]
 [&amp;lt;ffffffffa0621a38&amp;gt;] ? ldlm_resource_putref+0x128/0x2a0 [ptlrpc]
 [&amp;lt;ffffffffa061a168&amp;gt;] ? ldlm_lock_put+0xc8/0x530 [ptlrpc]
 [&amp;lt;ffffffffa0640200&amp;gt;] ? expired_lock_main+0x660/0x840 [ptlrpc]
 [&amp;lt;ffffffff81060250&amp;gt;] ? default_wake_function+0x0/0x20
 [&amp;lt;ffffffffa063fba0&amp;gt;] ? expired_lock_main+0x0/0x840 [ptlrpc]
 [&amp;lt;ffffffff8100c14a&amp;gt;] ? child_rip+0xa/0x20
 [&amp;lt;ffffffffa063fba0&amp;gt;] ? expired_lock_main+0x0/0x840 [ptlrpc]
 [&amp;lt;ffffffffa063fba0&amp;gt;] ? expired_lock_main+0x0/0x840 [ptlrpc]
 [&amp;lt;ffffffff8100c140&amp;gt;] ? child_rip+0x0/0x20
Lustre: 2532:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: [sent 1350287142/real 1350287142]  req@ffff88021cff4000 x1415872606374144/t0(0) o400-&amp;gt;MGC10.10.4.133@tcp@10.10.4.133@tcp:26/25 lens 224/224 e 0 to 1 dl 1350287149 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
LustreError: 166-1: MGC10.10.4.133@tcp: Connection to MGS (at 10.10.4.133@tcp) was lost; in progress operations using this service will fail
LustreError: 8550:0:(client.c:1125:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff880419f6a400 x1415872606374147/t0(0) o101-&amp;gt;MGC10.10.4.133@tcp@10.10.4.133@tcp:26/25 lens 328/384 e 0 to 0 dl 0 ref 2 fl Rpc:/0/ffffffff rc 0/-1
LustreError: 8550:0:(client.c:1125:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff880419f6a400 x1415872606374148/t0(0) o101-&amp;gt;MGC10.10.4.133@tcp@10.10.4.133@tcp:26/25 lens 328/384 e 0 to 0 dl 0 ref 2 fl Rpc:/0/ffffffff rc 0/-1
Lustre: OST lustre-OST0000 has stopped.
Lustre: Skipped 1 previous similar message
Lustre: server umount lustre-OST0000 complete
Lustre: 2521:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: [sent 1350287149/real 1350287149]  req@ffff88021ba68400 x1415872606374146/t0(0) o250-&amp;gt;MGC10.10.4.133@tcp@10.10.4.133@tcp:26/25 lens 400/544 e 0 to 1 dl 1350287155 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
LustreError: 8550:0:(client.c:1125:ptlrpc_import_delay_req()) @@@ IMP_INVALID  req@ffff88041d1b7000 x1415872606374153/t0(0) o101-&amp;gt;MGC10.10.4.133@tcp@10.10.4.133@tcp:26/25 lens 328/384 e 0 to 0 dl 0 ref 2 fl Rpc:/0/ffffffff rc 0/-1
LustreError: 8550:0:(client.c:1125:ptlrpc_import_delay_req()) Skipped 4 previous similar messages
Lustre: OST lustre-OST0001 has stopped.
Lustre: server umount lustre-OST0001 complete
Lustre: OST lustre-OST0002 has stopped.
Lustre: server umount lustre-OST0002 complete
Lustre: OST lustre-OST0003 has stopped.
Lustre: 2521:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: [sent 1350287159/real 1350287159]  req@ffff88021c977800 x1415872606374159/t0(0) o250-&amp;gt;MGC10.10.4.133@tcp@10.10.4.133@tcp:26/25 lens 400/544 e 0 to 1 dl 1350287170 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: server umount lustre-OST0004 complete
Lustre: Skipped 1 previous similar message
LustreError: 2484:0:(obd_mount.c:257:server_put_mount()) lustre-OST0005: mount busy, vfscount=15!
Lustre: OST lustre-OST0005 has stopped.
Lustre: Skipped 1 previous similar message
Lustre: 2521:0:(client.c:1917:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: [sent 1350287174/real 1350287174]  req@ffff88021cfef800 x1415872606374160/t0(0) o250-&amp;gt;MGC10.10.4.133@tcp@10.10.4.133@tcp:26/25 lens 400/544 e 0 to 1 dl 1350287190 ref 1 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: Mount still busy with 15 refs after 30 secs.
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="16363">LU-2180</key>
            <summary>Test failure on test suite parallel-scale-nfsv4: BUG: scheduling while atomic: ldlm_elt/8548/0x00000200</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="maloo">Maloo</reporter>
                        <labels>
                    </labels>
                <created>Mon, 15 Oct 2012 05:48:14 +0000</created>
                <updated>Fri, 22 Dec 2017 09:36:15 +0000</updated>
                            <resolved>Fri, 22 Dec 2017 09:36:15 +0000</resolved>
                                    <version>Lustre 2.3.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>6</watches>
                                                                            <comments>
                            <comment id="46558" author="johann" created="Mon, 15 Oct 2012 06:21:44 +0000"  >&lt;p&gt;I think Oleg reported a similar bug recently. It looks like expired_lock_main() shouldn&apos;t release lock reference while holding the waiting_locks_spinlock ...&lt;/p&gt;</comment>
                            <comment id="46576" author="green" created="Mon, 15 Oct 2012 11:09:11 +0000"  >&lt;p&gt;Yes, this one is real, though first time I see it really crash.&lt;/p&gt;</comment>
                            <comment id="217062" author="adilger" created="Fri, 22 Dec 2017 09:36:15 +0000"  >&lt;p&gt;Close old bug that has not been hit since initial report.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvadz:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>5218</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>