<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 03:34:43 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-17349] sanity-quota test_81: Kernel panic - not syncing: softlockup: hung tasks</title>
                <link>https://jira.whamcloud.com/browse/LU-17349</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;test_81 failed with the following error:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;trevis-68vm7 crashed during sanity-quota test_81
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Dmesg from vmcore:&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[42692.257956] Lustre: DEBUG MARKER: /usr/sbin/lctl mark == sanity-quota test 81: Race qmt_start_pool_recalc with qmt_pool_free =============================== 05:00:18 \(1700888418\)
[42692.536584] Lustre: DEBUG MARKER: == sanity-quota test 81: Race qmt_start_pool_recalc with qmt_pool_free =============================== 05:00:18 (1700888418)
[42692.749799] Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n osc.*MDT*.sync_*
[42693.171476] Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n osp.*.destroys_in_flight
[42694.725049] Lustre: DEBUG MARKER: lctl set_param fail_val=0 fail_loc=0
[42695.134885] Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n version 2&amp;gt;/dev/null
[42695.547239] Lustre: DEBUG MARKER: lfs --list-commands
[42695.915540] Lustre: DEBUG MARKER: /usr/sbin/lctl conf_param lustre.quota.ost=ugp
[42701.359171] Lustre: DEBUG MARKER: /usr/sbin/lctl mark User quota \(block hardlimit:20 MB\)
[42701.570888] Lustre: DEBUG MARKER: User quota (block hardlimit:20 MB)
[42701.765689] Lustre: DEBUG MARKER: lctl pool_new lustre.qpool1
[42704.157598] Lustre: DEBUG MARKER: lctl get_param -n lod.lustre-MDT0000-mdtlov.pools.qpool1 				2&amp;gt;/dev/null || echo foo
[42704.557285] Lustre: DEBUG MARKER: /usr/sbin/lctl set_param fail_loc=0x80000A07 fail_val=30
[42704.949475] Lustre: DEBUG MARKER: /usr/sbin/lctl pool_add lustre.qpool1 lustre-OST[0-0/1]
[42706.444500] LustreError: 12680:0:(fail.c:130:__cfs_fail_timeout_set()) cfs_fail_timeout id a07 sleeping for 30000ms
[42707.325568] Lustre: DEBUG MARKER: lctl get_param -n lod.lustre-MDT0000-mdtlov.pools.qpool1 |
				sort -u | tr &apos;\n&apos; &apos; &apos; 
[42707.702378] Lustre: DEBUG MARKER: grep -c /mnt/lustre-mds1&apos; &apos; /proc/mounts || true
[42708.083111] Lustre: DEBUG MARKER: umount -d -f /mnt/lustre-mds1
[42711.531547] sched: RT throttling activated
[42734.489940] NMI watchdog: BUG: soft lockup - CPU#1 stuck for 21s! [ldlm_bl_03:9064]
[42734.492361] Modules linked in: nfsd nfs_acl dm_flakey lustre(OE) obdecho(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) ptlrpc_gss(OE) osp(OE) mdd(OE) lod(OE) mdt(OE) lfsck(OE) mgs(OE) mgc(OE) osd_ldiskfs(OE) lquota(OE) fid(OE) fld(OE) ksocklnd(OE) ptlrpc(OE) obdclass(OE) lnet(OE) ldiskfs(OE) libcfs(OE) dm_mod rpcsec_gss_krb5 auth_rpcgss nfsv4 dns_resolver nfs lockd grace fscache rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod crc_t10dif crct10dif_generic ib_srp scsi_transport_srp scsi_tgt ib_ipoib rdma_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_core sunrpc iosf_mbi crc32_pclmul ghash_clmulni_intel ppdev aesni_intel lrw gf128mul glue_helper joydev ablk_helper cryptd pcspkr virtio_balloon i2c_piix4 parport_pc parport ip_tables ext4 mbcache jbd2 ata_generic pata_acpi
[42734.888235]  virtio_blk virtio_net net_failover failover crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata virtio_pci virtio_ring virtio floppy [last unloaded: dm_flakey]
[42734.926022] CPU: 1 PID: 9064 Comm: ldlm_bl_03 Kdump: loaded Tainted: G           OE  ------------   3.10.0-1160.83.1.el7_lustre.ddn17.x86_64 #1
[42734.928412] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
[42734.929490] task: ffff88382c5f1080 ti: ffff88382f584000 task.ti: ffff88382f584000
[42734.930874] RIP: 0010:[&amp;lt;ffffffff931b9e1b&amp;gt;]  [&amp;lt;ffffffff931b9e1b&amp;gt;] native_safe_halt+0xb/0x30
[42735.066856] RSP: 0018:ffff88382f587c50  EFLAGS: 00000202
[42735.067898] RAX: 0000000000000005 RBX: ffff88382f587c10 RCX: 0000000000000000
[42735.069221] RDX: 0000000000000005 RSI: 0000000000000005 RDI: ffff93b84581c000
[42735.070544] RBP: ffff88382f587c50 R08: 0000000000000000 R09: 00000000000000b0
[42735.071866] R10: 0000000000000000 R11: 0000000000000001 R12: 0000090900000010
[42735.073188] R13: 0000236800000000 R14: 000000000000154f R15: 0000000000533f52
[42735.074498] FS:  0000000000000000(0000) GS:ffff88387fd00000(0000) knlGS:0000000000000000
[42735.076001] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[42735.077070] CR2: 0000000000519001 CR3: 00000000bb992000 CR4: 00000000000606e0
[42735.089216] Call Trace:
[42735.097974]  [&amp;lt;ffffffff92a6fd95&amp;gt;] kvm_wait+0x65/0x70
[42735.103299]  [&amp;lt;ffffffff92b1e95e&amp;gt;] __pv_queued_spin_lock_slowpath+0x29e/0x2e0
[42735.111331]  [&amp;lt;ffffffff931ac21a&amp;gt;] queued_spin_lock_slowpath+0xb/0x13
[42735.112554]  [&amp;lt;ffffffff931ba734&amp;gt;] _raw_spin_lock+0x24/0x30
[42735.114265]  [&amp;lt;ffffffffc06cdec9&amp;gt;] cfs_hash_spin_lock+0x9/0x10 [libcfs]
[42735.115552]  [&amp;lt;ffffffffc06cea57&amp;gt;] cfs_hash_for_each_relax+0x147/0x450 [libcfs]
[42735.581459]  [&amp;lt;ffffffffc0bb9e30&amp;gt;] ? ldlm_lock_mode_downgrade+0x340/0x340 [ptlrpc]
[42735.582936]  [&amp;lt;ffffffffc0bb9e30&amp;gt;] ? ldlm_lock_mode_downgrade+0x340/0x340 [ptlrpc]
[42735.584409]  [&amp;lt;ffffffffc06d1ed5&amp;gt;] cfs_hash_for_each_nolock+0x75/0x1d0 [libcfs]
[42735.586451]  [&amp;lt;ffffffffc0bbac8c&amp;gt;] ldlm_reprocess_recovery_done+0x3c/0x120 [ptlrpc]
[42735.588445]  [&amp;lt;ffffffffc0bbba3c&amp;gt;] ldlm_export_cancel_locks+0x17c/0x190 [ptlrpc]
[42735.598443]  [&amp;lt;ffffffffc0be72f0&amp;gt;] ldlm_bl_thread_main+0x780/0xab0 [ptlrpc]
[42735.600524]  [&amp;lt;ffffffff92acc680&amp;gt;] ? wake_up_atomic_t+0x40/0x40
[42735.601684]  [&amp;lt;ffffffffc0be6b70&amp;gt;] ? ldlm_handle_bl_callback+0x400/0x400 [ptlrpc]
[42735.603122]  [&amp;lt;ffffffff92acb511&amp;gt;] kthread+0xd1/0xe0
[42735.604049]  [&amp;lt;ffffffff92acb440&amp;gt;] ? insert_kthread_work+0x40/0x40
[42735.605692]  [&amp;lt;ffffffff931c51f7&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[42735.606909]  [&amp;lt;ffffffff92acb440&amp;gt;] ? insert_kthread_work+0x40/0x40
[42735.608052] Code: ff ff 66 0f 1f 44 00 00 0f 00 2d fd e7 04 00 e9 14 ff ff ff 66 2e 0f 1f 84 00 00 00 00 00 66 90 55 48 89 e5 66 66 66 66 90 fb f4 &amp;lt;5d&amp;gt; c3 cc cc cc cc 0f 1f 80 00 00 00 00 0f 00 2d 21 f8 04 00 eb 
[42735.613869] Kernel panic - not syncing: softlockup: hung tasks
[42735.614975] CPU: 1 PID: 9064 Comm: ldlm_bl_03 Kdump: loaded Tainted: G           OEL ------------   3.10.0-1160.83.1.el7_lustre.ddn17.x86_64 #1
[42735.617327] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
[42735.618420] Call Trace:
[42735.618895]  &amp;lt;IRQ&amp;gt;  [&amp;lt;ffffffff931b1bec&amp;gt;] dump_stack+0x19/0x1f
[42735.620023]  [&amp;lt;ffffffff931ab708&amp;gt;] panic+0xe8/0x21f
[42735.626270]  [&amp;lt;ffffffff92b56f5a&amp;gt;] watchdog_timer_fn+0x20a/0x220
[42735.627420]  [&amp;lt;ffffffff92b56d50&amp;gt;] ? watchdog+0x50/0x50
[42735.628385]  [&amp;lt;ffffffff92acfa0e&amp;gt;] __hrtimer_run_queues+0x10e/0x270
[42735.629567]  [&amp;lt;ffffffff92acff9f&amp;gt;] hrtimer_interrupt+0xaf/0x1e0
[42735.636517]  [&amp;lt;ffffffff92a5fa3b&amp;gt;] local_apic_timer_interrupt+0x3b/0x70
[42735.637771]  [&amp;lt;ffffffff931ca053&amp;gt;] smp_apic_timer_interrupt+0x43/0x60
[42735.638948]  [&amp;lt;ffffffff931c63f2&amp;gt;] apic_timer_interrupt+0x172/0x180
[42735.640111]  &amp;lt;EOI&amp;gt;  [&amp;lt;ffffffff931b9e1b&amp;gt;] ? native_safe_halt+0xb/0x30
[42735.641339]  [&amp;lt;ffffffff92a6fd95&amp;gt;] kvm_wait+0x65/0x70
[42735.642265]  [&amp;lt;ffffffff92b1e95e&amp;gt;] __pv_queued_spin_lock_slowpath+0x29e/0x2e0
[42735.643592]  [&amp;lt;ffffffff931ac21a&amp;gt;] queued_spin_lock_slowpath+0xb/0x13
[42735.644781]  [&amp;lt;ffffffff931ba734&amp;gt;] _raw_spin_lock+0x24/0x30
[42735.645820]  [&amp;lt;ffffffffc06cdec9&amp;gt;] cfs_hash_spin_lock+0x9/0x10 [libcfs]
[42735.647043]  [&amp;lt;ffffffffc06cea57&amp;gt;] cfs_hash_for_each_relax+0x147/0x450 [libcfs]
[42735.648417]  [&amp;lt;ffffffffc0bb9e30&amp;gt;] ? ldlm_lock_mode_downgrade+0x340/0x340 [ptlrpc]
[42735.649822]  [&amp;lt;ffffffffc0bb9e30&amp;gt;] ? ldlm_lock_mode_downgrade+0x340/0x340 [ptlrpc]
[42735.651229]  [&amp;lt;ffffffffc06d1ed5&amp;gt;] cfs_hash_for_each_nolock+0x75/0x1d0 [libcfs]
[42735.652587]  [&amp;lt;ffffffffc0bbac8c&amp;gt;] ldlm_reprocess_recovery_done+0x3c/0x120 [ptlrpc]
[42735.654041]  [&amp;lt;ffffffffc0bbba3c&amp;gt;] ldlm_export_cancel_locks+0x17c/0x190 [ptlrpc]
[42735.655419]  [&amp;lt;ffffffffc0be72f0&amp;gt;] ldlm_bl_thread_main+0x780/0xab0 [ptlrpc]
[42735.656699]  [&amp;lt;ffffffff92acc680&amp;gt;] ? wake_up_atomic_t+0x40/0x40
[42735.657818]  [&amp;lt;ffffffffc0be6b70&amp;gt;] ? ldlm_handle_bl_callback+0x400/0x400 [ptlrpc]
[42735.659216]  [&amp;lt;ffffffff92acb511&amp;gt;] kthread+0xd1/0xe0
[42735.660139]  [&amp;lt;ffffffff92acb440&amp;gt;] ? insert_kthread_work+0x40/0x40
[42735.661276]  [&amp;lt;ffffffff931c51f7&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[42735.662488]  [&amp;lt;ffffffff92acb440&amp;gt;] ? insert_kthread_work+0x40/0x40 &lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;</description>
                <environment></environment>
        <key id="79489">LU-17349</key>
            <summary>sanity-quota test_81: Kernel panic - not syncing: softlockup: hung tasks</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="1">Fixed</resolution>
                                        <assignee username="scherementsev">Sergey Cheremencev</assignee>
                                    <reporter username="scherementsev">Sergey Cheremencev</reporter>
                        <labels>
                    </labels>
                <created>Fri, 8 Dec 2023 19:22:21 +0000</created>
                <updated>Tue, 9 Jan 2024 01:11:27 +0000</updated>
                            <resolved>Wed, 20 Dec 2023 02:46:38 +0000</resolved>
                                                    <fixVersion>Lustre 2.16.0</fixVersion>
                                        <due></due>
                            <votes>0</votes>
                                    <watches>3</watches>
                                                                            <comments>
                            <comment id="396079" author="gerrit" created="Fri, 8 Dec 2023 19:30:30 +0000"  >&lt;p&gt;&quot;Sergey Cheremencev &amp;lt;scherementsev@ddn.com&amp;gt;&quot; uploaded a new patch: &lt;a href=&quot;https://review.whamcloud.com/c/fs/lustre-release/+/53384&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/c/fs/lustre-release/+/53384&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-17349&quot; title=&quot;sanity-quota test_81: Kernel panic - not syncing: softlockup: hung tasks&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-17349&quot;&gt;&lt;del&gt;LU-17349&lt;/del&gt;&lt;/a&gt; tests: sanity-quota_81 decrease timeout&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: 1&lt;br/&gt;
Commit: e82c155c83c9aae3191ef040103e223497dce5ad&lt;/p&gt;</comment>
                            <comment id="397555" author="gerrit" created="Wed, 20 Dec 2023 02:00:33 +0000"  >&lt;p&gt;&quot;Oleg Drokin &amp;lt;green@whamcloud.com&amp;gt;&quot; merged in patch &lt;a href=&quot;https://review.whamcloud.com/c/fs/lustre-release/+/53384/&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://review.whamcloud.com/c/fs/lustre-release/+/53384/&lt;/a&gt;&lt;br/&gt;
Subject: &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-17349&quot; title=&quot;sanity-quota test_81: Kernel panic - not syncing: softlockup: hung tasks&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-17349&quot;&gt;&lt;del&gt;LU-17349&lt;/del&gt;&lt;/a&gt; tests: sanity-quota_81 decrease timeout&lt;br/&gt;
Project: fs/lustre-release&lt;br/&gt;
Branch: master&lt;br/&gt;
Current Patch Set: &lt;br/&gt;
Commit: b58219ef1edebcb266cbe0dfede491ba5de491d1&lt;/p&gt;</comment>
                            <comment id="397572" author="pjones" created="Wed, 20 Dec 2023 02:46:38 +0000"  >&lt;p&gt;Landed for 2.16&lt;/p&gt;</comment>
                            <comment id="398910" author="adilger" created="Tue, 9 Jan 2024 01:11:27 +0000"  >&lt;p&gt;Hit another lockup (which triggered a crash) like this in sanity-flr:&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sets/112570ae-2e64-4c60-bd13-b1447c7934fa&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sets/112570ae-2e64-4c60-bd13-b1447c7934fa&lt;/a&gt;&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[ 5945.178891] watchdog: BUG: soft lockup - CPU#0 stuck for 22s! [ldlm_bl_02:77462]
[ 5945.182887] watchdog: BUG: soft lockup - CPU#1 stuck for 22s! [ldlm_bl_03:80014]
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;I&apos;ve filed that as a separate bug in &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-17406&quot; title=&quot;sanity-flr test_50A: watchdog: BUG: soft lockup - CPU#0 stuck for 22s&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-17406&quot;&gt;LU-17406&lt;/a&gt;, but it looks very similar to this one.  I think there is a problem with threads contending on a DLM spinlock.&lt;/p&gt;</comment>
                    </comments>
                <issuelinks>
                            <issuelinktype id="10010">
                    <name>Duplicate</name>
                                                                <inwardlinks description="is duplicated by">
                                                        </inwardlinks>
                                    </issuelinktype>
                            <issuelinktype id="10011">
                    <name>Related</name>
                                            <outwardlinks description="is related to ">
                                        <issuelink>
            <issuekey id="79885">LU-17406</issuekey>
        </issuelink>
                            </outwardlinks>
                                                                <inwardlinks description="is related to">
                                        <issuelink>
            <issuekey id="79885">LU-17406</issuekey>
        </issuelink>
                            </inwardlinks>
                                    </issuelinktype>
                    </issuelinks>
                <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i044dr:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>