<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 02:48:46 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-11996] NMI watchdog in lfsck test 18</title>
                <link>https://jira.whamcloud.com/browse/LU-11996</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;Going through maloo crash reports ther&apos;s this NMI watchdog case that seems to only hit on b2_10 and related reviews:&lt;/p&gt;

&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;[ 9012.939072] Lustre: DEBUG MARKER: /usr/sbin/lctl mark == sanity-lfsck test 18g: Find out orphan OST-object and repair it \(7\) =============================== 07:44:51 \(1548229491\)
[ 9013.146289] Lustre: DEBUG MARKER: == sanity-lfsck test 18g: Find out orphan OST-object and repair it (7) =============================== 07:44:51 (1548229491)
[ 9013.350706] Lustre: 20100:0:(osd_internal.h:1155:osd_trans_exec_op()) lustre-OST0000: opcode 2: before 256 &amp;lt; left 274, rollback = 2
[ 9013.352818] Lustre: 20100:0:(osd_internal.h:1155:osd_trans_exec_op()) Skipped 16 previous similar messages
[ 9013.354412] Lustre: 20100:0:(osd_handler.c:1673:osd_trans_dump_creds())   create: 0/0/0, destroy: 0/0/0
[ 9013.356182] Lustre: 20100:0:(osd_handler.c:1673:osd_trans_dump_creds()) Skipped 16 previous similar messages
[ 9013.357856] Lustre: 20100:0:(osd_handler.c:1680:osd_trans_dump_creds())   attr_set: 1/1/0, xattr_set: 2/274/0
[ 9013.359530] Lustre: 20100:0:(osd_handler.c:1680:osd_trans_dump_creds()) Skipped 16 previous similar messages
[ 9013.361235] Lustre: 20100:0:(osd_handler.c:1690:osd_trans_dump_creds())   write: 0/0/0, punch: 0/0/0, quota 3/3/0
[ 9013.362959] Lustre: 20100:0:(osd_handler.c:1690:osd_trans_dump_creds()) Skipped 16 previous similar messages
[ 9013.364612] Lustre: 20100:0:(osd_handler.c:1697:osd_trans_dump_creds())   insert: 0/0/0, delete: 0/0/0
[ 9013.366183] Lustre: 20100:0:(osd_handler.c:1697:osd_trans_dump_creds()) Skipped 16 previous similar messages
[ 9013.367853] Lustre: 20100:0:(osd_handler.c:1704:osd_trans_dump_creds())   ref_add: 0/0/0, ref_del: 0/0/0
[ 9013.369531] Lustre: 20100:0:(osd_handler.c:1704:osd_trans_dump_creds()) Skipped 16 previous similar messages
[ 9054.660354] LNet: Service thread pid 9025 was inactive for 40.01s. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
[ 9054.663155] Pid: 9025, comm: ll_ost_out00_00 3.10.0-862.14.4.el7_lustre.x86_64 #1 SMP Mon Jan 7 20:20:29 UTC 2019
[ 9054.664802] Call Trace:
[ 9054.665266]  [&amp;lt;ffffffffc0c6b130&amp;gt;] ptlrpc_set_wait+0x4c0/0x920 [ptlrpc]
[ 9054.666720]  [&amp;lt;ffffffffc0c6b60d&amp;gt;] ptlrpc_queue_wait+0x7d/0x220 [ptlrpc]
[ 9054.667862]  [&amp;lt;ffffffffc076c4f4&amp;gt;] fld_client_rpc+0x104/0x540 [fld]
[ 9054.668923]  [&amp;lt;ffffffffc077131f&amp;gt;] fld_server_lookup+0x15f/0x320 [fld]
[ 9054.670006]  [&amp;lt;ffffffffc0f61a81&amp;gt;] lfsck_orphan_it_next+0x8c1/0xc90 [lfsck]
[ 9054.671267]  [&amp;lt;ffffffffc0f61ebe&amp;gt;] lfsck_orphan_it_load+0x6e/0x160 [lfsck]
[ 9054.672419]  [&amp;lt;ffffffffc0a52198&amp;gt;] dt_index_walk+0xf8/0x430 [obdclass]
[ 9054.673701]  [&amp;lt;ffffffffc0a5365c&amp;gt;] dt_index_read+0x44c/0x6b0 [obdclass]
[ 9054.674822]  [&amp;lt;ffffffffc0cd8d22&amp;gt;] tgt_obd_idx_read+0x612/0x860 [ptlrpc]
[ 9054.676042]  [&amp;lt;ffffffffc0cdb37a&amp;gt;] tgt_request_handle+0x92a/0x1370 [ptlrpc]
[ 9054.677249]  [&amp;lt;ffffffffc0c83e3b&amp;gt;] ptlrpc_server_handle_request+0x23b/0xaa0 [ptlrpc]
[ 9054.678554]  [&amp;lt;ffffffffc0c87582&amp;gt;] ptlrpc_main+0xa92/0x1e40 [ptlrpc]
[ 9054.679653]  [&amp;lt;ffffffff902bdf21&amp;gt;] kthread+0xd1/0xe0
[ 9054.680543]  [&amp;lt;ffffffff909255f7&amp;gt;] ret_from_fork_nospec_end+0x0/0x39
[ 9054.681615]  [&amp;lt;ffffffffffffffff&amp;gt;] 0xffffffffffffffff
[ 9054.682506] LustreError: dumping log to /tmp/lustre-log.1548229534.9025
[ 9061.992373] LustreError: 7499:0:(fid_handler.c:337:__seq_server_alloc_meta()) srv-lustre-OST0001: Allocated super-sequence failed: rc = -11
[ 9061.994622] LustreError: 7499:0:(fid_handler.c:337:__seq_server_alloc_meta()) Skipped 3569 previous similar messages
[ 9072.297316] NMI watchdog: BUG: soft lockup - CPU#1 stuck for 23s! [lfsck:7414]
[ 9072.302126] Modules linked in: osp(OE) ofd(OE) lfsck(OE) ost(OE) mgc(OE) osd_ldiskfs(OE) lquota(OE) fid(OE) fld(OE) ksocklnd(OE) ptlrpc(OE) obdclass(OE) lnet(OE) ldiskfs(OE) libcfs(OE) rpcsec_gss_krb5 auth_rpcgss nfsv4 dns_resolver nfs lockd grace fscache rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod crc_t10dif crct10dif_generic ib_srp scsi_transport_srp scsi_tgt ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_core sunrpc dm_mod iosf_mbi crc32_pclmul ppdev ghash_clmulni_intel aesni_intel parport_pc lrw gf128mul glue_helper parport ablk_helper virtio_balloon cryptd pcspkr joydev i2c_piix4 i2c_core ip_tables ext4 mbcache jbd2 ata_generic pata_acpi virtio_blk 8139too ata_piix crct10dif_pclmul crct10dif_common crc32c_intel libata 8139cp
[ 9072.315961]  serio_raw virtio_pci mii virtio_ring virtio floppy
[ 9072.316970] CPU: 1 PID: 7414 Comm: lfsck Kdump: loaded Tainted: G           OE  ------------   3.10.0-862.14.4.el7_lustre.x86_64 #1
[ 9072.318855] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
[ 9072.319786] task: ffff9f2669f40000 ti: ffff9f2604b1c000 task.ti: ffff9f2604b1c000
[ 9072.320985] RIP: 0010:[&amp;lt;ffffffff9055f6a9&amp;gt;]  [&amp;lt;ffffffff9055f6a9&amp;gt;] __write_lock_failed+0x9/0x20
[ 9072.322414] RSP: 0018:ffff9f2604b1fbd0  EFLAGS: 00000297
[ 9072.323282] RAX: ffff9f2642722000 RBX: ffff9f2604b1fbd8 RCX: 00000000007b1a63
[ 9072.324431] RDX: 0000000000000000 RSI: ffff9f2667af8000 RDI: ffff9f2660f3c884
[ 9072.325580] RBP: ffff9f2604b1fbd0 R08: 000000000001ba80 R09: ffffffffc0f5495d
[ 9072.326732] R10: ffff9f267fd1ba80 R11: ffffe9e601710340 R12: 0000123100000000
[ 9072.327877] R13: 00001cf600000000 R14: 00000000000e68c9 R15: 5c481b9300000001
[ 9072.329023] FS:  0000000000000000(0000) GS:ffff9f267fd00000(0000) knlGS:0000000000000000
[ 9072.330325] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 9072.331250] CR2: 00007fa602e2f000 CR3: 0000000035df6000 CR4: 00000000000606e0
[ 9072.332402] Call Trace:
[ 9072.332829]  [&amp;lt;ffffffff9091b527&amp;gt;] _raw_write_lock+0x17/0x20
[ 9072.333758]  [&amp;lt;ffffffffc0f542d6&amp;gt;] lfsck_rbtree_cleanup.isra.87+0x46/0x140 [lfsck]
[ 9072.334976]  [&amp;lt;ffffffffc0f54ac0&amp;gt;] lfsck_layout_slave_quit+0x290/0x340 [lfsck]
[ 9072.336152]  [&amp;lt;ffffffffc0f5d842&amp;gt;] lfsck_layout_slave_double_scan+0x102/0x1110 [lfsck]
[ 9072.337430]  [&amp;lt;ffffffff902d2010&amp;gt;] ? wake_up_state+0x20/0x20
[ 9072.338344]  [&amp;lt;ffffffffc0f2bd1f&amp;gt;] lfsck_double_scan+0x5f/0x210 [lfsck]
[ 9072.339438]  [&amp;lt;ffffffffc0a308c1&amp;gt;] ? lprocfs_counter_sub+0xc1/0x130 [obdclass]
[ 9072.340605]  [&amp;lt;ffffffffc0f30946&amp;gt;] lfsck_master_engine+0x4a6/0x1360 [lfsck]
[ 9072.341717]  [&amp;lt;ffffffff902d2010&amp;gt;] ? wake_up_state+0x20/0x20
[ 9072.342626]  [&amp;lt;ffffffffc0f304a0&amp;gt;] ? lfsck_master_oit_engine+0x1190/0x1190 [lfsck]
[ 9072.343838]  [&amp;lt;ffffffff902bdf21&amp;gt;] kthread+0xd1/0xe0
[ 9072.344633]  [&amp;lt;ffffffff902bde50&amp;gt;] ? insert_kthread_work+0x40/0x40
[ 9072.345627]  [&amp;lt;ffffffff909255f7&amp;gt;] ret_from_fork_nospec_begin+0x21/0x21
[ 9072.346676]  [&amp;lt;ffffffff902bde50&amp;gt;] ? insert_kthread_work+0x40/0x40
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;

&lt;p&gt;sample reports:&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sessions/a4bbd650-ac77-45a7-9d70-1d0eb7d6cfcf&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sessions/a4bbd650-ac77-45a7-9d70-1d0eb7d6cfcf&lt;/a&gt;&lt;br/&gt;
&lt;a href=&quot;https://testing.whamcloud.com/test_sessions/455af6f1-1cfe-4820-a7c2-c136ee3eceef&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sessions/455af6f1-1cfe-4820-a7c2-c136ee3eceef&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The very first time it crashed like that was in this review patch: &lt;a href=&quot;https://testing.whamcloud.com/test_sessions/b8c26f6b-4930-4aa8-937f-eedf046e3427&quot; class=&quot;external-link&quot; target=&quot;_blank&quot; rel=&quot;nofollow noopener&quot;&gt;https://testing.whamcloud.com/test_sessions/b8c26f6b-4930-4aa8-937f-eedf046e3427&lt;/a&gt; but the patch is unrelated (the failure was miscathegorized as &lt;a href=&quot;https://jira.whamcloud.com/browse/LU-11111&quot; title=&quot;crash doing LFSCK: orph_index_insert()) ASSERTION( !(obj-&amp;gt;mod_flags &amp;amp; ORPHAN_OBJ)&quot; class=&quot;issue-link&quot; data-issue-key=&quot;LU-11111&quot;&gt;&lt;del&gt;LU-11111&lt;/del&gt;&lt;/a&gt; but it does not look related)&lt;/p&gt;
</description>
                <environment></environment>
        <key id="54965">LU-11996</key>
            <summary>NMI watchdog in lfsck test 18</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="4" iconUrl="https://jira.whamcloud.com/images/icons/priorities/minor.svg">Minor</priority>
                        <status id="1" iconUrl="https://jira.whamcloud.com/images/icons/statuses/open.png" description="The issue is open and ready for the assignee to start work on it.">Open</status>
                    <statusCategory id="2" key="new" colorName="default"/>
                                    <resolution id="-1">Unresolved</resolution>
                                        <assignee username="wc-triage">WC Triage</assignee>
                                    <reporter username="green">Oleg Drokin</reporter>
                        <labels>
                    </labels>
                <created>Sun, 24 Feb 2019 05:49:30 +0000</created>
                <updated>Sun, 24 Feb 2019 05:49:30 +0000</updated>
                                            <version>Lustre 2.10.7</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>1</watches>
                                                                                    <attachments>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|i00c4v:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>9223372036854775807</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>