<!-- 
RSS generated by JIRA (9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c) at Sat Feb 10 01:30:09 UTC 2024

It is possible to restrict the fields that are returned in this document by specifying the 'field' parameter in your request.
For example, to request only the issue key and summary append 'field=key&field=summary' to the URL of your request.
-->
<rss version="0.92" >
<channel>
    <title>Whamcloud Community JIRA</title>
    <link>https://jira.whamcloud.com</link>
    <description>This file is an XML representation of an issue</description>
    <language>en-us</language>    <build-info>
        <version>9.4.14</version>
        <build-number>940014</build-number>
        <build-date>05-12-2023</build-date>
    </build-info>


<item>
            <title>[LU-3007] Umount OST caused a oops in ptlrpc_free_rqbd</title>
                <link>https://jira.whamcloud.com/browse/LU-3007</link>
                <project id="10000" key="LU">Lustre</project>
                    <description>&lt;p&gt;After finishing a test run with the 2.4 file system when we went to umount the entire file system one of the OSS expected a oops when it went to umount.&lt;br/&gt;
The oops is:&lt;/p&gt;

&lt;p&gt;Mar  9 20:11:12 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;112998.231220&amp;#93;&lt;/span&gt; Lustre: server umount routed1-OST00b1 complete&lt;br/&gt;
Mar  9 20:11:13 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;112999.326796&amp;#93;&lt;/span&gt; LustreError: 20552:0:(obd_mount.c:2115:server_put_super()) routed1-OST01d1: failed to disconnect lwp. (rc=-110)&lt;br/&gt;
Mar  9 20:11:13 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;112999.359171&amp;#93;&lt;/span&gt; LustreError: 20552:0:(obd_mount.c:2115:server_put_super()) Skipped 1 previous similar message&lt;br/&gt;
Mar  9 20:11:13 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;112999.370669&amp;#93;&lt;/span&gt; Lustre: server umount routed1-OST0171 complete&lt;br/&gt;
Mar  9 20:11:13 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;112999.370672&amp;#93;&lt;/span&gt; Lustre: Skipped 1 previous similar message&lt;br/&gt;
Mar  9 20:11:13 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;112999.429394&amp;#93;&lt;/span&gt; Lustre: Failing over routed1-OST01d1&lt;br/&gt;
Mar  9 20:11:13 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;112999.448337&amp;#93;&lt;/span&gt; Lustre: Skipped 1 previous similar message&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.073001&amp;#93;&lt;/span&gt; BUG: soft lockup - CPU#0 stuck for 67s! &lt;span class=&quot;error&quot;&gt;&amp;#91;obd_zombid:5676&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.073001&amp;#93;&lt;/span&gt; Modules linked in: osp(U) ofd(U) ost(U) mgc(U) fsfilt_ldiskfs(U) osd_ldiskfs(U) ldiskfs(U) mdd(U) lustre(U) lov(U) osc(U) mdc(U) lquota(U) fid(U) fld(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) libcfs(U) ext4 mbcache jbd2 sha512_generic sha256_generic ipmi_devintf ipmi_si ipmi_msghandler autofs4 dm_round_robin scsi_wait_scan sd_mod crc_t10dif ib_srp scsi_transport_srp scsi_tgt nf_conntrack_netbios_ns nf_conntrack_broadcast ipt_REJECT xt_comment nf_conntrack_ipv4 nf_defrag_ipv4 xt_state nf_conntrack iptable_filter ip_tables ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ib_sa dm_mirror dm_region_hash dm_log dm_multipath dm_mod sg sr_mod cdrom radeon ttm drm_kms_helper drm i2c_algo_bit i2c_core ib_mthca ib_mad ib_core shpchp microcode dcdbas serio_raw ata_generic pata_acpi ata_piix iTCO_wdt iTCO_vendor_support i5000_edac edac_core i5k_amb ioatdma dca ipv6 nfs lockd fscache nfs_acl auth_rpcgss sunrpc bnx2 &lt;span class=&quot;error&quot;&gt;&amp;#91;last unloaded: libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; Pid: 5676, comm: obd_zombid Not tainted 2.6.32-279.22.1.el6.head.x86_64 #1 Dell Inc. PowerEdge 1950/0TT740&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; RIP: 0010:&lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81147e76&amp;gt;&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81147e76&amp;gt;&amp;#93;&lt;/span&gt; remove_vm_area+0x86/0xa0&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; RSP: 0018:ffff880419427cb0  EFLAGS: 00000287&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; RAX: ffff8800a66fb540 RBX: ffff880419427cd0 RCX: ffff88011e073440&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; RDX: ffff88011e0733c0 RSI: 0000000000000001 RDI: ffffffff81abea30&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; RBP: ffffffff8100bb8e R08: 0000000000000000 R09: ffffc900cc6ee000&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; R10: ffff880028402500 R11: ffff88008acf3800 R12: ffff880419427c80&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; R13: 0000000000000297 R14: ffff8800000206c0 R15: ffffea0002410270&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; FS:  00007f4834066700(0000) GS:ffff880028200000(0000) knlGS:0000000000000000&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; CR2: 00000000004050b9 CR3: 0000000001a85000 CR4: 00000000000006f0&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; Process obd_zombid (pid: 5676, threadinfo ffff880419426000, task ffff8803e2329540)&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; Stack:&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  ffff88001b6df840 ffff8800a66fb540 ffff88008ad17850 ffffc900cc6ee000&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; &amp;lt;d&amp;gt; ffff880419427d00 ffffffff81147f4e 000000000000005a ffff88006f537800&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; &amp;lt;d&amp;gt; ffff8803d5e93400 ffff8803d5e93430 ffff880419427d10 ffffffff811480ba&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; Call Trace:&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81147f4e&amp;gt;&amp;#93;&lt;/span&gt; ? __vunmap+0x2e/0x120&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811480ba&amp;gt;&amp;#93;&lt;/span&gt; ? vfree+0x2a/0x40&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06e3b0e&amp;gt;&amp;#93;&lt;/span&gt; ? cfs_free_large+0xe/0x10 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0e307c2&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_free_rqbd+0xc2/0x3b0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0e33c45&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_unregister_service+0x5d5/0x10b0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81052223&amp;gt;&amp;#93;&lt;/span&gt; ? __wake_up+0x53/0x70&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa1253292&amp;gt;&amp;#93;&lt;/span&gt; ? ost_cleanup+0x52/0x1a0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ost&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07f1d92&amp;gt;&amp;#93;&lt;/span&gt; ? class_decref+0x212/0x580 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06f32e1&amp;gt;&amp;#93;&lt;/span&gt; ? libcfs_debug_msg+0x41/0x50 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf549&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_cull+0x309/0x5d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf8d5&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0xc5/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8105fa40&amp;gt;&amp;#93;&lt;/span&gt; ? default_wake_function+0x0/0x20&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf810&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0x0/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c0ca&amp;gt;&amp;#93;&lt;/span&gt; ? child_rip+0xa/0x20&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf810&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0x0/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf810&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0x0/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c0c0&amp;gt;&amp;#93;&lt;/span&gt; ? child_rip+0x0/0x20&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; Code: 30 ea ab 81 48 89 45 e8 e8 88 4d 3a 00 48 8b 15 19 73 e6 00 48 c7 c1 78 f1 fa 81 48 8b 45 e8 48 39 d3 74 0c 90 48 89 d1 48 8b 12 &amp;lt;48&amp;gt; 39 d3 75 f5 48 8b 13 48 89 11 f0 81 05 a4 6b 97 00 00 00 00&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt; Call Trace:&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81147f4e&amp;gt;&amp;#93;&lt;/span&gt; ? __vunmap+0x2e/0x120&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff811480ba&amp;gt;&amp;#93;&lt;/span&gt; ? vfree+0x2a/0x40&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06e3b0e&amp;gt;&amp;#93;&lt;/span&gt; ? cfs_free_large+0xe/0x10 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0e307c2&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_free_rqbd+0xc2/0x3b0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa0e33c45&amp;gt;&amp;#93;&lt;/span&gt; ? ptlrpc_unregister_service+0x5d5/0x10b0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ptlrpc&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff81052223&amp;gt;&amp;#93;&lt;/span&gt; ? __wake_up+0x53/0x70&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa1253292&amp;gt;&amp;#93;&lt;/span&gt; ? ost_cleanup+0x52/0x1a0 &lt;span class=&quot;error&quot;&gt;&amp;#91;ost&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07f1d92&amp;gt;&amp;#93;&lt;/span&gt; ? class_decref+0x212/0x580 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa06f32e1&amp;gt;&amp;#93;&lt;/span&gt; ? libcfs_debug_msg+0x41/0x50 &lt;span class=&quot;error&quot;&gt;&amp;#91;libcfs&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf549&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_cull+0x309/0x5d0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf8d5&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0xc5/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8105fa40&amp;gt;&amp;#93;&lt;/span&gt; ? default_wake_function+0x0/0x20&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf810&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0x0/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c0ca&amp;gt;&amp;#93;&lt;/span&gt; ? child_rip+0xa/0x20&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf810&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0x0/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffffa07cf810&amp;gt;&amp;#93;&lt;/span&gt; ? obd_zombie_impexp_thread+0x0/0x1c0 &lt;span class=&quot;error&quot;&gt;&amp;#91;obdclass&amp;#93;&lt;/span&gt;&lt;br/&gt;
Mar  9 20:12:30 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113076.102203&amp;#93;&lt;/span&gt;  &lt;span class=&quot;error&quot;&gt;&amp;#91;&amp;lt;ffffffff8100c0c0&amp;gt;&amp;#93;&lt;/span&gt; ? child_rip+0x0/0x20&lt;br/&gt;
Mar  9 20:13:35 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113140.803317&amp;#93;&lt;/span&gt; Lustre: server umount routed1-OST0231 complete&lt;br/&gt;
Mar  9 20:13:35 widow-oss11c2 kernel: &lt;span class=&quot;error&quot;&gt;&amp;#91;113140.819063&amp;#93;&lt;/span&gt; Lustre: Skipped 1 previous similar message&lt;/p&gt;</description>
                <environment>OSS server runing Lustre 2.3.61</environment>
        <key id="18036">LU-3007</key>
            <summary>Umount OST caused a oops in ptlrpc_free_rqbd</summary>
                <type id="1" iconUrl="https://jira.whamcloud.com/secure/viewavatar?size=xsmall&amp;avatarId=11303&amp;avatarType=issuetype">Bug</type>
                                            <priority id="3" iconUrl="https://jira.whamcloud.com/images/icons/priorities/major.svg">Major</priority>
                        <status id="5" iconUrl="https://jira.whamcloud.com/images/icons/statuses/resolved.png" description="A resolution has been taken, and it is awaiting verification by reporter. From here issues are either reopened, or are closed.">Resolved</status>
                    <statusCategory id="3" key="done" colorName="success"/>
                                    <resolution id="5">Cannot Reproduce</resolution>
                                        <assignee username="bfaccini">Bruno Faccini</assignee>
                                    <reporter username="simmonsja">James A Simmons</reporter>
                        <labels>
                    </labels>
                <created>Thu, 21 Mar 2013 14:28:40 +0000</created>
                <updated>Wed, 17 Jul 2013 14:48:08 +0000</updated>
                            <resolved>Wed, 17 Jul 2013 14:48:08 +0000</resolved>
                                    <version>Lustre 2.4.0</version>
                                                        <due></due>
                            <votes>0</votes>
                                    <watches>5</watches>
                                                                            <comments>
                            <comment id="54593" author="bfaccini" created="Thu, 21 Mar 2013 18:59:20 +0000"  >&lt;p&gt;James,&lt;br/&gt;
It is only a soft-lockup trace/stack, was it a one-shoot or were there multiple ??&lt;br/&gt;
How many OSTs to be stopped ??&lt;br/&gt;
A full log may be welcome.&lt;/p&gt;</comment>
                            <comment id="54594" author="green" created="Thu, 21 Mar 2013 19:02:02 +0000"  >&lt;p&gt;so it seems the unmount actually succeeded from further messages.&lt;br/&gt;
the only issue is we are probably taking too long to unregister some buffers when there are many of them, so should not be holding any locks over that?&lt;/p&gt;</comment>
                            <comment id="54651" author="simmonsja" created="Fri, 22 Mar 2013 13:16:40 +0000"  >&lt;p&gt;Bruno the soft-lockup only happened once on single OSS when unmounting this particular OST. The total OSTs were 376 if I remember correctly. The other were fine.&lt;/p&gt;</comment>
                            <comment id="54678" author="bfaccini" created="Fri, 22 Mar 2013 17:52:17 +0000"  >&lt;p&gt;Humm, Having a look to some assembly of the remove_vm_area() routine it seems that the soft-lockup occured during the for loop to find the matching vm_area :&lt;/p&gt;
&lt;div class=&quot;preformatted panel&quot; style=&quot;border-width: 1px;&quot;&gt;&lt;div class=&quot;preformattedContent panelContent&quot;&gt;
&lt;pre&gt;struct vm_struct *remove_vm_area(const void *addr)
{
        struct vmap_area *va;

        va = find_vmap_area((unsigned long)addr);
        if (va &amp;amp;&amp;amp; va-&amp;gt;flags &amp;amp; VM_VM_AREA) {
                struct vm_struct *vm = va-&amp;gt;private;

                if (!(vm-&amp;gt;flags &amp;amp; VM_UNLIST)) {
                        struct vm_struct *tmp, **p;
                        /*
                         * remove from list and disallow access to
                         * this vm_struct before unmap. (address range
                         * confliction is maintained by vmap.)
                         */
                        write_lock(&amp;amp;vmlist_lock);
                        for (p = &amp;amp;vmlist; (tmp = *p) != vm; p = &amp;amp;tmp-&amp;gt;next)  &amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;&amp;lt;
                                ;
                        *p = tmp-&amp;gt;next;
                        write_unlock(&amp;amp;vmlist_lock);
                }

                vmap_debug_free_range(va-&amp;gt;va_start, va-&amp;gt;va_end);
                free_unmap_vmap_area(va);
                vm-&amp;gt;size -= PAGE_SIZE;

                return vm;
        }
        return NULL;
}
&lt;/pre&gt;
&lt;/div&gt;&lt;/div&gt;
&lt;p&gt;Thus it is kind of strange that this thread was the only &quot;impacted&quot;!!...&lt;br/&gt;
Also is it reasonable to think that it spent more than 67s in this loop and finally exited from it ?&lt;br/&gt;
May be it spent most of this time trying to write_lock(&amp;amp;vmlist_lock) before ?&lt;/p&gt;</comment>
                            <comment id="55003" author="simmonsja" created="Thu, 28 Mar 2013 14:07:49 +0000"  >&lt;p&gt;I found out from our admin it was more than OST umounting that had a problem. The logs you saw was for one that completed. The rest never completed and after 30 minutes the admin rebooted the machine instead.&lt;/p&gt;</comment>
                            <comment id="55010" author="green" created="Thu, 28 Mar 2013 14:48:35 +0000"  >&lt;p&gt;can we get logs from one of the nodes that did not complete unmounting, please?&lt;/p&gt;</comment>
                            <comment id="55017" author="simmonsja" created="Thu, 28 Mar 2013 16:42:11 +0000"  >&lt;p&gt;Here is the entire log of the day for this OSS that crashed when unmounting.&lt;/p&gt;</comment>
                            <comment id="55601" author="bfaccini" created="Fri, 5 Apr 2013 13:20:48 +0000"  >&lt;p&gt;Thank&apos;s for the log, but next time this kind of hung situation requires a hard reboot/reset, is it possible that you take a crash-dump instead ??&lt;/p&gt;</comment>
                            <comment id="56296" author="bfaccini" created="Mon, 15 Apr 2013 07:24:18 +0000"  >&lt;p&gt;James,&lt;br/&gt;
There is nothing more of interest in the log, thus we need at least a full back trace (Alt+SysRq+T, or echo t &amp;gt; /proc/sysrq-trigger) to be taken at the time the soft-lockup is reported, and a forced crash-dump would be a must.&lt;/p&gt;</comment>
                            <comment id="56806" author="simmonsja" created="Tue, 23 Apr 2013 13:29:05 +0000"  >&lt;p&gt;We have changed are policy now for test shots. For the last test shot we did not encounter this problem. If we don&apos;t encounter this at the next test shot I say we can close the ticket.&lt;/p&gt;</comment>
                            <comment id="62466" author="bfaccini" created="Wed, 17 Jul 2013 09:16:42 +0000"  >&lt;p&gt;Hello James, &lt;br/&gt;
How went the further test shots finally ??&lt;br/&gt;
If ok, can we close this ticket? If not, have you been able to take a crash-dump upon a new occurrence?&lt;/p&gt;</comment>
                            <comment id="62476" author="simmonsja" created="Wed, 17 Jul 2013 13:23:20 +0000"  >&lt;p&gt;Yes you can close this ticket. If we encounter it we can reopen but we haven&apos;t seen this problem lately.&lt;/p&gt;</comment>
                            <comment id="62484" author="bfaccini" created="Wed, 17 Jul 2013 14:48:08 +0000"  >&lt;p&gt;Ok thank&apos;s James.&lt;/p&gt;</comment>
                    </comments>
                    <attachments>
                            <attachment id="12439" name="lu-3007.log" size="45734" author="simmonsja" created="Thu, 28 Mar 2013 16:42:11 +0000"/>
                    </attachments>
                <subtasks>
                    </subtasks>
                <customfields>
                                                                                                                                                                                            <customfield id="customfield_10890" key="com.atlassian.jira.plugins.jira-development-integration-plugin:devsummary">
                        <customfieldname>Development</customfieldname>
                        <customfieldvalues>
                            
                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        <customfield id="customfield_10390" key="com.pyxis.greenhopper.jira:gh-lexo-rank">
                        <customfieldname>Rank</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>1|hzvlv3:</customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                <customfield id="customfield_10090" key="com.pyxis.greenhopper.jira:gh-global-rank">
                        <customfieldname>Rank (Obsolete)</customfieldname>
                        <customfieldvalues>
                            <customfieldvalue>7320</customfieldvalue>
                        </customfieldvalues>
                    </customfield>
                                                                                            <customfield id="customfield_10060" key="com.atlassian.jira.plugin.system.customfieldtypes:select">
                        <customfieldname>Severity</customfieldname>
                        <customfieldvalues>
                                <customfieldvalue key="10022"><![CDATA[3]]></customfieldvalue>

                        </customfieldvalues>
                    </customfield>
                                                                                                                                                                                                                                                                                                                                                        </customfields>
    </item>
</channel>
</rss>