[LU-6409] sleeping while atomic in nodemap_destroy Created: 28/Mar/15  Updated: 23/Mar/17  Resolved: 29/Jun/15

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: None
Fix Version/s: Lustre 2.8.0

Type: Bug Priority: Major
Reporter: Oleg Drokin Assignee: WC Triage
Resolution: Fixed Votes: 0
Labels: patch

Issue Links:
Related
is related to LU-5092 nodemap: transfer idmaps between MGS,... Resolved
is related to LU-7811 BUG: sleeping function called from in... Resolved
Severity: 3
Rank (Obsolete): 9223372036854775807

 Description   

nodemap_destroy seems to be using mutexes when called under spinlock/rwlock.

Here's the relevant error message:

[ 6003.169249] Lustre: server umount lustre-OST0001 complete
[ 6007.190173] BUG: sleeping function called from invalid context at kernel/mutex.c:576
[ 6007.190482] in_atomic(): 1, irqs_disabled(): 0, pid: 20641, name: rmmod
[ 6007.190659] 1 lock held by rmmod/20641:
[ 6007.190805]  #0:  (&new_bkts[i]->hsb_lock.rw){......}, at: [<ffffffffa015151d>] cfs_hash_rw_lock+0xd/0x20 [libcfs]
[ 6007.191177] CPU: 3 PID: 20641 Comm: rmmod Tainted: GF       W  O--------------   3.10.0-debug #5
[ 6007.191480] Hardware name: Bochs Bochs, BIOS Bochs 01/01/2011
[ 6007.191648]  ffff880083aac540 000000008428f159 ffff8800967e7d50 ffffffff816ccb68
[ 6007.191954]  ffff8800967e7d68 ffffffff810a8fd9 ffff8800a8d2b0c0 ffff8800967e7de8
[ 6007.192275]  ffffffff816d0dd5 0000000000000000 ffff8800967e7de0 0000000000000246
[ 6007.192582] Call Trace:
[ 6007.192716]  [<ffffffff816ccb68>] dump_stack+0x19/0x1b
[ 6007.192879]  [<ffffffff810a8fd9>] __might_sleep+0xe9/0x110
[ 6007.193063]  [<ffffffff816d0dd5>] mutex_lock_nested+0x55/0x490
[ 6007.193281]  [<ffffffffa057cfa4>] ? nodemap_destroy+0x74/0x1f0 [ptlrpc]
[ 6007.193491]  [<ffffffffa05812ab>] nm_member_reclassify_nodemap+0x1b/0x40 [ptlrpc]
[ 6007.193799]  [<ffffffffa057cfbc>] nodemap_destroy+0x8c/0x1f0 [ptlrpc]
[ 6007.193993]  [<ffffffffa057d14d>] nodemap_putref+0x2d/0xa0 [ptlrpc]
[ 6007.194201]  [<ffffffffa057e1f5>] nodemap_hs_put_locked+0x15/0x20 [ptlrpc]
[ 6007.194392]  [<ffffffffa0151f25>] cfs_hash_bd_del_locked+0x75/0xf0 [libcfs]
[ 6007.194578]  [<ffffffffa015579c>] cfs_hash_putref+0x1cc/0x500 [libcfs]
[ 6007.194771]  [<ffffffffa057e22a>] nodemap_cleanup_all+0x2a/0x30 [ptlrpc]
[ 6007.194967]  [<ffffffffa057e23e>] nodemap_mod_exit+0xe/0x20 [ptlrpc]
[ 6007.195177]  [<ffffffffa0583bbd>] ptlrpc_exit+0x9/0x44c [ptlrpc]
[ 6007.195354]  [<ffffffff810e769b>] SyS_delete_module+0x16b/0x2d0
[ 6007.195526]  [<ffffffff816df369>] system_call_fastpath+0x16/0x1b


 Comments   
Comment by Gerrit Updater [ 30/Mar/15 ]

Kit Westneat (kit.westneat@gmail.com) uploaded a new patch: http://review.whamcloud.com/14275
Subject: LU-6409 nodemap: replace mutex with spinlock
Project: fs/lustre-release
Branch: master
Current Patch Set: 1
Commit: c334985339b345fc90128f3d9d55d0779e2dd8ca

Comment by Oleg Drokin [ 11/Apr/15 ]

Please note that with this latest patch I still get this warning:

Apr 11 00:01:22 centos6-1 kernel: BUG: sleeping function called from invalid context at /home/green/git/lustre-release/libcfs/libcfs/hash.c:1148
Apr 11 00:01:22 centos6-1 kernel: in_atomic(): 1, irqs_disabled(): 0, pid: 17878, name: rmmod
Apr 11 00:01:22 centos6-1 kernel: 1 lock held by rmmod/17878:
Apr 11 00:01:22 centos6-1 kernel: #0:  (&new_bkts[i]->hsb_lock.rw){......}, at: [<ffffffffa01c030d>] cfs_hash_rw_lock+0xd/0x20 [libcfs]
Apr 11 00:01:22 centos6-1 kernel: CPU: 7 PID: 17878 Comm: rmmod Tainted: GF       W  O--------------   3.10.0-debug #5
Apr 11 00:01:22 centos6-1 kernel: Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011
Apr 11 00:01:22 centos6-1 kernel: ffff880077774740 0000000096882851 ffff880091d93d80 ffffffff816ccb68
Apr 11 00:01:22 centos6-1 kernel: ffff880091d93d98 ffffffff810a8fd9 0000000000000003 ffff880091d93de8
Apr 11 00:01:22 centos6-1 kernel: ffffffffa01c45f3 0000000096882851 ffff88007aeebcc0 ffff880000000008
Apr 11 00:01:22 centos6-1 kernel: Call Trace:
Apr 11 00:01:22 centos6-1 kernel: [<ffffffff816ccb68>] dump_stack+0x19/0x1b
Apr 11 00:01:22 centos6-1 kernel: [<ffffffff810a8fd9>] __might_sleep+0xe9/0x110
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa01c45f3>] cfs_hash_putref+0x233/0x500 [libcfs]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa05ac15f>] nm_member_delete_hash+0x2f/0x40 [ptlrpc]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa05a8014>] nodemap_destroy+0xa4/0x1f0 [ptlrpc]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa05a818d>] nodemap_putref+0x2d/0xa0 [ptlrpc]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa05a9235>] nodemap_hs_put_locked+0x15/0x20 [ptlrpc]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa01c0d15>] cfs_hash_bd_del_locked+0x75/0xf0 [libcfs]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa01c458c>] cfs_hash_putref+0x1cc/0x500 [libcfs]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa05a926a>] nodemap_cleanup_all+0x2a/0x30 [ptlrpc]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa05a927e>] nodemap_mod_exit+0xe/0x20 [ptlrpc]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffffa05aebc2>] ptlrpc_exit+0x9/0x447 [ptlrpc]
Apr 11 00:01:22 centos6-1 kernel: [<ffffffff810e769b>] SyS_delete_module+0x16b/0x2d0
Apr 11 00:01:22 centos6-1 kernel: [<ffffffff816df369>] system_call_fastpath+0x16/0x1b

This is because cfs_hash_putref might call cfs_hash_destroy if that was the last reference relesed and that might sleep.

Comment by Oleg Drokin [ 11/Apr/15 ]

Aha! so this is also the reason for quite a bunch of spinlock bugs I was having where spinlock debugging code was complaining that a lock was taken on one cpu, but released on another, I guess:

Apr 11 00:30:38 centos6-7 kernel: BUG: sleeping function called from invalid context at /home/green/git/lustre-release/libcfs/libcfs/hash.c:1148
Apr 11 00:30:38 centos6-7 kernel: in_atomic(): 1, irqs_disabled(): 0, pid: 3620, name: rmmod
Apr 11 00:30:38 centos6-7 kernel: 1 lock held by rmmod/3620:
Apr 11 00:30:38 centos6-7 kernel: #0:  (&new_bkts[i]->hsb_lock.rw){......}, at: [<ffffffffa014e30d>] cfs_hash_rw_lock+0xd/0x20 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: CPU: 2 PID: 3620 Comm: rmmod Tainted: GF       W  O--------------   3.10.0-debug #5
Apr 11 00:30:38 centos6-7 kernel: Hardware name: Bochs Bochs, BIOS Bochs 01/01/2011
Apr 11 00:30:38 centos6-7 kernel: ffff88003ae74780 00000000d7e1d1ee ffff88008473bd80 ffffffff816ccb68
Apr 11 00:30:38 centos6-7 kernel: ffff88008473bd98 ffffffff810a8fd9 0000000000000003 ffff88008473bde8
Apr 11 00:30:38 centos6-7 kernel: ffffffffa01525f3 00000000d7e1d1ee ffff88008d62bc00 ffff880000000008
Apr 11 00:30:38 centos6-7 kernel: Call Trace:
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816ccb68>] dump_stack+0x19/0x1b
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff810a8fd9>] __might_sleep+0xe9/0x110
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa01525f3>] cfs_hash_putref+0x233/0x500 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05cb15f>] nm_member_delete_hash+0x2f/0x40 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c7014>] nodemap_destroy+0xa4/0x1f0 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c718d>] nodemap_putref+0x2d/0xa0 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c8235>] nodemap_hs_put_locked+0x15/0x20 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa014ed15>] cfs_hash_bd_del_locked+0x75/0xf0 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa015258c>] cfs_hash_putref+0x1cc/0x500 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c826a>] nodemap_cleanup_all+0x2a/0x30 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c827e>] nodemap_mod_exit+0xe/0x20 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05cdbc2>] ptlrpc_exit+0x9/0x447 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff810e769b>] SyS_delete_module+0x16b/0x2d0
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816df369>] system_call_fastpath+0x16/0x1b
Apr 11 00:30:38 centos6-7 kernel: BUG: scheduling while atomic: rmmod/3620/0x10000002
Apr 11 00:30:38 centos6-7 kernel: 1 lock held by rmmod/3620:
Apr 11 00:30:38 centos6-7 kernel: #0:  (&new_bkts[i]->hsb_lock.rw){......}, at: [<ffffffffa014e30d>] cfs_hash_rw_lock+0xd/0x20 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: Modules linked in: ptlrpc(OF-) obdclass(OF) ksocklnd(OF) lnet(OF) libcfs(OF) loop(F) jbd2(F) sha512_generic(F) mbcache(F) rpcsec_gss_krb5(F) ata_generic(F) pata_acpi(F) ttm(F) drm_kms_helper(F) drm(F) i2c_piix4(F) ata_piix(F) virtio_blk(F) i2c_core(F) virtio_balloon(F) virtio_console(F) floppy(F) libata(F) serio_raw(F) pcspkr(F) [last unloaded: lquota]
Apr 11 00:30:38 centos6-7 kernel: CPU: 2 PID: 3620 Comm: rmmod Tainted: GF       W  O--------------   3.10.0-debug #5
Apr 11 00:30:38 centos6-7 kernel: Hardware name: Bochs Bochs, BIOS Bochs 01/01/2011
Apr 11 00:30:38 centos6-7 kernel: ffff88003ae74780 00000000d7e1d1ee ffff88008473bcf8 ffffffff816ccb68
Apr 11 00:30:38 centos6-7 kernel: ffff88008473bd10 ffffffff816c827a ffff88008473bfd8 ffff88008473bd70
Apr 11 00:30:38 centos6-7 kernel: ffffffff816d3d1c ffff88003ae74780 ffff88008473bfd8 ffff88008473bfd8
Apr 11 00:30:38 centos6-7 kernel: Call Trace:
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816ccb68>] dump_stack+0x19/0x1b
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816c827a>] __schedule_bug+0x59/0x68
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816d3d1c>] __schedule+0x75c/0x810
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff810acab6>] __cond_resched+0x26/0x30
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816d41ea>] _cond_resched+0x3a/0x50
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa01525f8>] cfs_hash_putref+0x238/0x500 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05cb15f>] nm_member_delete_hash+0x2f/0x40 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c7014>] nodemap_destroy+0xa4/0x1f0 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c718d>] nodemap_putref+0x2d/0xa0 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c8235>] nodemap_hs_put_locked+0x15/0x20 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa014ed15>] cfs_hash_bd_del_locked+0x75/0xf0 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa015258c>] cfs_hash_putref+0x1cc/0x500 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c826a>] nodemap_cleanup_all+0x2a/0x30 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c827e>] nodemap_mod_exit+0xe/0x20 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05cdbc2>] ptlrpc_exit+0x9/0x447 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff810e769b>] SyS_delete_module+0x16b/0x2d0
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816df369>] system_call_fastpath+0x16/0x1b
Apr 11 00:30:38 centos6-7 kernel: BUG: rwlock wrong CPU on CPU#3, rmmod/3620, ffff88005ca55dc0
Apr 11 00:30:38 centos6-7 kernel: CPU: 3 PID: 3620 Comm: rmmod Tainted: GF       W  O--------------   3.10.0-debug #5
Apr 11 00:30:38 centos6-7 kernel: Hardware name: Bochs Bochs, BIOS Bochs 01/01/2011
Apr 11 00:30:38 centos6-7 kernel: ffffffff81a971b4 00000000d7e1d1ee ffff88008473be10 ffffffff816ccb68
Apr 11 00:30:38 centos6-7 kernel: ffff88008473be30 ffffffff816ccc6f ffff88005ca55dc0 ffff88005ca57e00
Apr 11 00:30:38 centos6-7 kernel: ffff88008473be48 ffffffff813707c0 ffff88005ca55dc0 ffff88008473be60
Apr 11 00:30:38 centos6-7 kernel: Call Trace:
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816ccb68>] dump_stack+0x19/0x1b
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816ccc6f>] rwlock_bug+0x4e/0x53
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff813707c0>] do_raw_write_unlock+0x60/0x80
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816d6107>] _raw_write_unlock+0x27/0x40
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa014e2ed>] cfs_hash_rw_unlock+0xd/0x20 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa01525e0>] cfs_hash_putref+0x220/0x500 [libcfs]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c826a>] nodemap_cleanup_all+0x2a/0x30 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05c827e>] nodemap_mod_exit+0xe/0x20 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffffa05cdbc2>] ptlrpc_exit+0x9/0x447 [ptlrpc]
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff810e769b>] SyS_delete_module+0x16b/0x2d0
Apr 11 00:30:38 centos6-7 kernel: [<ffffffff816df369>] system_call_fastpath+0x16/0x1b
Comment by James A Simmons [ 05/May/15 ]

I rebased the patch from Kit. Oh I see bugs remain.

Comment by Kit Westneat [ 06/May/15 ]

I realized that the patch I submitted was actually making things worse, as you aren't allowed to call cfs_hash_for_each while holding a spinlock. I instead modified nodemap_cleanup_all to convert the hash to a list and then iterate the list, calling nodemap_putref on each nodemap. Does this sound reasonable?

Comment by Oleg Drokin [ 03/Jun/15 ]

Hm, so I have several patches from you:
93ccc4a LU-5092 nodemap: add structure to hold nodemap config
e4d2197 LU-5092 nodemap: convert member hash to a list
3f0bb87 LU-6409 nodemap: don't destroy nodemaps in cfs_hash_for_each

And this is not enough.

I still get this in the logs:

leeping function called from invalid context at kernel/mutex.c:576
May 30 22:19:09 centos6-0 kernel: in_atomic(): 1, irqs_disabled(): 0, pid: 9196,
 name: mdt00_002
May 30 22:19:09 centos6-0 kernel: 2 locks held by mdt00_002/9196:
May 30 22:19:09 centos6-0 kernel: #0:  (active_config_lock){......}, at: [<fffff
fffa05cb147>] nodemap_lock_active_ranges+0x17/0x30 [ptlrpc]
May 30 22:19:09 centos6-0 kernel: #1:  (&config->nmc_range_tree_lock){......}, a
t: [<ffffffffa05cb157>] nodemap_lock_active_ranges+0x27/0x30 [ptlrpc] PID: 9196 Comm: mdt00_002 Tainted: GF       W  O--------------   3.10.0-debug #
re name: Red Hat KVM, BIOS 0.5.1 01/01/2011
0019366680 00000000b6a08141 ffff88003ac9b9a0 ffffffff816ccb68
003ac9b9b8 ffffffff810a8fd9 ffff8800aa50
0e00 ffff88003ac9ba40
ff816d0dd5 0000000000000000 0000000000000002 0000000000000000
race:
ffff816ccb68>] dump_stack+0x19/0x1b
ffff810a8fd9>] __might_sleep+0xe9/0x110
ffff816d0dd5>] mutex_lock_nested+0x55/0x490
ffffa05cb157>] ? nodemap_lock_active_ranges+0x27/0x30 [ptlrpc]
ffffa05cfbe8>] nm_member_add+0x78/0x180 [ptlrpc]
ffffa05cc80d>] nodemap_add_member+0x2d/0x50 [ptlrpc]
ffffa0b147f3>] mdt_obd_connect+0x3b3/0x760 [mdt]
ffffa01c4a87>] ? libcfs_debug_msg+0x57/0x80 [libcfs]
ffffa051ded3>] target_handle_connect+0x12c3/0x2be0 [ptlrpc]
ffff81363263>] ? number.isra.2+0x323/0x360
ffffa02dbcbd>] ? keys_fill+0x1d/0x130 [obdclass]
ffffa02dbcd2>] ? keys_fill+0x32/0x130 [obdclass]
ffffa05b445f>] tgt_request_handle+0x3ef/0x1070 [ptlrpc]
ffffa0565c01>] ptlrpc_server_handle_request+0x231/0xab0 [ptlrpc]
ffffa0563768>] ? ptlrpc_wait_event+0xb8/0x370 [ptlrpc]
ffffa0569b70>] ptlrpc_main+0xae0/0x1ee0 [ptlrpc]
ffffa0569090>] ? ptlrpc_register_service+0xf30/0xf30 [ptlrpc]
ffff8109c00a>] kthread+0xea/0xf0
ffff8109bf20>] ? kthread_create_on_node+0x140/0x140
ffff816df2bc>] ret_from_fork+0x7c/0xb0
ffff8109bf20>] ? kthread_create_on_node+0x140/0x140

seemingly from every target.

Comment by Kit Westneat [ 03/Jun/15 ]

Ah ok, I see what's going on, I'll update the patches I have in gerrit.

Can you describe what tests/config options you are running to get these warnings? I am not getting them on my build and I don't want to waste your time if I can catch them before I submit.

Comment by Oleg Drokin [ 04/Jun/15 ]

You need to use rhel7 for this, there's a new kernel option you need to enable:

CONFIG_DEBUG_ATOMIC_SLEEP=y

Turn this on and run something like sanity, you'll get a bunch of warnings. Also apply patch from LU-6447 (I know it's imperfect) to quiet down unrelated warnings.

Comment by Kit Westneat [ 08/Jun/15 ]

I finally got the debug kernel working and reproduced the warning you saw. Should I report any other non-nodemap warnings I find? There is a lock inversion warning with s_dquot.dqio_mutex and i_data_sem that I couldn't find any Jira tickets about, but I'm only getting it during mkfs so far.

Comment by Oleg Drokin [ 09/Jun/15 ]

Sure, feel free to report the other ones you see too.

Comment by Kit Westneat [ 09/Jun/15 ]

Ok, I reported the potential lock inversion as LU-6700.

I updated the nodemap patches and no longer get the warning messages.

Comment by Oleg Drokin [ 29/Jun/15 ]

I tried the three patches from LU-5092plus the one patch from this ticket and hit a deadlock when running sanity 133g:

[15921.648643] LustreError: 7937:0:(lproc_fid.c:182:lprocfs_server_fid_width_seq_write()) Skipped 17 previous similar messages
[16080.884140] INFO: task badarea_io:7981 blocked for more than 120 seconds.
[16080.884666]       Tainted: P           ---------------    2.6.32-rhe6.6-debug #1
[16080.885381] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[16080.886109] badarea_io    D 0000000000000002  2792  7981   7003 0x00000000
[16080.886582]  ffff880046b5fd78 0000000000000086 0000000000000000 ffffffff8114e4a0
[16080.887360]  0000000000000000 ffffffff8104e658 00000e7b1ce2fe1b ffff880046b5fd58
[16080.888121]  ffffffffa1395a20 00000001003b97ab ffff88005f2708f8 ffff880046b5ffd8
[16080.888904] Call Trace:
[16080.889234]  [<ffffffff8114e4a0>] ? handle_pte_fault+0x90/0xa50
[16080.889663]  [<ffffffff8104e658>] ? __change_page_attr_set_clr+0x808/0xcc0
[16080.890154]  [<ffffffffa1395a20>] ? lprocfs_single_release+0x0/0x10 [obdclass]
[16080.890876]  [<ffffffff8151ff50>] __mutex_lock_slowpath+0x120/0x2e0
[16080.891364]  [<ffffffffa166d8d0>] ? nodemap_del_nodemap_write+0x0/0xc0 [ptlrpc]
[16080.892098]  [<ffffffff81520141>] mutex_lock+0x31/0x50
[16080.892598]  [<ffffffffa16699bc>] nodemap_del+0x3c/0x1c0 [ptlrpc]
[16080.893118]  [<ffffffffa166d8d0>] ? nodemap_del_nodemap_write+0x0/0xc0 [ptlrpc]
[16080.896192]  [<ffffffffa166d975>] nodemap_del_nodemap_write+0xa5/0xc0 [ptlrpc]
[16080.896810]  [<ffffffff811fb8c5>] proc_reg_write+0x85/0xc0
[16080.897094]  [<ffffffff8118f4a8>] vfs_write+0xb8/0x1a0
[16080.897388]  [<ffffffff8118fde1>] sys_write+0x51/0x90
[16080.897719]  [<ffffffff8100b0b2>] system_call_fastpath+0x16/0x1b
Comment by Oleg Drokin [ 29/Jun/15 ]

ah, I see it's likely a bug in http://review.whamcloud.com/#/c/11813/8

Comment by Gerrit Updater [ 29/Jun/15 ]

Oleg Drokin (oleg.drokin@intel.com) merged in patch http://review.whamcloud.com/14275/
Subject: LU-6409 nodemap: don't destroy nodemaps in cfs_hash_for_each
Project: fs/lustre-release
Branch: master
Current Patch Set:
Commit: 6299d78713ef2b6f69f6dd046256aa91a85147a8

Comment by Peter Jones [ 29/Jun/15 ]

Landed for 2.8

Generated at Sat Feb 10 01:59:58 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.