[LU-6287] sanity test 17n ASSERTION( atomic_read(&lu->ld_ref) == 0 ) failed Created: 25/Feb/15  Updated: 09/Sep/16  Resolved: 31/Mar/15

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: Lustre 2.8.0
Fix Version/s: Lustre 2.8.0

Type: Bug Priority: Minor
Reporter: Maloo Assignee: Di Wang
Resolution: Fixed Votes: 0
Labels: dne2

Issue Links:
Related
is related to LU-3534 async update cross-MDTs Resolved
Severity: 3
Rank (Obsolete): 17619

 Description   

This issue was created by maloo for wangdi <di.wang@intel.com>

This issue relates to the following test suite run: https://testing.hpdd.intel.com/test_sets/8443c2d2-bd2c-11e4-a946-5254006e85c2.

The sub-test test_17n failed with the following error:

test failed to respond and timed out
13:07:23:Lustre: DEBUG MARKER: == sanity test 17n: run e2fsck against master/slave MDT which contains remote dir == 13:07:13 (1424869633)
13:07:23:Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n version
13:07:23:Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n version
13:07:23:Lustre: DEBUG MARKER: grep -c /mnt/mds1' ' /proc/mounts
13:07:23:Lustre: DEBUG MARKER: umount -d /mnt/mds1
13:07:23:Lustre: Failing over lustre-MDT0000
13:07:23:LustreError: 2880:0:(lod_dev.c:1311:lod_device_free()) ASSERTION( atomic_read(&lu->ld_ref) == 0 ) failed: lu is ffff88005bee2000
13:07:23:LustreError: 2880:0:(lod_dev.c:1311:lod_device_free()) LBUG
13:07:23:Pid: 2880, comm: obd_zombid
13:07:23:
13:07:23:Call Trace:
13:07:23: [<ffffffffa0491895>] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
13:07:23: [<ffffffffa0491e97>] lbug_with_loc+0x47/0xb0 [libcfs]
13:07:23: [<ffffffffa0fb524e>] lod_device_free+0x1ee/0x220 [lod]
13:07:23: [<ffffffffa05dd95d>] class_decref+0x3ed/0x4d0 [obdclass]
13:07:23: [<ffffffffa05c69e1>] obd_zombie_impexp_cull+0x611/0x970 [obdclass]
13:07:23: [<ffffffffa05c6da5>] obd_zombie_impexp_thread+0x65/0x190 [obdclass]
13:07:23: [<ffffffff81064b90>] ? default_wake_function+0x0/0x20
13:07:23: [<ffffffffa05c6d40>] ? obd_zombie_impexp_thread+0x0/0x190 [obdclass]
13:07:23: [<ffffffff8109e66e>] kthread+0x9e/0xc0
13:07:23: [<ffffffff8100c20a>] child_rip+0xa/0x20
13:07:23: [<ffffffff8109e5d0>] ? kthread+0x0/0xc0
13:07:23: [<ffffffff8100c200>] ? child_rip+0x0/0x20
13:07:23:
13:07:23:LustreError: 11118:0:(mdt_handler.c:4471:mdt_fini()) ASSERTION( atomic_read(&d->ld_ref) == 0 ) failed: 
13:07:23:LustreError: 11118:0:(mdt_handler.c:4471:mdt_fini()) LBUG
13:07:23:Pid: 11118, comm: umount
13:07:23:
13:07:23:Call Trace:
13:07:23: [<ffffffffa0491895>] libcfs_debug_dumpstack+0x55/0x80 [libcfs]
13:07:23: [<ffffffffa0491e97>] lbug_with_loc+0x47/0xb0 [libcfs]
13:07:23: [<ffffffffa0edde84>] mdt_device_fini+0x12f4/0x13a0 [mdt]
13:07:23: [<ffffffffa05c7986>] ? class_disconnect_exports+0x116/0x2f0 [obdclass]
13:07:23: [<ffffffffa05e41a2>] class_cleanup+0x552/0xd10 [obdclass]
13:07:23: [<ffffffffa05c4f26>] ? class_name2dev+0x56/0xe0 [obdclass]
13:07:23: [<ffffffffa05e694a>] class_process_config+0x1fea/0x27c0 [obdclass]
13:07:23: [<ffffffff81174d4c>] ? __kmalloc+0x21c/0x230
13:07:23: [<ffffffffa05df955>] ? lustre_cfg_new+0x435/0x630 [obdclass]
13:07:23: [<ffffffffa05e7241>] class_manual_cleanup+0x121/0x870 [obdclass]
13:07:23: [<ffffffffa05c4f26>] ? class_name2dev+0x56/0xe0 [obdclass]
13:07:23: [<ffffffffa061f7f7>] server_put_super+0xb37/0xe50 [obdclass]
13:07:23: [<ffffffff8119083b>] generic_shutdown_super+0x5b/0xe0
13:07:23: [<ffffffff81190926>] kill_anon_super+0x16/0x60
13:07:23: [<ffffffffa05e9496>] lustre_kill_super+0x36/0x60 [obdclass]
13:07:23: [<ffffffff811910c7>] deactivate_super+0x57/0x80
13:07:23: [<ffffffff811b0cff>] mntput_no_expire+0xbf/0x110
13:07:23: [<ffffffff811b184b>] sys_umount+0x7b/0x3a0
13:07:23: [<ffffffff8100b072>] system_call_fastpath+0x16/0x1b
13:07:23:
13:07:23:Kernel panic - not syncing: LBUG

Please provide additional information about the failure here.

Info required for matching: sanity 17n



 Comments   
Comment by Di Wang [ 31/Mar/15 ]

The fix has been incorporated into http://review.whamcloud.com/#/c/11572

Generated at Sat Feb 10 01:58:55 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.