[LU-3330] mds crashed during umount after mds-survey run Created: 13/May/13  Updated: 14/May/13  Resolved: 14/May/13

Status: Closed
Project: Lustre
Component/s: None
Affects Version/s: None
Fix Version/s: None

Type: Bug Priority: Minor
Reporter: Minh Diep Assignee: WC Triage
Resolution: Duplicate Votes: 0
Labels: None

Issue Links:
Duplicate
duplicates LU-3021 replay-vbr test 11a: RIP: ldiskfs_mb_... Resolved
Severity: 3
Rank (Obsolete): 8221

 Description   

mds crashed during umount after mds-survey run ended with error

LustreError: 9487:0:(echo_client.c:1959:echo_md_destroy_internal()) Can not unlink child tests: rc = -30
LustreError: 9668:0:(echo_client.c:1959:echo_md_destroy_internal()) Can not unlink child tests: rc = -30
LustreError: 9668:0:(echo_client.c:1959:echo_md_destroy_internal()) Skipped 63 previous similar messages
Lustre: Failing over lustre-MDT0000
LDISKFS-fs error (device sdc1): ldiskfs_mb_release_inode_pa: pa free mismatch: [pa ffff8805c3244208] [phy 134080] [logic 0] [len 1024] [free 1024] [error 1] [inode 15] [freed 0]
LDISKFS-fs error (device sdc1): ldiskfs_mb_release_inode_pa: free 0, pa_free 1024
-----------[ cut here ]-----------
kernel BUG at /var/lib/jenkins/workspace/lustre-master/arch/x86_64/build_type/server/distro/el6/ib_stack/inkernel/BUILD/BUILD/lustre-ldiskfs-4.1.0/ldiskfs/mballoc.c:3786!
invalid opcode: 0000 1 SMP
last sysfs file: /sys/devices/pci0000:00/0000:00:03.0/0000:02:00.0/host0/port-0:0/expander-0:0/port-0:0:4/end_device-0:0:4/target0:0:4/0:0:4:0/block/sdl/sdl1/uevent
CPU 11
Modules linked in: obdecho(U) osp(U) lod(U) mdt(U) mgs(U) mgc(U) fsfilt_ldiskfs(U) osd_ldiskfs(U) lquota(U) mdd(U) lustre(U) lov(U) osc(U) mdc(U) fid(U) fld(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) sha512_generic sha256_generic libcfs(U) ldiskfs(U) jbd2 nfsd exportfs nfs lockd fscache auth_rpcgss nfs_acl sunrpc cpufreq_ondemand acpi_cpufreq freq_table mperf ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_addr ipv6 microcode serio_raw mlx4_ib ib_sa ib_mad ib_core mlx4_en mlx4_core i2c_i801 i2c_core iTCO_wdt iTCO_vendor_support ioatdma i7core_edac edac_core ses enclosure sg igb dca ptp pps_core ext3 jbd mbcache sr_mod cdrom sd_mod crc_t10dif pata_acpi ata_generic ata_piix mpt2sas scsi_transport_sas raid_class dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]

Pid: 9812, comm: umount Not tainted 2.6.32-358.6.1.el6_lustre.x86_64 #1 Supermicro X8DTH-i/6/iF/6F/X8DTH
RIP: 0010:[<ffffffffa048ab76>] [<ffffffffa048ab76>] ldiskfs_mb_release_inode_pa+0x346/0x360 [ldiskfs]
RSP: 0018:ffff8807ea2497e8 EFLAGS: 00010287
RAX: 0000000000000400 RBX: 0000000000000000 RCX: ffff880541de2c00
RDX: 0000000000000000 RSI: 0000000000000046 RDI: ffff8804fe0f8200
RBP: ffff8807ea249898 R08: ffffffff81c07720 R09: 0000000000000000
R10: 0000000000000003 R11: 000000000000005a R12: ffff88064b8c0898
R13: ffff8804f06d6748 R14: 0000000000000c00 R15: ffff8805c3244208
FS: 00007f0885262740(0000) GS:ffff8800282e0000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00000000016ad018 CR3: 00000007e0d68000 CR4: 00000000000007e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process umount (pid: 9812, threadinfo ffff8807ea248000, task ffff88082ee67540)
Stack:
ffff880700000400 0000000000000400 ffff880400000001 000000000000000f
<d> 0000000000000000 0000000000000008 ffff8807ea249828 ffffffff811b60b6
<d> ffff880541de2c00 ffff88051293d800 ffff8805c3244208 000000000001ffc0
Call Trace:
[<ffffffff811b60b6>] ? __wait_on_buffer+0x26/0x30
[<ffffffffa048f1ee>] ldiskfs_discard_preallocations+0x1fe/0x490 [ldiskfs]
[<ffffffffa04a5166>] ldiskfs_clear_inode+0x16/0x50 [ldiskfs]
[<ffffffff8119d00c>] clear_inode+0xac/0x140
[<ffffffff8119d0e0>] dispose_list+0x40/0x120
[<ffffffff8119d58a>] invalidate_inodes+0xea/0x190
[<ffffffff8118333c>] generic_shutdown_super+0x4c/0xe0
[<ffffffff81183401>] kill_block_super+0x31/0x50
[<ffffffff81183bd7>] deactivate_super+0x57/0x80
[<ffffffff811a1c4f>] mntput_no_expire+0xbf/0x110
[<ffffffffa0d355f9>] osd_umount+0x79/0x150 [osd_ldiskfs]
[<ffffffffa0d3d037>] osd_device_fini+0x147/0x190 [osd_ldiskfs]
[<ffffffffa063d937>] class_cleanup+0x577/0xda0 [obdclass]
[<ffffffffa0612af6>] ? class_name2dev+0x56/0xe0 [obdclass]
[<ffffffffa063f21c>] class_process_config+0x10bc/0x1c80 [obdclass]
[<ffffffffa0638a43>] ? lustre_cfg_new+0x353/0x7e0 [obdclass]
[<ffffffffa063ff59>] class_manual_cleanup+0x179/0x6f0 [obdclass]
[<ffffffffa04f1717>] ? cfs_waitq_broadcast+0x17/0x20 [libcfs]
[<ffffffffa0610ea6>] ? class_export_put+0xf6/0x2b0 [obdclass]
[<ffffffffa0d3dfe5>] osd_obd_disconnect+0x1c5/0x1d0 [osd_ldiskfs]
[<ffffffffa0641f8e>] lustre_put_lsi+0x17e/0x1100 [obdclass]
[<ffffffffa064ace8>] lustre_common_put_super+0x5f8/0xc40 [obdclass]
[<ffffffffa06753aa>] server_put_super+0x1ca/0xf00 [obdclass]
[<ffffffff8118334b>] generic_shutdown_super+0x5b/0xe0
[<ffffffff81183436>] kill_anon_super+0x16/0x60
[<ffffffffa0641db6>] lustre_kill_super+0x36/0x60 [obdclass]
[<ffffffff81183bd7>] deactivate_super+0x57/0x80
[<ffffffff811a1c4f>] mntput_no_expire+0xbf/0x110
[<ffffffff811a26bb>] sys_umount+0x7b/0x3a0
[<ffffffff8100b072>] system_call_fastpath+0x16/0x1b
Code: 55 c8 e9 39 fe ff ff 31 db 41 83 7f 4c 00 0f 84 7e fd ff ff 0f 0b eb fe 0f 0b eb fe 0f 0b 0f 1f 80 00 00 00 00 eb f7 0f 0b eb fe <0f> 0b 0f 1f 84 00 00 00 00 00 eb f6 66 66 66 66 66 2e 0f 1f 84
RIP [<ffffffffa048ab76>] ldiskfs_mb_release_inode_pa+0x346/0x360 [ldiskfs]
RSP <ffff8807ea2497e8>



 Comments   
Comment by Niu Yawei (Inactive) [ 14/May/13 ]

This looks like same issue of LU-3021.

Comment by Jodi Levi (Inactive) [ 14/May/13 ]

Duplicate of LU-3021

Generated at Sat Feb 10 01:32:59 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.