Details
-
Bug
-
Resolution: Duplicate
-
Minor
-
None
-
Lustre 2.8.0
-
None
-
server: lustre-master build # 3120 RHEL7.1
client: SLES11 SP3
-
3
-
9223372036854775807
Description
This issue was created by maloo for sarah_lw <wei3.liu@intel.com>
This issue relates to the following test suite run: https://testing.hpdd.intel.com/test_sets/d747913c-36eb-11e5-8aa3-5254006e85c2.
The sub-test conf-sanity failed with the following error:
test failed to respond and timed out
MDS console
============----- Wed Jul 29 14:28:52 PDT 2015 21:29:48:[ 6939.261454] Lustre: DEBUG MARKER: -----============= acceptance-small: conf-sanity ============----- Wed Jul 29 14:28:52 PDT 2015 21:29:48:[ 6939.725706] Lustre: DEBUG MARKER: /usr/sbin/lctl mark excepting tests: 32newtarball 21:29:48:[ 6939.843346] Lustre: DEBUG MARKER: excepting tests: 32newtarball 21:29:48:[ 6941.017504] Lustre: DEBUG MARKER: grep -c /mnt/mds1' ' /proc/mounts 21:29:48:[ 6941.263857] Lustre: DEBUG MARKER: umount -d -f /mnt/mds1 21:29:48:[ 6944.448137] LustreError: 2957:0:(client.c:1144:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88007853aa00 x1508060907555292/t0(0) o13->lustre-OST0004-osc-MDT0000@10.2.4.186@tcp:7/4 lens 224/368 e 0 to 0 dl 0 ref 1 fl Rpc:/0/ffffffff rc 0/-1 21:29:48:[ 6944.454124] LustreError: 2957:0:(client.c:1144:ptlrpc_import_delay_req()) Skipped 4 previous similar messages 21:29:48:[ 6945.325038] Lustre: lustre-MDT0000: Not available for connect from 10.2.4.186@tcp (stopping) 21:29:48:[ 6945.325039] Lustre: lustre-MDT0000: Not available for connect from 10.2.4.186@tcp (stopping) 21:29:48:[ 6945.334071] Lustre: Skipped 5 previous similar messages 21:29:48:[ 6953.424097] Lustre: 15445:0:(client.c:2020:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1438205342/real 1438205342] req@ffff88006bb46400 x1508060907555340/t0(0) o251->MGC10.2.4.185@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1438205348 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1 21:29:48:[ 6953.940790] Lustre: server umount lustre-MDT0000 complete 21:29:48:[ 6954.052294] Lustre: DEBUG MARKER: lsmod | grep lnet > /dev/null && lctl dl | grep ' ST ' 21:29:48:[ 6967.843149] Lustre: DEBUG MARKER: grep -c /mnt/mds1' ' /proc/mounts 21:29:48:[ 6968.122930] Lustre: DEBUG MARKER: lsmod | grep lnet > /dev/null && lctl dl | grep ' ST ' 21:29:48:[ 6968.433002] Lustre: DEBUG MARKER: mkfs.lustre --mgs --fsname=lustre --mdt --index=0 --param=sys.timeout=20 --param=lov.stripesize=1048576 --param=lov.stripecount=0 --param=mdt.identity_upcall=/usr/sbin/l_getidentity --backfstype=ldiskfs --device-size=200000 --mkfsoptions="-E lazy_itable_i 21:29:48:[ 6969.075173] LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. Opts: user_xattr,errors=remount-ro 21:29:48:[ 6985.644418] Lustre: DEBUG MARKER: mkdir -p /mnt/mds1 21:29:48:[ 6985.873774] Lustre: DEBUG MARKER: test -b /dev/lvm-Role_MDS/P1 21:29:48:[ 6986.104783] Lustre: DEBUG MARKER: mkdir -p /mnt/mds1; mount -t lustre /dev/lvm-Role_MDS/P1 /mnt/mds1 21:29:48:[ 6986.315220] LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. Opts: user_xattr,errors=remount-ro 21:29:48:[ 6986.561765] LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. Opts: user_xattr,errors=remount-ro,no_mbcache 21:29:48:[ 6986.607695] Lustre: Setting parameter lustre-MDT0000-mdtlov.lov.stripesize in log lustre-MDT0000 21:29:48:[ 6986.775915] Lustre: 15982:0:(osd_internal.h:1085:osd_trans_exec_check()) op 3: used 8, used now 8, reserved 4 21:29:48:[ 6986.781997] Lustre: 15982:0:(osd_internal.h:1085:osd_trans_exec_check()) Skipped 2 previous similar messages 21:29:48:[ 6986.784728] Lustre: 15982:0:(osd_handler.c:902:osd_trans_dump_creds()) create: 1/4/8, destroy: 0/0/0 21:29:48:[ 6986.786932] Lustre: 15982:0:(osd_handler.c:902:osd_trans_dump_creds()) Skipped 7 previous similar messages 21:29:48:[ 6986.789329] Lustre: 15982:0:(osd_handler.c:909:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 0/0/0 21:29:48:[ 6986.791595] Lustre: 15982:0:(osd_handler.c:909:osd_trans_dump_creds()) Skipped 7 previous similar messages 21:29:48:[ 6986.793997] Lustre: 15982:0:(osd_handler.c:919:osd_trans_dump_creds()) write: 0/0/0, punch: 0/0/0, quota 2/2/0 21:29:48:[ 6986.796392] Lustre: 15982:0:(osd_handler.c:919:osd_trans_dump_creds()) Skipped 7 previous similar messages 21:29:48:[ 6986.798858] Lustre: 15982:0:(osd_handler.c:926:osd_trans_dump_creds()) insert: 1/17/0, delete: 0/0/0 21:29:48:[ 6986.801203] Lustre: 15982:0:(osd_handler.c:926:osd_trans_dump_creds()) Skipped 7 previous similar messages 21:29:48:[ 6986.803701] Lustre: 15982:0:(osd_handler.c:933:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 21:29:48:[ 6986.805925] Lustre: 15982:0:(osd_handler.c:933:osd_trans_dump_creds()) Skipped 7 previous similar messages 21:29:48:[ 6986.808305] LustreError: 15982:0:(osd_internal.h:1088:osd_trans_exec_check()) LBUG 21:29:48:[ 6986.810317] Pid: 15982, comm: llog_process_th 21:29:48:[ 6986.812333] 21:29:48:[ 6986.812333] Call Trace: 21:29:48:[ 6986.815553] [<ffffffffa06147d3>] libcfs_debug_dumpstack+0x53/0x80 [libcfs] 21:29:48:[ 6986.817690] [<ffffffffa0614d75>] lbug_with_loc+0x45/0xc0 [libcfs] 21:29:48:[ 6986.819616] [<ffffffffa0bf481e>] osd_it_ea_rec.part.94+0x0/0x36 [osd_ldiskfs] 21:29:48:[ 6986.821735] [<ffffffffa0bcd8c7>] osd_object_ea_create+0x927/0xb80 [osd_ldiskfs] 21:29:48:[ 6986.823728] [<ffffffffa075afa8>] dt_find_or_create+0x598/0x8d0 [obdclass] 21:29:48:[ 6986.825719] [<ffffffffa05f4d40>] fld_index_init+0x1a0/0xbb0 [fld] 21:29:48:[ 6986.827708] [<ffffffffa05f2485>] fld_server_init+0xa5/0x3e0 [fld] 21:29:48:[ 6986.829555] [<ffffffffa0d8dcf4>] mdt_init0+0x4f4/0x12b0 [mdt] 21:29:48:[ 6986.831523] [<ffffffffa0d8eb29>] mdt_device_alloc+0x79/0x110 [mdt] 21:29:48:[ 6986.833346] [<ffffffffa073ba84>] obd_setup+0x114/0x2a0 [obdclass] 21:29:48:[ 6986.835297] [<ffffffffa073cb77>] class_setup+0x2f7/0x8d0 [obdclass] 21:29:48:[ 6986.837065] [<ffffffffa0744c96>] class_process_config+0x1c36/0x2db0 [obdclass] 21:29:48:[ 6986.839037] [<ffffffff812de5d9>] ? simple_strtoul+0x9/0x10 21:29:48:[ 6986.840735] [<ffffffffa0748fb0>] ? target_name2index+0x90/0xc0 [obdclass] 21:29:48:[ 6986.842695] [<ffffffff811acbe3>] ? __kmalloc+0x1f3/0x230 21:29:48:[ 6986.844363] [<ffffffffa073e6bb>] ? lustre_cfg_new+0x8b/0x400 [obdclass] 21:29:48:[ 6986.846252] [<ffffffffa07466bd>] class_config_llog_handler+0x8ad/0x1d20 [obdclass] 21:29:48:[ 6986.848028] [<ffffffffa0707345>] llog_process_thread+0x5f5/0x1020 [obdclass] 21:29:48:[ 6986.849909] [<ffffffffa07087c0>] ? llog_process_thread_daemonize+0x0/0x80 [obdclass] 21:29:48:[ 6986.851690] [<ffffffffa070880c>] llog_process_thread_daemonize+0x4c/0x80 [obdclass] 21:29:48:[ 6986.853606] [<ffffffff8109739f>] kthread+0xcf/0xe0 21:29:48:[ 6986.855145] [<ffffffff810972d0>] ? kthread+0x0/0xe0 21:29:48:[ 6986.856831] [<ffffffff81615018>] ret_from_fork+0x58/0x90 21:29:48:[ 6986.858357] [<ffffffff810972d0>] ? kthread+0x0/0xe0 21:29:48:[ 6986.859975] 21:29:48:[ 6986.861579] Kernel panic - not syncing: LBUG 21:29:48:[ 6986.862575] CPU: 1 PID: 15982 Comm: llog_process_th Tainted: GF O-------------- 3.10.0-229.7.2.el7_lustre.g8bca546.x86_64 #1 21:29:48:[ 6986.862575] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2007 21:29:48:[ 6986.862575] ffffffffa0631faf 000000004569312b ffff880067503870 ffffffff816051aa 21:29:48:[ 6986.862575] ffff8800675038f0 ffffffff815fea1e ffffffff00000008 ffff880067503900 21:29:48:[ 6986.862575] ffff8800675038a0 000000004569312b ffffffffa0bf67d0 0000000000000246 21:29:48:[ 6986.862575] Call Trace: 21:29:48:[ 6986.862575] [<ffffffff816051aa>] dump_stack+0x19/0x1b 21:29:48:[ 6986.862575] [<ffffffff815fea1e>] panic+0xd8/0x1e7 21:29:48:[ 6986.862575] [<ffffffffa0614ddb>] lbug_with_loc+0xab/0xc0 [libcfs] 21:29:48:[ 6986.862575] [<ffffffffa0bf481e>] osd_trans_exec_check.part.91+0x1a/0x1a [osd_ldiskfs] 21:29:48:[ 6986.862575] [<ffffffffa0bcd8c7>] osd_object_ea_create+0x927/0xb80 [osd_ldiskfs] 21:29:48:[ 6986.862575] [<ffffffffa075afa8>] dt_find_or_create+0x598/0x8d0 [obdclass] 21:29:48:[ 6986.862575] [<ffffffffa05f4d40>] fld_index_init+0x1a0/0xbb0 [fld] 21:29:48:[ 6986.862575] [<ffffffffa05f2485>] fld_server_init+0xa5/0x3e0 [fld] 21:29:48:[ 6986.862575] [<ffffffffa0d8dcf4>] mdt_init0+0x4f4/0x12b0 [mdt] 21:29:48:[ 6986.862575] [<ffffffffa0d8eb29>] mdt_device_alloc+0x79/0x110 [mdt] 21:29:48:[ 6986.862575] [<ffffffffa073ba84>] obd_setup+0x114/0x2a0 [obdclass] 21:29:48:[ 6986.862575] [<ffffffffa073cb77>] class_setup+0x2f7/0x8d0 [obdclass] 21:29:48:[ 6986.862575] [<ffffffffa0744c96>] class_process_config+0x1c36/0x2db0 [obdclass] 21:29:48:[ 6986.862575] [<ffffffff812de5d9>] ? simple_strtoul+0x9/0x10 21:29:48:[ 6986.862575] [<ffffffffa0748fb0>] ? target_name2index+0x90/0xc0 [obdclass] 21:29:48:[ 6986.862575] [<ffffffff811acbe3>] ? __kmalloc+0x1f3/0x230 21:29:48:[ 6986.862575] [<ffffffffa073e6bb>] ? lustre_cfg_new+0x8b/0x400 [obdclass] 21:29:48:[ 6986.862575] [<ffffffffa07466bd>] class_config_llog_handler+0x8ad/0x1d20 [obdclass] 21:29:48:[ 6986.862575] [<ffffffffa0707345>] llog_process_thread+0x5f5/0x1020 [obdclass] 21:29:48:[ 6986.862575] [<ffffffffa07087c0>] ? llog_backup+0x500/0x500 [obdclass] 21:29:48:[ 6986.862575] [<ffffffffa070880c>] llog_process_thread_daemonize+0x4c/0x80 [obdclass] 21:29:48:[ 6986.862575] [<ffffffff8109739f>] kthread+0xcf/0xe0 21:29:48:[ 6986.862575] [<ffffffff810972d0>] ? kthread_create_on_node+0x140/0x140 21:29:48:[ 6986.862575] [<ffffffff81615018>] ret_from_fork+0x58/0x90 21:29:48:[ 6986.862575] [<ffffffff810972d0>] ? kthread_create_on_node+0x140/0x140 21:29:48:[ 6986.862575] drm_kms_helper: panic occurred, switching back to text console 21:29:48:[ 6986.862575] ------------[ cut here ]------------ 21:29:48:[ 6986.862575] kernel BUG at arch/x86/mm/pageattr.c:216! 21:29:48:[ 6986.862575] invalid opcode: 0000 [#1] SMP 21:29:48:[ 6986.862575] Modules linked in: osp(OF) mdd(OF) lod(OF) mdt(OF) lfsck(OF) mgs(OF) mgc(OF) osd_ldiskfs(OF) lquota(OF) fid(OF) fld(OF) ksocklnd(OF) ptlrpc(OF) obdclass(OF) lnet(OF) sha512_generic libcfs(OF) ldiskfs(OF) dm_mod rpcsec_gss_krb5 auth_rpcgss nfsv4 dns_resolver nfs lockd fscache xprtrdma sunrpc ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp scsi_transport_srp scsi_tgt ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ppdev ib_sa parport_pc virtio_balloon pcspkr serio_raw parport i2c_piix4 ib_mad ib_core ib_addr ext4 mbcache jbd2 ata_generic pata_acpi cirrus syscopyarea sysfillrect virtio_blk sysimgblt 8139too drm_kms_helper ttm virtio_pci virtio_ring virtio 8139cp mii drm i2c_core ata_piix libata floppy 21:29:48:[ 6986.862575] CPU: 1 PID: 15982 Comm: llog_process_th Tainted: GF O-------------- 3.10.0-229.7.2.el7_lustre.g8bca546.x86_64 #1
It looks like the
LU-6827patch did land after that test run so we can close this as a duplicate and reopen if that is proven not to be the case.