[LU-8462] OSS keeps dropping into KDB Created: 02/Aug/16  Updated: 26/Nov/16  Resolved: 26/Nov/16

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: Lustre 2.5.3
Fix Version/s: None

Type: Bug Priority: Major
Reporter: Herbert Yeung Assignee: nasf (Inactive)
Resolution: Fixed Votes: 0
Labels: None

Issue Links:
Related
is related to LU-7114 ldiskfs: corrupted bitmaps handling p... Resolved
is related to LU-8463 OSSes drop into KDB during recovery Resolved
Severity: 2
Rank (Obsolete): 9223372036854775807

 Description   

On Tuesday, July 26th, the RAID controller firmware was updated, causing two OSSes to drop into KDB due to missing paths. Since then, we have had a spate of crashes on the file system:

Tue Jul 26 - 394 - 2028 KDB (I/O)

  • 395 - 2000 KDB (I/O)
    Fri jUl 29 - s393 - 0910 KDB (aborting journal)
    Fri Jul 29 - s394 - 1922 KDB (disk)
    Fri Jul 29 - s393 - 2123 KDB (disk)

s394 Tue Jul 26 - 2028 :

Buffer I/O error on device dm-21, logical block 137126919
lost page write due to I/O error on dm-21
end_request: I/O error, dev dm-21, sector 1073743416
LDISKFS-fs error (device dm-21): ldiskfs_read_block_bitmap: Cannot read block bitmap - block_group = 4295, block_bitmap = 134217927
Entering kdb (current=0xffff880360d92ab0, pid 63652) on processor 0 Oops: (null)
due to oops @ 0x0
kdba_dumpregs: pt_regs not available, use bt* or pid to select a different task
[0]kdb> [-- MARK -- Tue Jul 26 18:00:00 2016]

s395 Tue Jul 26 - 2000 :

end_request: I/O error, dev dm-18, sector 759488504
end_request: I/O error, dev dm-18, sector 759488504
end_request: I/O error, dev dm-18, sector 295072
LustreError: 63719:0:(osd_io.c:1179:osd_ldiskfs_write_record()) dm-18: error reading offset 1342464 (block 327): rc = -5
LustreError: 63719:0:(osd_handler.c:1051:osd_trans_stop()) Failure in transaction hook: -5
------------[ cut here ]------------
WARNING: at fs/buffer.c:1186 mark_buffer_dirty+0x82/0xa0() (Tainted: G        W  ---------------   )
Hardware name: C1104G-RP5
Modules linked in: osp(U) ofd(U) lfsck(U) ost(U) mgc(U) fsfilt_ldiskfs(U) osd_ldiskfs(U) lquota(U) ldiskfs(U) jbd2 lustre(U) lov(U) osc(U) mdc(U) fid(U) fld(U) ko2iblnd(U) ptlrpc(U) obdclass(U) lnet(U) lvfs(U) sha512_generic crc32c_intel libcfs(U) sunrpc dm_round_robin bonding ib_ucm(U) rdma_ucm(U) rdma_cm(U) iw_cm(U) configfs ib_srp(U) scsi_transport_srp(U) ib_ipoib(U) ib_cm(U) ib_uverbs(U) ib_umad(U) dm_multipath microcode acpi_pad iTCO_wdt iTCO_vendor_support igb hwmon dca i2c_algo_bit ptp pps_core i2c_i801 i2c_core sg lpc_ich mfd_core tcp_bic ext3 jbd sd_mod crc_t10dif ahci isci libsas scsi_transport_sas wmi mlx4_ib(U) ib_sa(U) ib_mad(U) ib_core(U) ib_addr(U) ipv6 mlx4_core(U) mlx_compat(U) memtrack(U) dm_mirror dm_region_hash dm_log dm_mod gru [last unloaded: scsi_wait_scan]
Pid: 36266, comm: jbd2/dm-6 Tainted: G        W  ---------------    2.6.32-504.30.3.el6.20151008.x86_64.lustre253 #1
Call Trace:
 [<ffffffff81074127>] ? warn_slowpath_common+0x87/0xc0
 [<ffffffff8107417a>] ? warn_slowpath_null+0x1a/0x20
 [<ffffffff811c1ca2>] ? mark_buffer_dirty+0x82/0xa0
 [<ffffffffa0b6b385>] ? __jbd2_journal_temp_unlink_buffer+0xa5/0x140 [jbd2]
 [<ffffffffa0b6b436>] ? __jbd2_journal_unfile_buffer+0x16/0x30 [jbd2]
 [<ffffffffa0b6b798>] ? __jbd2_journal_refile_buffer+0xc8/0xf0 [jbd2]
 [<ffffffffa0b6e548>] ? jbd2_journal_commit_transaction+0xdc8/0x15a0 [jbd2]
 [<ffffffff810873eb>] ? try_to_del_timer_sync+0x7b/0xe0
 [<ffffffffa0b73c48>] ? kjournald2+0xb8/0x220 [jbd2]
 [<ffffffff8109e120>] ? autoremove_wake_function+0x0/0x40
 [<ffffffffa0b73b90>] ? kjournald2+0x0/0x220 [jbd2]
 [<ffffffff8109dc8e>] ? kthread+0x9e/0xc0
 [<ffffffff8100c28a>] ? child_rip+0xa/0x20
 [<ffffffff8109dbf0>] ? kthread+0x0/0xc0
 [<ffffffff8100c280>] ? child_rip+0x0/0x20
---[ end trace 8870bed99a68f953 ]---
srp_daemon[3088]: No response to inform info registration
srp_daemon[3088]: Fail to register to traps, maybe there is no opensm running on fabric or IB port is down
srp_daemon[3088]: bad MAD status (110) from lid 1
scsi host7: ib_srp: Got failed path rec status -110
scsi host7: ib_srp: Path record query failed
scsi host7: reconnect attempt 4 failed (-110)
------------[ cut here ]------------
kernel BUG at /usr/src/redhat/BUILD/lustre-2.5.3/ldiskfs/mballoc.c:3189!

Entering kdb (current=0xffff881ae3470040, pid 63719) on processor 9 due to KDB_ENTER()
[9]kdb> [-- MARK -- Tue Jul 26 18:00:00 2016]

s393 Fri Jul 29 - 0910:

LDISKFS-fs error (device dm-18): ldiskfs_mb_release_inode_pa: pa free mismatch: [pa ffff8801ce41bf28] [phy 1437833216] [logic 334848] [len 1024] [free 542] [error 0] [inode 207111] [freed 1024]
Aborting journal on device dm-4.
Kernel panic - not syncing: LDISKFS-fs (device dm-18): panic forced after error

Pid: 165, comm: kswapd1 Tainted: G        W  ---------------    2.6.32-504.30.3.el6.20151008.x86_64.lustre253 #1
Call Trace:
 [<ffffffff81564fb9>] ? panic+0xa7/0x190
 [<ffffffffa0bc5118>] ? ldiskfs_commit_super+0x188/0x210 [ldiskfs]
 [<ffffffffa0bc5724>] ? ldiskfs_handle_error+0xc4/0xd0 [ldiskfs]
 [<ffffffffa0bc5ac2>] ? __ldiskfs_error+0x82/0x90 [ldiskfs]
 [<ffffffffa0ba920f>] ? ldiskfs_mb_release_inode_pa+0x26f/0x360 [ldiskfs]

Entering kdb (current=0xffff881065789520, pid 165) on processor 8 Oops: (null)
due to oops @ 0x0
kdba_dumpregs: pt_regs not available, use bt* or pid to select a different task
[8]kdb> [-- Fri Jul 29 08:55:12 2016]

s394 Fri Jul 29 - 1922:

Buffer I/O error on device dm-21, logical block 137126919
lost page write due to I/O error on dm-21
end_request: I/O error, dev dm-21, sector 1073743416
LDISKFS-fs error (device dm-21): ldiskfs_read_block_bitmap: Cannot read block bitmap - block_group = 4295, block_bitmap = 134217927
Entering kdb (current=0xffff880360d92ab0, pid 63652) on processor 0 Oops: (null)
due to oops @ 0x0
kdba_dumpregs: pt_regs not available, use bt* or pid to select a different task
[0]kdb> [-- MARK -- Tue Jul 26 18:00:00 2016]

s393 Fri Jul 29 -2123:

[-- MARK -- Fri Jul 29 18:00:00 2016]
LDISKFS-fs error (device dm-27): ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 4891corrupted: 21438 blocks free in bitmap, 25535 - in gd

Aborting journal on device dm-2.
LDISKFS-fs error (device dm-27): ldiskfs_journal_start_sb:
Kernel panic - not syncing: LDISKFS-fs (device dm-27): panic forced after error

Pid: 95204, comm: ll_ost_io01_021 Tainted: G        W  ---------------    2.6.32-504.30.3.el6.20151008.x86_64.lustre253 #1
Call Trace:
 [<ffffffff81564fb9>] ? panic+0xa7/0x190
 [<ffffffffa0eb0118>] ? ldiskfs_commit_super+0x188/0x210 [ldiskf s ]14
 out of 16 cpus in kdb, waiting for the rest, timeout in 10 second(s)
 [<ffffffffa0eb0724>] ? ldiskfs_handle_error+0xc4/0xd0 [ldiskfs]
 [<ffffffffa0eb0ac2>] ? __ldiskfs_error+0x82/0x90 [ldiskfs]
 [<ffffffff810f2059>] ? delayacct_end+0x89/0xa0
 [<ffffffffa0e92a29>] ? ldiskfs_mb_check_ondisk_bitmap+0x149/0x150 [ldiskfs]
 [<ffffffffa0e92aad>] ? ldiskfs_mb_generate_from_pa+0x7d/0x180 [ldiskfs]
 [<ffffffff8109e1a0>] ? wake_bit_function+0x0/0x50
 [<ffffffffa0e9528b>] ? ldiskfs_mb_init_cache+0x55b/0xa30 [ldiskfs]
 [<ffffffffa0e9587e>] ? ldiskfs_mb_init_group+0x11e/0x210 [ldiskfs]
 [<ffffffffa0e95dd5>] ? ldiskfs_mb_load_buddy+0x355/0x390 [ldiskfs]
 [<ffffffffa0e7c00b>] ? __ldiskfs_handle_dirty_metadata+0x7b/0x100 [ldiskfs]
 [<ffffffffa0e96b7d>] ? ldiskfs_mb_find_by_goal+0x6d/0x2e0 [ldiskfs]
 [<ffffffffa0e97019>] ? ldiskfs_mb_regular_allocator+0x59/0x410 [ldiskfs]
 [<ffffffffa0eb12e8>] ? __ldiskfs_journal_stop+0x68/0xa0 [ldiskfs]
 [<ffffffffa0e91872>] ? ldiskfs_mb_normalize_request+0x2c2/0x3d0 [ldiskfs]
 [<ffffffffa0e9923d>] ? ldiskfs_mb_new_blocks+0x47d/0x630 [ldiskfs]
 [<ffffffffa0eb1390>] ? ldiskfs_journal_start_sb+0x70/0x170 [ldiskfs]
 [<ffffffffa02369ff>] ? ldiskfs_ext_new_extent_cb+0x57f/0x6cc [fsfilt_ldiskfs]
 [<ffffffffa0e7ecd2>] ? ldiskfs_ext_walk_space+0x142/0x310 [ldiskfs]
 [<ffffffffa0236480>] ? ldiskfs_ext_new_extent_cb+0x0/0x6cc [fsfilt_ldiskfs]
 [<ffffffffa02361cc>] ? fsfilt_map_nblocks+0xcc/0xf0 [fsfilt_ldiskfs]
 [<ffffffffa02362f0>] ? fsfilt_ldiskfs_map_ext_inode_pages+0x100/0x200 [fsfilt_ldiskfs]
 [<ffffffffa0236475>] ? fsfilt_ldiskfs_map_inode_pages+0x85/0x90 [fsfilt_ldiskfs]
 [<ffffffffa0eb1884>] ? ldiskfs_dquot_initialize+0x94/0xd0 [ldiskfs]
 [<ffffffffa0f89342>] ? osd_write_commit+0x302/0x620 [osd_ldiskfs]
 [<ffffffffa137c094>] ? ofd_commitrw_write+0x684/0x11b0 [ofd]
 [<ffffffffa137edfd>] ? ofd_commitrw+0x5cd/0xbb0 [ofd]
 [<ffffffffa0372861>] ? lprocfs_counter_add+0x151/0x1d6 [lvfs]
 [<ffffffffa131319d>] ? obd_commitrw+0x11d/0x390 [ost]
 [<ffffffffa131cf71>] ? ost_brw_write+0xea1/0x15d0 [ost]
 [<ffffffff812b9076>] ? vsnprintf+0x336/0x5e0
 [<ffffffffa0686500>] ? target_bulk_timeout+0x0/0xc0 [ptlrpc]
 [<ffffffffa13235cc>] ? ost_handle+0x439c/0x44d0 [ost]
 [<ffffffffa03f8a44>] ? libcfs_id2str+0x74/0xb0 [libcfs]
 [<ffffffffa06d60c5>] ? ptlrpc_server_handle_request+0x385/0xc00 [ptlrpc]
 [<ffffffffa03fe8d5>] ? lc_watchdog_touch+0x65/0x170 [libcfs]
 [<ffffffffa06cea69>] ? ptlrpc_wait_event+0xa9/0x2d0 [ptlrpc]
 [<ffffffffa06d889d>] ? ptlrpc_main+0xafd/0x1780 [ptlrpc]
 [<ffffffff8100c28a>] ? child_rip+0xa/0x20
 [<ffffffffa06d7da0>] ? ptlrpc_main+0x0/0x1780 [ptlrpc]
LDISKFS-fs error (device dm-27): ldiskfs_journal_start_sb: Detected aborted journal [<ffffffff8100c280>] ? child_rip+0x0/0x20

.  15 out of 16 cpus in kdb, waiting for the rest, timeout in 9 second(s)
..1 cpu is not in kdb, its state is unknown

Entering kdb (current=0xffff880e989cc040, pid 95204) on processor 7 Oops: (null)
due to oops @ 0x0
kdba_dumpregs: pt_regs not available, use bt* or pid to select a different task
[7]kdb> [-- MARK -- Fri Jul 29 19:00:00 2016]


 Comments   
Comment by Oleg Drokin [ 02/Aug/16 ]

It looks like when you upgraded your RAID controller firmware, it makes your RAID controller to return IO errors when Lustre tries to read data from the array.

I imagine this needs to be resolved first to have any sort of stable operations. Then after that is resolved you'll probably need to run e2fsck to fix whatever disk problems happened (this is only after IO errors are totally gone! or you are risking making things even worse).

As for the crash - yes, it should not really be happening during IO errors, but it's like a third order after effect. These failures are fixed in newer Lustre releases.

Comment by Jay Lan (Inactive) [ 02/Aug/16 ]

I saw two crash dumps on service395 (nbp6-oss16.) The first crash happened after the system had been up running for 40 days before crash. The second crash happened during recovery after 1st crash.

Here is the stack trace of the first crash:

KERNEL: ../vmlinux
DUMPFILE: vmcore [PARTIAL DUMP]
CPUS: 16
DATE: Tue Jul 26 17:49:06 2016
UPTIME: 40 days, 05:53:18
LOAD AVERAGE: 1.70, 1.15, 0.58
TASKS: 1526
NODENAME: nbp6-oss16
RELEASE: 2.6.32-504.30.3.el6.20151008.x86_64.lustre253
VERSION: #1 SMP Tue Oct 27 21:45:38 EDT 2015
MACHINE: x86_64 (2599 Mhz)
MEMORY: 128 GB
PANIC: "kernel BUG at /usr/src/redhat/BUILD/lustre-2.5.3/ldiskfs/mballoc.c:3189!"
PID: 63719
COMMAND: "ll_ost_io02_020"
TASK: ffff881ae3470040 [THREAD_INFO: ffff881233f5a000]
CPU: 9
STATE: TASK_RUNNING (PANIC)

crash> bt
PID: 63719 TASK: ffff881ae3470040 CPU: 9 COMMAND: "ll_ost_io02_020"
#0 [ffff881233f5b030] machine_kexec at ffffffff8103b5db
#1 [ffff881233f5b090] crash_kexec at ffffffff810c9432
#2 [ffff881233f5b160] kdb_kdump_check at ffffffff8129ad17
#3 [ffff881233f5b170] kdb_main_loop at ffffffff8129df07
#4 [ffff881233f5b280] kdb_save_running at ffffffff8129806e
#5 [ffff881233f5b290] kdba_main_loop at ffffffff814806a8
#6 [ffff881233f5b2d0] kdb at ffffffff8129b206
#7 [ffff881233f5b340] report_bug at ffffffff812ae6b3
#8 [ffff881233f5b370] die at ffffffff8101106f
#9 [ffff881233f5b3a0] do_trap at ffffffff81569614
#10 [ffff881233f5b400] do_invalid_op at ffffffff8100d025
#11 [ffff881233f5b4a0] invalid_op at ffffffff8100c01b
[exception RIP: ldiskfs_mb_use_inode_pa+186]
RIP: ffffffffa0ba5d2a RSP: ffff881233f5b550 RFLAGS: 00010202
RAX: 000000000000000b RBX: ffff881e610fd898 RCX: ffff881e610fd8cc
RDX: 0000000000000400 RSI: 0000000005eec7f4 RDI: 0000000000008000
RBP: ffff881233f5b570 R8: ffff881e610fd8d0 R9: 0000000000000000
R10: 0000000000000010 R11: 0000000000000000 R12: ffff882069a3dd68
R13: 0000000005eec800 R14: 000000000000000c R15: 00000000ffffffff
ORIG_RAX: ffffffffffffffff CS: 0010 SS: 0018
#12 [ffff881233f5b578] ldiskfs_mb_use_preallocated at ffffffffa0ba5f59 [ldiskfs]
#13 [ffff881233f5b5d8] ldiskfs_mb_new_blocks at ffffffffa0bae0bd [ldiskfs]
#14 [ffff881233f5b678] ldiskfs_ext_new_extent_cb at ffffffffa01be9ff [fsfilt_ldiskfs]
#15 [ffff881233f5b728] ldiskfs_ext_walk_space at ffffffffa0b93cd2 [ldiskfs]
#16 [ffff881233f5b7b8] fsfilt_map_nblocks at ffffffffa01be1cc [fsfilt_ldiskfs]
#17 [ffff881233f5b818] fsfilt_ldiskfs_map_ext_inode_pages at ffffffffa01be2f0 [fsfilt_ldiskfs]
#18 [ffff881233f5b898] fsfilt_ldiskfs_map_inode_pages at ffffffffa01be475 [fsfilt_ldiskfs]
#19 [ffff881233f5b8d8] osd_write_commit at ffffffffa1488342 [osd_ldiskfs]
#20 [ffff881233f5b948] ofd_commitrw_write at ffffffffa1ae5094 [ofd]
#21 [ffff881233f5b9d8] ofd_commitrw at ffffffffa1ae7dfd [ofd]
#22 [ffff881233f5ba68] obd_commitrw at ffffffffa1a7c19d [ost]
#23 [ffff881233f5bae8] ost_brw_write at ffffffffa1a85f71 [ost]
#24 [ffff881233f5bc78] ost_handle at ffffffffa1a8c5cc [ost]
#25 [ffff881233f5bdc8] ptlrpc_server_handle_request at ffffffffa06dd0c5 [ptlrpc]
#26 [ffff881233f5bea8] ptlrpc_main at ffffffffa06df89d [ptlrpc]
#27 [ffff881233f5bf48] kernel_thread at ffffffff8100c28a
crash>

Line 3189 of ldiskfs/mballoc.c is in ldksifs_mb_use_inode_pa()
...
BUG_ON(pa->pa_free < len);

Comment by Herbert Yeung [ 02/Aug/16 ]

Copied the wrong kdb output for s394 Fri Jul 29 - 1922

LDISKFS-fs error (device dm-27): ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 4891corrupted: 21438 blocks free in bitmap, 25535 - in gd

Aborting journal on device dm-2.
LDISKFS-fs error (device dm-27): ldiskfs_journal_start_sb:
Kernel panic - not syncing: LDISKFS-fs (device dm-27): panic forced after error

Pid: 95204, comm: ll_ost_io01_021 Tainted: G W --------------- 2.6.32-504.30.3.el6.20151008.x86_64.lustre253 #1
Call Trace:
[<ffffffff81564fb9>] ? panic+0xa7/0x190
[<ffffffffa0eb0118>] ? ldiskfs_commit_super+0x188/0x210 [ldiskf s ]14
out of 16 cpus in kdb, waiting for the rest, timeout in 10 second(s)
[<ffffffffa0eb0724>] ? ldiskfs_handle_error+0xc4/0xd0 [ldiskfs]
[<ffffffffa0eb0ac2>] ? __ldiskfs_error+0x82/0x90 [ldiskfs]
[<ffffffff810f2059>] ? delayacct_end+0x89/0xa0
[<ffffffffa0e92a29>] ? ldiskfs_mb_check_ondisk_bitmap+0x149/0x150 [ldiskfs]
[<ffffffffa0e92aad>] ? ldiskfs_mb_generate_from_pa+0x7d/0x180 [ldiskfs]
[<ffffffff8109e1a0>] ? wake_bit_function+0x0/0x50
[<ffffffffa0e9528b>] ? ldiskfs_mb_init_cache+0x55b/0xa30 [ldiskfs]
[<ffffffffa0e9587e>] ? ldiskfs_mb_init_group+0x11e/0x210 [ldiskfs]
[<ffffffffa0e95dd5>] ? ldiskfs_mb_load_buddy+0x355/0x390 [ldiskfs]
[<ffffffffa0e7c00b>] ? __ldiskfs_handle_dirty_metadata+0x7b/0x100 [ldiskfs]
[<ffffffffa0e96b7d>] ? ldiskfs_mb_find_by_goal+0x6d/0x2e0 [ldiskfs]
[<ffffffffa0e97019>] ? ldiskfs_mb_regular_allocator+0x59/0x410 [ldiskfs]
[<ffffffffa0eb12e8>] ? __ldiskfs_journal_stop+0x68/0xa0 [ldiskfs]
[<ffffffffa0e91872>] ? ldiskfs_mb_normalize_request+0x2c2/0x3d0 [ldiskfs]
[<ffffffffa0e9923d>] ? ldiskfs_mb_new_blocks+0x47d/0x630 [ldiskfs]
[<ffffffffa0eb1390>] ? ldiskfs_journal_start_sb+0x70/0x170 [ldiskfs]
[<ffffffffa02369ff>] ? ldiskfs_ext_new_extent_cb+0x57f/0x6cc [fsfilt_ldiskfs]
[<ffffffffa0e7ecd2>] ? ldiskfs_ext_walk_space+0x142/0x310 [ldiskfs]
[<ffffffffa0236480>] ? ldiskfs_ext_new_extent_cb+0x0/0x6cc [fsfilt_ldiskfs]
[<ffffffffa02361cc>] ? fsfilt_map_nblocks+0xcc/0xf0 [fsfilt_ldiskfs]
[<ffffffffa02362f0>] ? fsfilt_ldiskfs_map_ext_inode_pages+0x100/0x200 [fsfilt_ldiskfs]
[<ffffffffa0236475>] ? fsfilt_ldiskfs_map_inode_pages+0x85/0x90 [fsfilt_ldiskfs]
[<ffffffffa0eb1884>] ? ldiskfs_dquot_initialize+0x94/0xd0 [ldiskfs]
[<ffffffffa0f89342>] ? osd_write_commit+0x302/0x620 [osd_ldiskfs]
[<ffffffffa137c094>] ? ofd_commitrw_write+0x684/0x11b0 [ofd]
[<ffffffffa137edfd>] ? ofd_commitrw+0x5cd/0xbb0 [ofd]
[<ffffffffa0372861>] ? lprocfs_counter_add+0x151/0x1d6 [lvfs]
[<ffffffffa131319d>] ? obd_commitrw+0x11d/0x390 [ost]
[<ffffffffa131cf71>] ? ost_brw_write+0xea1/0x15d0 [ost]
[<ffffffff812b9076>] ? vsnprintf+0x336/0x5e0
[<ffffffffa0686500>] ? target_bulk_timeout+0x0/0xc0 [ptlrpc]
[<ffffffffa13235cc>] ? ost_handle+0x439c/0x44d0 [ost]
[<ffffffffa03f8a44>] ? libcfs_id2str+0x74/0xb0 [libcfs]
[<ffffffffa06d60c5>] ? ptlrpc_server_handle_request+0x385/0xc00 [ptlrpc]
[<ffffffffa03fe8d5>] ? lc_watchdog_touch+0x65/0x170 [libcfs]
[<ffffffffa06cea69>] ? ptlrpc_wait_event+0xa9/0x2d0 [ptlrpc]
[<ffffffffa06d889d>] ? ptlrpc_main+0xafd/0x1780 [ptlrpc]
[<ffffffff8100c28a>] ? child_rip+0xa/0x20
[<ffffffffa06d7da0>] ? ptlrpc_main+0x0/0x1780 [ptlrpc]
LDISKFS-fs error (device dm-27): ldiskfs_journal_start_sb: Detected aborted journal [<ffffffff8100c280>] ? child_rip+0x0/0x20

. 15 out of 16 cpus in kdb, waiting for the rest, timeout in 9 second(s)
..1 cpu is not in kdb, its state is unknown

Entering kdb (current=0xffff880e989cc040, pid 95204) on processor 7 Oops: (null)
due to oops @ 0x0
kdba_dumpregs: pt_regs not available, use bt* or pid to select a different task
[7]kdb> [-- MARK -- Fri Jul 29 19:00:00 2016]

Comment by Herbert Yeung [ 02/Aug/16 ]

To clarify, the controllers were being downgraded as the latest version had a firmware bug. They were downgraded to the version that all of our other RAIDs are running at (which do not experience this problem). During the upgrade process, all access to the disks were lost, resulting in (expected) I/O errors. So the crashes on Tuesday are understandable (though I am glad we won't be running into them in the future once we upgrade).

I have scanned through the controller logs and do not see any signs that there is some ongoing disk corruption problem (though it could be invisible). When the disks were yanked from lustre, could this have corrupted the file system to cause the following crashes? If so, would a lfsck clear up the problems?

Comment by Oleg Drokin [ 03/Aug/16 ]

From the messages it looks like the bitmaps got corrupted in your case.
Patches in bug LU-7114 should convert that from a fatal even into a non-fatal. Recent e2fsck should also fix that I believe.

Comment by Peter Jones [ 03/Aug/16 ]

Fan Yong

Is the fix from LU-7114 sufficient to address this issue? If so could you please port it to the 2.7 FE branch?

Thanks

Peter

Comment by nasf (Inactive) [ 04/Aug/16 ]

Originally, the case of the bitmap corruption is fatal, just like this ticket described, the system becomes unavailable. The patch from LU-7114 is NOT to fix bitmap corruption, instead, it marks the ground(s) with bad bitmap(s) and skip errors, then the system is still usable under such trouble case. I have back ported the patch to b2_7_fe (http://review.whamcloud.com/21705), but it just make the system to be available, the customer still needs to run e2fsck to fix the corrupted bitmap(s).

Comment by Herbert Yeung [ 09/Aug/16 ]

With the patch, will we still get error messages/warnings like below? Want to create a SEC rule to notify us if it happens again.

LDISKFS-fs error (device dm-27): ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 4891corrupted: 21438 blocks free in bitmap, 25535 - in gd

Comment by nasf (Inactive) [ 12/Aug/16 ]

With the patch applied, there may be still some message. But the message will be warning (#define KERN_WARNING "<4>"), not error (#define KERN_CRIT "<2>").

Comment by nasf (Inactive) [ 24/Aug/16 ]

Any feedback? Thanks!

Comment by Nathan Dauchy (Inactive) [ 29/Aug/16 ]

What kind of feedback were you looking for?

We did run e2fsck on the related servers and it corrected some errors.

It looks like the backport patch requires review and landing to the 2.7 FE branch.

Comment by nasf (Inactive) [ 29/Aug/16 ]

I mean that have you tried the patch? If yes, any results? Besides, any others you want us to do?

Comment by Nathan Dauchy (Inactive) [ 30/Aug/16 ]

We have not tried the patch at NASA yet, were waiting on completed review and testing and landing. Since we fixed the on-disk corruption with e2fsck, I don't believe we have a test case for it anymore anyway. As you described it seems like it will be sufficient though, so please proceed with getting it landed in the 2.7 FE branch and we will pull it into a future build. Thanks!

Comment by nasf (Inactive) [ 26/Nov/16 ]

http://review.whamcloud.com/#/c/21705/ has been landed to b2_7_fe.

Generated at Sat Feb 10 02:17:45 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.