[LU-15633] OSS hit txg_sync hung and then OOM during recovery Created: 08/Mar/22  Updated: 08/Mar/22

Status: Open
Project: Lustre
Component/s: None
Affects Version/s: Lustre 2.15.0
Fix Version/s: None

Type: Bug Priority: Major
Reporter: Sarah Liu Assignee: WC Triage
Resolution: Unresolved Votes: 0
Labels: soak
Environment:

soak is running lustre-master-ib build #647 tag-2.15.0_RC2_2_g94f4e1f


Attachments: Text File soak-7.log    
Severity: 3
Rank (Obsolete): 9223372036854775807

 Description   

After switch soak to master-ib with EL7.9, mofed 4.9, OSS crashed consistently.

When doing oss failover, during recovery, the paired OSS hit following error

2022-03-07 23:42:50,054:fsmgmt.fsmgmt:INFO     ... soaked-OST0014 mounted successfully on soak-7
2022-03-07 23:42:50,055:fsmgmt.fsmgmt:INFO     ... soaked-OST0014 failed over
2022-03-07 23:42:50,055:fsmgmt.fsmgmt:INFO     Wait for recovery to complete
2022-03-07 23:42:50,949:fsmgmt.fsmgmt:DEBUG    Recovery Result Record: {'soak-7': {'soaked-OST000b': 'INACTIVE', 'soaked-OST000a': 'RECOVERING', 'soa
ked-OST000f': 'INACTIVE', 'soaked-OST000e': 'RECOVERING', 'soaked-OST0002': 'COMPLETE', 'soaked-OST0003': 'INACTIVE', 'soaked-OST0006': 'RECOVERING',
 'soaked-OST0007': 'INACTIVE'}}
2022-03-07 23:42:50,950:fsmgmt.fsmgmt:INFO     Node soak-7: 'soaked-OST0002' recovery completed
2022-03-07 23:42:50,950:fsmgmt.fsmgmt:INFO     soaked-OST000a in status 'RECOVERING'.
2022-03-07 23:42:50,950:fsmgmt.fsmgmt:INFO     soaked-OST0006 in status 'RECOVERING'.
2022-03-07 23:42:50,951:fsmgmt.fsmgmt:INFO     soaked-OST000e in status 'RECOVERING'.
2022-03-07 23:42:50,951:fsmgmt.fsmgmt:INFO     Next recovery check in 15s...
...
2022-03-07 23:43:03,746:sched.sched :DEBUG    JobID: 1074530 type: kcompile #tasks: 1 timelimit: 05:00:00
cmd: kcompile.sh /mnt/soaked/soaktest/test
2022-03-07 23:44:38,215:fsmgmt.fsmgmt:DEBUG    Recovery Result Record: {'soak-7': 'error: 255'}
2022-03-07 23:44:38,215:fsmgmt.fsmgmt:ERROR    Node 'soak-7', no recovery status information available for soaked-OST000a.
2022-03-07 23:44:38,215:fsmgmt.fsmgmt:ERROR    Node 'soak-7', no recovery status information available for soaked-OST0006.
2022-03-07 23:44:38,216:fsmgmt.fsmgmt:ERROR    Node 'soak-7', no recovery status information available for soaked-OST000e.
2022-03-07 23:44:38,216:fsmgmt.fsmgmt:INFO     Next recovery check in 15s...

on soak-7, console shows

[19313.048562] Lustre: soaked-OST000e: deleting orphan objects from 0x680000400:3566 to 0x680000400:5057
[19440.127775] INFO: task txg_sync:15186 blocked for more than 120 seconds.
[19440.135262] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[19440.144004] txg_sync        D ffff8c21ebfb9080     0 15186      2 0x00000080
[19440.151895] Call Trace:
[19440.154633]  [<ffffffffb7d89179>] schedule+0x29/0x70
[19440.160179]  [<ffffffffb7d86d88>] schedule_timeout+0x168/0x2d0
[19440.166693]  [<ffffffffb76ad8b0>] ? requeue_timers+0x170/0x170
[19440.173208]  [<ffffffffb7706992>] ? ktime_get_ts64+0x52/0xf0
[19440.179526]  [<ffffffffb7d88a2d>] io_schedule_timeout+0xad/0x130
[19440.186244]  [<ffffffffc069c5cd>] __cv_timedwait_common+0x13d/0x190 [spl]
[19440.193829]  [<ffffffffb76c6f50>] ? wake_up_atomic_t+0x30/0x30
[19440.200346]  [<ffffffffc069c659>] __cv_timedwait_io+0x19/0x20 [spl]
[19440.207391]  [<ffffffffc1876eb3>] zio_wait+0x143/0x290 [zfs]
[19440.213746]  [<ffffffffc17cf30f>] dsl_pool_sync+0xdf/0x500 [zfs]
[19440.220484]  [<ffffffffc1811e30>] ? spa_suspend_async_destroy+0x60/0x60 [zfs]
[19440.228481]  [<ffffffffc17fdab3>] spa_sync+0x5c3/0x10e0 [zfs]
[19440.234928]  [<ffffffffc181591b>] ? spa_txg_history_init_io+0x10b/0x120 [zfs]
[19440.242936]  [<ffffffffc181a00f>] txg_sync_thread+0x2cf/0x4d0 [zfs]
[19440.249964]  [<ffffffffc1819d40>] ? txg_fini+0x280/0x280 [zfs]
[19440.256485]  [<ffffffffc06a3873>] thread_generic_wrapper+0x73/0x80 [spl]
[19440.263969]  [<ffffffffc06a3800>] ? __thread_exit+0x20/0x20 [spl]
[19440.270774]  [<ffffffffb76c5e61>] kthread+0xd1/0xe0
[19440.276221]  [<ffffffffb76c5d90>] ? insert_kthread_work+0x40/0x40
[19440.283028]  [<ffffffffb7d95df7>] ret_from_fork_nospec_begin+0x21/0x21
[19440.290317]  [<ffffffffb76c5d90>] ? insert_kthread_work+0x40/0x40
[19469.327166] Lustre: soaked-OST0002: Client dd15d031-530f-484f-875a-a585acc53921 (at 192.168.1.125@o2ib) reconnecting
[19469.848067] Lustre: soaked-OST0003: Client bd579c0e-c6d9-4026-8251-a17bb20e4467 (at 192.168.1.119@o2ib) reconnecting
[19469.859817] Lustre: Skipped 5 previous similar messages
[19524.058650] Lustre: ll_ost_io01_037: service thread pid 23417 was inactive for 200.092 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one.
[19524.058652] Lustre: ll_ost_io00_087: service thread pid 23495 was inactive for 200.087 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
[19524.058654] Lustre: ll_ost_io00_068: service thread pid 23475 was inactive for 200.015 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
[19524.058655] Lustre: ll_ost_io00_037: service thread pid 23405 was inactive for 200.090 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
[19524.058658] Pid: 23495, comm: ll_ost_io00_087 3.10.0-1160.49.1.el7_lustre.x86_64 #1 SMP Mon Dec 13 22:00:59 UTC 2021
[19524.058660] Call Trace:
[19524.058685] [<0>] cv_wait_common+0x125/0x150 [spl]
[19524.058688] [<0>] __cv_wait+0x15/0x20 [spl]
[19524.058729] [<0>] dbuf_read+0x286/0x570 [zfs]
[19524.058748] [<0>] dbuf_hold_impl+0x4fd/0x5d0 [zfs]
[19524.058767] [<0>] dbuf_hold_impl+0x4e2/0x5d0 [zfs]
[19524.058785] [<0>] dbuf_hold_level+0x35/0x60 [zfs]
[19524.058803] [<0>] dbuf_hold+0x16/0x20 [zfs]
[19524.058823] [<0>] dmu_buf_hold_array_by_dnode+0xee/0x4e0 [zfs]
[19524.058843] [<0>] dmu_buf_hold_array_by_bonus+0x69/0x90 [zfs]
[19524.058852] [<0>] osd_bufs_get+0x7c0/0xfb0 [osd_zfs]
[19524.058864] [<0>] ofd_preprw+0x801/0x18d0 [ofd]
[19524.058926] [<0>] tgt_brw_read+0x9d4/0x2040 [ptlrpc]
[19524.058956] [<0>] tgt_request_handle+0x92f/0x19c0 [ptlrpc]
[19524.058984] [<0>] ptlrpc_server_handle_request+0x253/0xc30 [ptlrpc]
[19524.059010] [<0>] ptlrpc_main+0xbf4/0x15e0 [ptlrpc]
[19524.059015] [<0>] kthread+0xd1/0xe0
[19524.059018] [<0>] ret_from_fork_nospec_begin+0x21/0x21
[19524.059033] [<0>] 0xfffffffffffffffe
[19524.059034] Pid: 23475, comm: ll_ost_io00_068 3.10.0-1160.49.1.el7_lustre.x86_64 #1 SMP Mon Dec 13 22:00:59 UTC 2021
[19524.059034] Call Trace:
...

[19581.563091] Lustre: Skipped 2 previous similar messages
[19581.869624] Lustre: soaked-OST000b: Bulk IO read error with b5d7b2be-72fd-4f21-9635-1a800193ea85 (at 192.168.1.121@o2ib), client will retry: rc -107
[19583.779078] obd_memory max: 533504839, obd_memory current: 532366911
[19583.786625] kthreadd invoked oom-killer: gfp_mask=0x3000d0, order=2, oom_score_adj=0
[19583.795274] kthreadd cpuset=/ mems_allowed=0-1
[19583.800235] CPU: 14 PID: 2 Comm: kthreadd Kdump: loaded Tainted: P           OE  ------------   3.10.0-1160.49.1.el7_lustre.x86_64 #1
[19583.813622] Hardware name: Intel Corporation S2600GZ ........../S2600GZ, BIOS SE5C600.86B.01.08.0003.022620131521 02/26/2013
[19583.826137] Call Trace:
[19583.828872]  [<ffffffffb7d83539>] dump_stack+0x19/0x1b
[19583.834609]  [<ffffffffb7d7e5d8>] dump_header+0x90/0x229
[19583.840540]  [<ffffffffb7d7d3d8>] ? printk+0x60/0x77
[19583.846085]  [<ffffffffb77c254d>] oom_kill_process+0x2cd/0x490
[19583.852596]  [<ffffffffb77c1f3d>] ? oom_unkillable_task+0xcd/0x120
[19583.859498]  [<ffffffffb77c2c3a>] out_of_memory+0x31a/0x500
[19583.865720]  [<ffffffffb77c9854>] __alloc_pages_nodemask+0xad4/0xbe0
[19583.872817]  [<ffffffffb76c5d90>] ? insert_kthread_work+0x40/0x40
[19583.879624]  [<ffffffffb769882d>] copy_process+0x1dd/0x1a80
[19583.885846]  [<ffffffffb769a281>] do_fork+0x91/0x330
[19583.891390]  [<ffffffffb7d88df0>] ? __schedule+0x320/0x680
[19583.897516]  [<ffffffffb769a546>] kernel_thread+0x26/0x30
[19583.903548]  [<ffffffffb76c6841>] kthreadd+0x2e1/0x320
[19583.909286]  [<ffffffffb76c6560>] ? kthread_create_on_cpu+0x60/0x60
[19583.916287]  [<ffffffffb7d95df7>] ret_from_fork_nospec_begin+0x21/0x21
[19583.923575]  [<ffffffffb76c6560>] ? kthread_create_on_cpu+0x60/0x60
[19583.930571] Mem-Info:
...

Generated at Sat Feb 10 03:20:00 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.