Details
-
Bug
-
Resolution: Duplicate
-
Major
-
None
-
None
-
None
-
3
-
9223372036854775807
Description
Had this happen yesterday testing master + a whole bunch of patches that I am going to land to master shortly.
First time to hit this assertion according to jira.
[141250.438757] Lustre: DEBUG MARKER: == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ======== 17:51:18 (1492465878) [141250.778639] Turning device loop0 (0x700000) read-only [141250.807567] Lustre: DEBUG MARKER: mds1 REPLAY BARRIER on lustre-MDT0000 [141250.813942] Lustre: DEBUG MARKER: local REPLAY BARRIER on lustre-MDT0000 [141251.024752] Removing read-only on unknown block (0x700000) [141252.084549] LDISKFS-fs (loop0): mounted filesystem with ordered data mode. Opts: user_xattr,errors=remount-ro,no_mbcache,nodelalloc [141252.498259] LustreError: 3776:0:(mdt_handler.c:6065:mdt_iocontrol()) lustre-MDT0000: Aborting recovery for device [141252.505283] LustreError: 3776:0:(ldlm_lib.c:2603:target_stop_recovery_thread()) lustre-MDT0000: Aborting recovery [141252.506489] Lustre: 3824:0:(ldlm_lib.c:2056:target_recovery_overseer()) recovery is aborted, evict exports in recovery [141252.508782] Lustre: 3824:0:(ldlm_lib.c:2056:target_recovery_overseer()) Skipped 2 previous similar messages [141252.509890] Lustre: lustre-MDT0000: disconnecting 1 stale clients [141252.555800] Lustre: lustre-OST0001: deleting orphan objects from 0x0:1447 to 0x0:1473 [141252.555805] Lustre: lustre-OST0000: deleting orphan objects from 0x0:1448 to 0x0:1473 [141252.558108] LustreError: 3797:0:(mdt_open.c:310:mdt_prep_ma_buf_from_rep()) ASSERTION( ma->ma_lmv == ((void *)0) && ma->ma_lmm == ((void *)0) ) failed: [141252.558110] LustreError: 3797:0:(mdt_open.c:310:mdt_prep_ma_buf_from_rep()) LBUG [141252.558111] Pid: 3797, comm: mdt00_002 [141252.558112] Call Trace: [141252.558124] [<ffffffffa01ce7ce>] libcfs_call_trace+0x4e/0x60 [libcfs] [141252.558129] [<ffffffffa01ce85c>] lbug_with_loc+0x4c/0xb0 [libcfs] [141252.558145] [<ffffffffa0c31847>] mdt_prep_ma_buf_from_rep.isra.31+0xe7/0xf0 [mdt] [141252.558156] [<ffffffffa0c391b7>] mdt_reint_open+0x2287/0x3270 [mdt] [141252.558184] [<ffffffffa039069c>] ? upcall_cache_get_entry+0x3ec/0x8e0 [obdclass] [141252.558194] [<ffffffffa0c1ea83>] ? ucred_set_jobid+0x53/0x70 [mdt] [141252.558203] [<ffffffffa0c2e2c0>] mdt_reint_rec+0x80/0x210 [mdt] [141252.558212] [<ffffffffa0c10e9a>] mdt_reint_internal+0x58a/0x960 [mdt] [141252.558220] [<ffffffffa0c113d2>] mdt_intent_reint+0x162/0x420 [mdt] [141252.558229] [<ffffffffa0c1ba5e>] mdt_intent_policy+0x41e/0xbf0 [mdt] [141252.558264] [<ffffffffa05560ea>] ldlm_lock_enqueue+0x35a/0x940 [ptlrpc] [141252.558272] [<ffffffffa01df4b5>] ? cfs_hash_rw_unlock+0x15/0x20 [libcfs] [141252.558301] [<ffffffffa057f833>] ldlm_handle_enqueue0+0x9e3/0x1670 [ptlrpc] [141252.558331] [<ffffffffa05a78c0>] ? lustre_swab_ldlm_request+0x0/0x30 [ptlrpc] [141252.558367] [<ffffffffa06014c2>] tgt_enqueue+0x62/0x210 [ptlrpc] [141252.558399] [<ffffffffa0605905>] tgt_request_handle+0x925/0x1330 [ptlrpc] [141252.558427] [<ffffffffa05b0aa1>] ptlrpc_server_handle_request+0x231/0xac0 [ptlrpc] [141252.558455] [<ffffffffa05ae658>] ? ptlrpc_wait_event+0xb8/0x370 [ptlrpc] [141252.558482] [<ffffffffa05b4838>] ptlrpc_main+0xa58/0x1dd0 [ptlrpc] [141252.558487] [<ffffffff81706487>] ? _raw_spin_unlock_irq+0x27/0x50 [141252.558514] [<ffffffffa05b3de0>] ? ptlrpc_main+0x0/0x1dd0 [ptlrpc] [141252.558517] [<ffffffff810a2eda>] kthread+0xea/0xf0 [141252.558520] [<ffffffff810a2df0>] ? kthread+0x0/0xf0 [141252.558523] [<ffffffff8170fbd8>] ret_from_fork+0x58/0x90 [141252.558525] [<ffffffff810a2df0>] ? kthread+0x0/0xf0 [141252.558525] [141252.558526] Kernel panic - not syncing: LBUG
Crashdump: /exports/crashdumps/192.168.10.221-2017-04-17-17:51:28
tag in my tree master-20170417
Attachments
Issue Links
- is related to
-
LU-12914 (mdt_open.c:312:mdt_prep_ma_buf_from_rep()) ASSERTION( ma->ma_lmv == ((void *)0) && ma->ma_lmm == ((void *)0) ) failed
-
- Resolved
-