Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-9363

mdt_open.c:310:mdt_prep_ma_buf_from_rep()) ASSERTION( ma->ma_lmv == ((void *)0) && ma->ma_lmm == ((void *)0) ) failed

    XMLWordPrintable

Details

    • Bug
    • Resolution: Duplicate
    • Major
    • None
    • None
    • None
    • 3
    • 9223372036854775807

    Description

      Had this happen yesterday testing master + a whole bunch of patches that I am going to land to master shortly.
      First time to hit this assertion according to jira.

      [141250.438757] Lustre: DEBUG MARKER: == replay-single test 34: abort recovery before client does replay (test mds_cleanup_orphans) ======== 17:51:18 (1492465878)
      [141250.778639] Turning device loop0 (0x700000) read-only
      [141250.807567] Lustre: DEBUG MARKER: mds1 REPLAY BARRIER on lustre-MDT0000
      [141250.813942] Lustre: DEBUG MARKER: local REPLAY BARRIER on lustre-MDT0000
      [141251.024752] Removing read-only on unknown block (0x700000)
      [141252.084549] LDISKFS-fs (loop0): mounted filesystem with ordered data mode. Opts: user_xattr,errors=remount-ro,no_mbcache,nodelalloc
      [141252.498259] LustreError: 3776:0:(mdt_handler.c:6065:mdt_iocontrol()) lustre-MDT0000: Aborting recovery for device
      [141252.505283] LustreError: 3776:0:(ldlm_lib.c:2603:target_stop_recovery_thread()) lustre-MDT0000: Aborting recovery
      [141252.506489] Lustre: 3824:0:(ldlm_lib.c:2056:target_recovery_overseer()) recovery is aborted, evict exports in recovery
      [141252.508782] Lustre: 3824:0:(ldlm_lib.c:2056:target_recovery_overseer()) Skipped 2 previous similar messages
      [141252.509890] Lustre: lustre-MDT0000: disconnecting 1 stale clients
      [141252.555800] Lustre: lustre-OST0001: deleting orphan objects from 0x0:1447 to 0x0:1473
      [141252.555805] Lustre: lustre-OST0000: deleting orphan objects from 0x0:1448 to 0x0:1473
      [141252.558108] LustreError: 3797:0:(mdt_open.c:310:mdt_prep_ma_buf_from_rep()) ASSERTION( ma->ma_lmv == ((void *)0) && ma->ma_lmm == ((void *)0) ) failed: 
      [141252.558110] LustreError: 3797:0:(mdt_open.c:310:mdt_prep_ma_buf_from_rep()) LBUG
      [141252.558111] Pid: 3797, comm: mdt00_002
      [141252.558112] 
      Call Trace:
      [141252.558124]  [<ffffffffa01ce7ce>] libcfs_call_trace+0x4e/0x60 [libcfs]
      [141252.558129]  [<ffffffffa01ce85c>] lbug_with_loc+0x4c/0xb0 [libcfs]
      [141252.558145]  [<ffffffffa0c31847>] mdt_prep_ma_buf_from_rep.isra.31+0xe7/0xf0 [mdt]
      [141252.558156]  [<ffffffffa0c391b7>] mdt_reint_open+0x2287/0x3270 [mdt]
      [141252.558184]  [<ffffffffa039069c>] ? upcall_cache_get_entry+0x3ec/0x8e0 [obdclass]
      [141252.558194]  [<ffffffffa0c1ea83>] ? ucred_set_jobid+0x53/0x70 [mdt]
      [141252.558203]  [<ffffffffa0c2e2c0>] mdt_reint_rec+0x80/0x210 [mdt]
      [141252.558212]  [<ffffffffa0c10e9a>] mdt_reint_internal+0x58a/0x960 [mdt]
      [141252.558220]  [<ffffffffa0c113d2>] mdt_intent_reint+0x162/0x420 [mdt]
      [141252.558229]  [<ffffffffa0c1ba5e>] mdt_intent_policy+0x41e/0xbf0 [mdt]
      [141252.558264]  [<ffffffffa05560ea>] ldlm_lock_enqueue+0x35a/0x940 [ptlrpc]
      [141252.558272]  [<ffffffffa01df4b5>] ? cfs_hash_rw_unlock+0x15/0x20 [libcfs]
      [141252.558301]  [<ffffffffa057f833>] ldlm_handle_enqueue0+0x9e3/0x1670 [ptlrpc]
      [141252.558331]  [<ffffffffa05a78c0>] ? lustre_swab_ldlm_request+0x0/0x30 [ptlrpc]
      [141252.558367]  [<ffffffffa06014c2>] tgt_enqueue+0x62/0x210 [ptlrpc]
      [141252.558399]  [<ffffffffa0605905>] tgt_request_handle+0x925/0x1330 [ptlrpc]
      [141252.558427]  [<ffffffffa05b0aa1>] ptlrpc_server_handle_request+0x231/0xac0 [ptlrpc]
      [141252.558455]  [<ffffffffa05ae658>] ? ptlrpc_wait_event+0xb8/0x370 [ptlrpc]
      [141252.558482]  [<ffffffffa05b4838>] ptlrpc_main+0xa58/0x1dd0 [ptlrpc]
      [141252.558487]  [<ffffffff81706487>] ? _raw_spin_unlock_irq+0x27/0x50
      [141252.558514]  [<ffffffffa05b3de0>] ? ptlrpc_main+0x0/0x1dd0 [ptlrpc]
      [141252.558517]  [<ffffffff810a2eda>] kthread+0xea/0xf0
      [141252.558520]  [<ffffffff810a2df0>] ? kthread+0x0/0xf0
      [141252.558523]  [<ffffffff8170fbd8>] ret_from_fork+0x58/0x90
      [141252.558525]  [<ffffffff810a2df0>] ? kthread+0x0/0xf0
      [141252.558525] 
      [141252.558526] Kernel panic - not syncing: LBUG
      

      Crashdump: /exports/crashdumps/192.168.10.221-2017-04-17-17:51:28
      tag in my tree master-20170417

      Attachments

        Issue Links

          Activity

            People

              wc-triage WC Triage
              green Oleg Drokin
              Votes:
              0 Vote for this issue
              Watchers:
              1 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: