Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-11125

ofd_create_hdl() destroys_in_progress already cleared

Details

    • Bug
    • Resolution: Fixed
    • Major
    • Lustre 2.12.0, Lustre 2.10.6
    • Lustre 2.10.4
    • lustre-2.10.4_1.chaos-1.ch6.x86_64 servers
      RHEL 7.5
      DNE1 file system
    • 3
    • 9223372036854775807

    Description

      Servers were restarted and appeared to recover normally.  They briefly appeared to be handling the same (heavy) workload from before they were powered off, then started logging the "system was overloaded" message.  The kernel then reported several stacks like this:

      INFO: task ll_ost00_007:108440 blocked for more than 120 seconds.
      "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
      ll_ost00_007 D ffff8ba4dc35bf40 0 108440 2 0x00000080
      Call Trace:
      [<ffffffffaad38919>] schedule_preempt_disabled+0x39/0x90
      [<ffffffffaad3654f>] __mutex_lock_slowpath+0x10f/0x250
      [<ffffffffaad357f2>] mutex_lock+0x32/0x42
      [<ffffffffc1669afb>] ofd_create_hdl+0xdcb/0x2090 [ofd]
      [<ffffffffc1322007>] ? lustre_msg_add_version+0x27/0xa0 [ptlrpc]
      [<ffffffffc132235f>] ? lustre_pack_reply_v2+0x14f/0x290 [ptlrpc]
      [<ffffffffc1322691>] ? lustre_pack_reply+0x11/0x20 [ptlrpc]
      [<ffffffffc138653a>] tgt_request_handle+0x92a/0x1370 [ptlrpc]
      [<ffffffffc132db5b>] ptlrpc_server_handle_request+0x23b/0xaa0 [ptlrpc]
      [<ffffffffc132b26b>] ? ptlrpc_wait_event+0xab/0x350 [ptlrpc]
      [<ffffffffaa6d5c32>] ? default_wake_function+0x12/0x20
      [<ffffffffaa6cb01b>] ? __wake_up_common+0x5b/0x90
      [<ffffffffc1331c70>] ptlrpc_main+0xae0/0x1e90 [ptlrpc]
      [<ffffffffc1331190>] ? ptlrpc_register_service+0xe30/0xe30 [ptlrpc]
      [<ffffffffaa6c0ad1>] kthread+0xd1/0xe0
      [<ffffffffaa6c0a00>] ? insert_kthread_work+0x40/0x40
      [<ffffffffaad44837>] ret_from_fork_nospec_begin+0x21/0x21
      [<ffffffffaa6c0a00>] ? insert_kthread_work+0x40/0x40

       And lustre began reporting:
      LustreError: 108448:0:(ofd_dev.c:1627:ofd_create_hdl()) lquake-OST0003:[27917288460] destroys_in_progress already cleared

       

      Attachments

        Issue Links

          Activity

            [LU-11125] ofd_create_hdl() destroys_in_progress already cleared
            ofaaland Olaf Faaland added a comment -

            Is there a ticket for reducing the contention on os_create_lock? Thanks.

            LU-11399

            ofaaland Olaf Faaland added a comment - Is there a ticket for reducing the contention on os_create_lock? Thanks. LU-11399

            Is there a ticket for reducing the contention on os_create_lock? Thanks.

            ofaaland Olaf Faaland added a comment - Is there a ticket for reducing the contention on os_create_lock? Thanks.

            John L. Hammond (jhammond@whamcloud.com) merged in patch https://review.whamcloud.com/33100/
            Subject: LU-11125 ofd: decrease message level
            Project: fs/lustre-release
            Branch: b2_10
            Current Patch Set:
            Commit: 483e2e745c5294ae6dc2b6f85cc8976eb9f90fa3

            gerrit Gerrit Updater added a comment - John L. Hammond (jhammond@whamcloud.com) merged in patch https://review.whamcloud.com/33100/ Subject: LU-11125 ofd: decrease message level Project: fs/lustre-release Branch: b2_10 Current Patch Set: Commit: 483e2e745c5294ae6dc2b6f85cc8976eb9f90fa3

            Minh Diep (mdiep@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/33100
            Subject: LU-11125 ofd: decrease message level
            Project: fs/lustre-release
            Branch: b2_10
            Current Patch Set: 1
            Commit: d266dbc4d62ab018d345159bb388f7f44fdca9d3

            gerrit Gerrit Updater added a comment - Minh Diep (mdiep@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/33100 Subject: LU-11125 ofd: decrease message level Project: fs/lustre-release Branch: b2_10 Current Patch Set: 1 Commit: d266dbc4d62ab018d345159bb388f7f44fdca9d3
            pjones Peter Jones added a comment -

            This first patch landed for 2.12. Mike, can you please link in the ticket tracking the subsequent improvement?

            pjones Peter Jones added a comment - This first patch landed for 2.12. Mike, can you please link in the ticket tracking the subsequent improvement?

            Oleg Drokin (green@whamcloud.com) merged in patch https://review.whamcloud.com/32985/
            Subject: LU-11125 ofd: decrease message level
            Project: fs/lustre-release
            Branch: master
            Current Patch Set:
            Commit: c25c4d60ac5fbede50a41da12f91de6e402b3c68

            gerrit Gerrit Updater added a comment - Oleg Drokin (green@whamcloud.com) merged in patch https://review.whamcloud.com/32985/ Subject: LU-11125 ofd: decrease message level Project: fs/lustre-release Branch: master Current Patch Set: Commit: c25c4d60ac5fbede50a41da12f91de6e402b3c68

            Mike Pershin (mpershin@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/32985
            Subject: LU-11125 ofd: decrease message level
            Project: fs/lustre-release
            Branch: master
            Current Patch Set: 1
            Commit: 68c740bcc319d019ec545eb50f25417a4bd9f7a9

            gerrit Gerrit Updater added a comment - Mike Pershin (mpershin@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/32985 Subject: LU-11125 ofd: decrease message level Project: fs/lustre-release Branch: master Current Patch Set: 1 Commit: 68c740bcc319d019ec545eb50f25417a4bd9f7a9

            Olaf, this message is not an error so it can be just made less noisy as first step and then I will create new improvement ticket to investigate this area thoughtfully to lower contention on that mutex during OST recovery

            tappro Mikhail Pershin added a comment - Olaf, this message is not an error so it can be just made less noisy as first step and then I will create new improvement ticket to investigate this area thoughtfully to lower contention on that mutex during OST recovery
            ofaaland Olaf Faaland added a comment -

            Hello Mikhail, have you learned anything more of interest?  Thanks.

            ofaaland Olaf Faaland added a comment - Hello Mikhail, have you learned anything more of interest?  Thanks.

            It still may indicate the problem there, there are two different operations are possible on OST restart - MDT initiates orphan destroys, but also new precreates may be initiated by write replays. Both uses the same mutex 'os_create_lock' though one process destroys non-used objects and other creates new objects. That causes high contention on that mutex exactly on OST restart and I am checking now is that possible to separate locking for both processes. That would relieve load on that mutex.

            tappro Mikhail Pershin added a comment - It still may indicate the problem there, there are two different operations are possible on OST restart - MDT initiates orphan destroys, but also new precreates may be initiated by write replays. Both uses the same mutex 'os_create_lock' though one process destroys non-used objects and other creates new objects. That causes high contention on that mutex exactly on OST restart and I am checking now is that possible to separate locking for both processes. That would relieve load on that mutex.

            People

              tappro Mikhail Pershin
              ofaaland Olaf Faaland
              Votes:
              0 Vote for this issue
              Watchers:
              4 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: