Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-5719

target_queue_recovery_request() ASSERTION failed

Details

    • Bug
    • Resolution: Duplicate
    • Major
    • Lustre 2.5.4
    • Lustre 2.5.2
    • None
    • RHEL 6.5, kernel 2.6.32-431.17.1.el6
    • 3
    • 16046

    Description

      Opening this issue to track our lineage, but we basically hit LU-5651 in production today.

      Attachments

        Issue Links

          Activity

            [LU-5719] target_queue_recovery_request() ASSERTION failed
            yujian Jian Yu added a comment -

            Close this ticket as a duplicate of LU-5651.

            yujian Jian Yu added a comment - Close this ticket as a duplicate of LU-5651 .
            yujian Jian Yu added a comment -

            Hi Jason, can we close this ticket now?

            yujian Jian Yu added a comment - Hi Jason, can we close this ticket now?
            yujian Jian Yu added a comment -

            Patches for LU-5651 landed to master and b2_5 branches. The current b2_5 branch doesn't contain the patch for LU-793. Can we close this ticket?

            yujian Jian Yu added a comment - Patches for LU-5651 landed to master and b2_5 branches. The current b2_5 branch doesn't contain the patch for LU-793 . Can we close this ticket?
            jamesanunez James Nunez (Inactive) added a comment - - edited

            Thanks for the update and it's good to hear you're back up and running. I'll lower the priority of this ticket.

            jamesanunez James Nunez (Inactive) added a comment - - edited Thanks for the update and it's good to hear you're back up and running. I'll lower the priority of this ticket.

            The production down issue was resolved by removing LU-793. Please lower criticality. Thanks for the quick engagement and resolution- we had a power event that caused part of the cluster to go down which caused us to go through recovery and exposed this issue. ~6 hours of downtime on our first day in Lustre 2.5.2, but not the worst day in production so far.

            hilljjornl Jason Hill (Inactive) added a comment - The production down issue was resolved by removing LU-793 . Please lower criticality. Thanks for the quick engagement and resolution- we had a power event that caused part of the cluster to go down which caused us to go through recovery and exposed this issue. ~6 hours of downtime on our first day in Lustre 2.5.2, but not the worst day in production so far.
            pjones Peter Jones added a comment -

            Thanks Jason

            pjones Peter Jones added a comment - Thanks Jason

            Oleg and James Simmons talked and instead of incorporating the patch from LU-5651 we are removing the patch from LU-793. RPMs are generated and we are working to reboot the filesystem with these RPMs.


            -Jason

            hilljjornl Jason Hill (Inactive) added a comment - Oleg and James Simmons talked and instead of incorporating the patch from LU-5651 we are removing the patch from LU-793 . RPMs are generated and we are working to reboot the filesystem with these RPMs. – -Jason

            People

              green Oleg Drokin
              hilljjornl Jason Hill (Inactive)
              Votes:
              0 Vote for this issue
              Watchers:
              8 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: