Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-19530

sanity-lfsck: lfsck_assistant_engine()) ASSERTION( lad->lad_post_result > 0 ) failed

    XMLWordPrintable

Details

    • Bug
    • Resolution: Duplicate
    • Medium
    • None
    • None
    • None
    • 3
    • 9223372036854775807

    Description

      This issue was created by maloo for Arshad <arshad.hussain@aeoncomputing.com>

      This issue relates to the following test suite run: https://testing.whamcloud.com/test_sets/6f38cd49-52f9-4612-80a3-1ab1a83aa906

      Test session details:
      clients: https://build.whamcloud.com/job/lustre-reviews/117942 - 5.14.0-503.40.1.el9_5.x86_64
      servers: https://build.whamcloud.com/job/lustre-reviews/117942 - 5.14.0-503.40.1_lustre.el9.x86_64

      <<Please provide additional information about the failure here>>

      MDS

      [ 4709.655725] Lustre: DEBUG MARKER: /usr/sbin/lctl set_param fail_loc=0x80001609^M
      [ 4710.020654] Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n >>>>>>>>>>>>>>>mdd.lustre-MDT0000.lfsck_namespace |^M
      [ 4710.020654] >>>>>>>>>awk '/^status/ { print $2 }'^M                          
      [ 4710.238744] LustreError: 234943:0:(lfsck_engine.c:1607:lfsck_assistant_engine()) ASSERTION( lad->lad_post_^M
      [ 4710.238744] LustreError: 234943:0:(lfsck_engine.c:1607:lfsck_assistant_engine()) ASSERTION( lad->lad_post_result > 0 ) failed: ^M
      [ 4710.238841] LustreError: 234943:0:(lfsck_engine.c:1607:lfsck_assistant_engine()) LBUG^M
      [ 4710.238931] CPU: 0 PID: 234943 Comm: lfsck_namespace Kdump: loaded Tainted: G           OE     -------  ---  5.14.0-503.40.1_lustre.el9.x86_64 #1^M
      [ 4710.239003] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011^M              
      [ 4710.239071] Call Trace:^M                                                    
      [ 4710.239212]  <TASK>^M                                                        
      [ 4710.239362]  dump_stack_lvl+0x34/0x48^M                                      
      [ 4710.239625]  lbug_with_loc.cold+0x5/0x43 [libcfs]^M                          
      [ 4710.239853]  lfsck_assistant_engine+0x14e0/0x16a0 [lfsck]^M                  
      [ 4710.240188]  ? _raw_spin_unlock_irqrestore+0xa/0x30^M                        
      [ 4710.240358]  ? __pfx_var_wake_function+0x10/0x10^M                           
      [ 4710.240502]  ? __pfx_lfsck_assistant_engine+0x10/0x10 [lfsck]^M              
      [ 4710.240624]  kthread+0xe0/0x100^M                                            
      [ 4710.240719]  ? __pfx_kthread+0x10/0x10^M                                     
      [ 4710.240848]  ret_from_fork+0x2c/0x50^M                                       
      [ 4710.240962]  </TASK>^M                                                       
      [ 4710.241092] Kernel panic - not syncing: LBUG^M     

      Client

      == sanity-lfsck test 6b: LFSCK resumes from last checkpoint (2) ========================================================== 17:46:12 (1761587172)
      preparing... 5 * 5 files will be created Mon Oct 27 05:46:13 PM UTC 2025.
      total: 5 mkdir in 0.00 seconds: 1766.77 ops/second
      total: 5 create in 0.00 seconds: 1898.22 ops/second
      total: 5 mkdir in 0.00 seconds: 1774.99 ops/second
      prepared Mon Oct 27 05:46:13 PM UTC 2025.
      CMD: trevis-44vm7 /usr/sbin/lctl set_param fail_val=1 fail_loc=0x1601
      fail_val=1
      fail_loc=0x1601
      CMD: trevis-44vm7 /usr/sbin/lctl lfsck_start -M lustre-MDT0000 -t namespace -r
      Started LFSCK on the device lustre-MDT0000: scrub namespace
      CMD: trevis-44vm7 /usr/sbin/lctl get_param -n mdd.lustre-MDT0000.lfsck_namespace
      CMD: trevis-44vm7 /usr/sbin/lctl set_param fail_loc=0x80001609
      fail_loc=0x80001609
      CMD: trevis-44vm7 /usr/sbin/lctl get_param -n         mdd.lustre-MDT0000.lfsck_namespace |
              awk '/^status/ { print \$2 }'
      Waiting 32s for 'failed'
      CMD: trevis-44vm7 /usr/sbin/lctl get_param -n         mdd.lustre-MDT0000.lfsck_namespace |
              awk '/^status/ { print \$2 }'
      trevis-44vm7: ssh: connect to host trevis-44vm7 port 22: Connection refused
      pdsh@trevis-44vm1: trevis-44vm7: ssh exited with exit code 255
      CMD: trevis-44vm7 /usr/sbin/lctl get_param -n         mdd.lustre-MDT0000.lfsck_namespace |
              awk '/^status/ { print \$2 }'
      trevis-44vm7: ssh: connect to host trevis-44vm7 port 22: Connection refused
      pdsh@trevis-44vm1: trevis-44vm7: ssh exited with exit code 255
      CMD: trevis-44vm7 /usr/sbin/lctl get_param -n         mdd.lustre-MDT0000.lfsck_namespace |
              awk '/^status/ { print \$2 }'

      Attachments

        Issue Links

          Activity

            People

              wc-triage WC Triage
              maloo Maloo
              Votes:
              0 Vote for this issue
              Watchers:
              1 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: