Details
-
Bug
-
Resolution: Duplicate
-
Medium
-
None
-
None
-
None
-
3
-
9223372036854775807
Description
This issue was created by maloo for Arshad <arshad.hussain@aeoncomputing.com>
This issue relates to the following test suite run: https://testing.whamcloud.com/test_sets/6f38cd49-52f9-4612-80a3-1ab1a83aa906
Test session details:
clients: https://build.whamcloud.com/job/lustre-reviews/117942 - 5.14.0-503.40.1.el9_5.x86_64
servers: https://build.whamcloud.com/job/lustre-reviews/117942 - 5.14.0-503.40.1_lustre.el9.x86_64
<<Please provide additional information about the failure here>>
MDS
[ 4709.655725] Lustre: DEBUG MARKER: /usr/sbin/lctl set_param fail_loc=0x80001609^M
[ 4710.020654] Lustre: DEBUG MARKER: /usr/sbin/lctl get_param -n >>>>>>>>>>>>>>>mdd.lustre-MDT0000.lfsck_namespace |^M
[ 4710.020654] >>>>>>>>>awk '/^status/ { print $2 }'^M
[ 4710.238744] LustreError: 234943:0:(lfsck_engine.c:1607:lfsck_assistant_engine()) ASSERTION( lad->lad_post_^M
[ 4710.238744] LustreError: 234943:0:(lfsck_engine.c:1607:lfsck_assistant_engine()) ASSERTION( lad->lad_post_result > 0 ) failed: ^M
[ 4710.238841] LustreError: 234943:0:(lfsck_engine.c:1607:lfsck_assistant_engine()) LBUG^M
[ 4710.238931] CPU: 0 PID: 234943 Comm: lfsck_namespace Kdump: loaded Tainted: G OE ------- --- 5.14.0-503.40.1_lustre.el9.x86_64 #1^M
[ 4710.239003] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011^M
[ 4710.239071] Call Trace:^M
[ 4710.239212] <TASK>^M
[ 4710.239362] dump_stack_lvl+0x34/0x48^M
[ 4710.239625] lbug_with_loc.cold+0x5/0x43 [libcfs]^M
[ 4710.239853] lfsck_assistant_engine+0x14e0/0x16a0 [lfsck]^M
[ 4710.240188] ? _raw_spin_unlock_irqrestore+0xa/0x30^M
[ 4710.240358] ? __pfx_var_wake_function+0x10/0x10^M
[ 4710.240502] ? __pfx_lfsck_assistant_engine+0x10/0x10 [lfsck]^M
[ 4710.240624] kthread+0xe0/0x100^M
[ 4710.240719] ? __pfx_kthread+0x10/0x10^M
[ 4710.240848] ret_from_fork+0x2c/0x50^M
[ 4710.240962] </TASK>^M
[ 4710.241092] Kernel panic - not syncing: LBUG^M
Client
== sanity-lfsck test 6b: LFSCK resumes from last checkpoint (2) ========================================================== 17:46:12 (1761587172)
preparing... 5 * 5 files will be created Mon Oct 27 05:46:13 PM UTC 2025.
total: 5 mkdir in 0.00 seconds: 1766.77 ops/second
total: 5 create in 0.00 seconds: 1898.22 ops/second
total: 5 mkdir in 0.00 seconds: 1774.99 ops/second
prepared Mon Oct 27 05:46:13 PM UTC 2025.
CMD: trevis-44vm7 /usr/sbin/lctl set_param fail_val=1 fail_loc=0x1601
fail_val=1
fail_loc=0x1601
CMD: trevis-44vm7 /usr/sbin/lctl lfsck_start -M lustre-MDT0000 -t namespace -r
Started LFSCK on the device lustre-MDT0000: scrub namespace
CMD: trevis-44vm7 /usr/sbin/lctl get_param -n mdd.lustre-MDT0000.lfsck_namespace
CMD: trevis-44vm7 /usr/sbin/lctl set_param fail_loc=0x80001609
fail_loc=0x80001609
CMD: trevis-44vm7 /usr/sbin/lctl get_param -n mdd.lustre-MDT0000.lfsck_namespace |
awk '/^status/ { print \$2 }'
Waiting 32s for 'failed'
CMD: trevis-44vm7 /usr/sbin/lctl get_param -n mdd.lustre-MDT0000.lfsck_namespace |
awk '/^status/ { print \$2 }'
trevis-44vm7: ssh: connect to host trevis-44vm7 port 22: Connection refused
pdsh@trevis-44vm1: trevis-44vm7: ssh exited with exit code 255
CMD: trevis-44vm7 /usr/sbin/lctl get_param -n mdd.lustre-MDT0000.lfsck_namespace |
awk '/^status/ { print \$2 }'
trevis-44vm7: ssh: connect to host trevis-44vm7 port 22: Connection refused
pdsh@trevis-44vm1: trevis-44vm7: ssh exited with exit code 255
CMD: trevis-44vm7 /usr/sbin/lctl get_param -n mdd.lustre-MDT0000.lfsck_namespace |
awk '/^status/ { print \$2 }'
Attachments
Issue Links
- duplicates
-
LU-19505 sanity-lfsck: test 6b fails on ASSERTION( lad->lad_post_result > 0)
-
- Resolved
-