Details
-
Bug
-
Resolution: Fixed
-
Minor
-
Lustre 2.11.0, Lustre 2.10.3, Lustre 2.10.6, Lustre 2.10.7
-
None
-
3
-
9223372036854775807
Description
sanity-scrub test 9 fails with
sanity-scrub test_9: @@@@@@ FAIL: (9) Expected 'scanning' on mds3
From the test_log, the last thing we see is:
CMD: trevis-8vm4 /usr/sbin/lctl lfsck_start -M lustre-MDT0000 -t scrub 8 -s 100 -r Started LFSCK on the device lustre-MDT0000: scrub CMD: trevis-8vm8 /usr/sbin/lctl lfsck_start -M lustre-MDT0001 -t scrub 8 -s 100 -r Started LFSCK on the device lustre-MDT0001: scrub CMD: trevis-8vm4 /usr/sbin/lctl lfsck_start -M lustre-MDT0002 -t scrub 8 -s 100 -r Started LFSCK on the device lustre-MDT0002: scrub CMD: trevis-8vm8 /usr/sbin/lctl lfsck_start -M lustre-MDT0003 -t scrub 8 -s 100 -r Started LFSCK on the device lustre-MDT0003: scrub CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0000.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0000.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm8 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0001.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm8 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0001.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' Waiting 6 secs for update CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' CMD: trevis-8vm4 /usr/sbin/lctl get_param -n osd-ldiskfs.lustre-MDT0002.oi_scrub | awk '/^status/ { print \$2 }' Update not seen after 6s: wanted 'scanning' got 'completed' sanity-scrub test_9: @@@@@@ FAIL: (9) Expected 'scanning' on mds3
There are no obvious errors in the console and debug logs.
So far, this looks like it only impacts DNE Lustre configurations (reivew-dne-part-*)
This test started to fail on June 2, 2017.
Logs for this test failures are at:
https://testing.hpdd.intel.com/test_sets/f934f42a-772a-11e7-8e04-5254006e85c2
https://testing.hpdd.intel.com/test_sets/6112f68a-76f1-11e7-8db2-5254006e85c2
https://testing.hpdd.intel.com/test_sets/57e14740-7626-11e7-9a53-5254006e85c2
https://testing.hpdd.intel.com/test_sets/f429dfd0-733e-11e7-8d7d-5254006e85c2
https://testing.hpdd.intel.com/test_sets/cd0041d8-6eac-11e7-a055-5254006e85c2
https://testing.hpdd.intel.com/test_sets/200eed24-6d32-11e7-a052-5254006e85c2
https://testing.hpdd.intel.com/test_sets/76ef4bb6-6841-11e7-a74b-5254006e85c2
Attachments
Issue Links
- mentioned in
-
Page Loading...