Details
-
Bug
-
Resolution: Cannot Reproduce
-
Minor
-
None
-
Lustre 2.13.0
-
None
-
3
-
9223372036854775807
Description
conf-sanity test_111 fails with ‘add mds1 failed with new params’. Looking at the client test_log, we see that the problem is with the --mkfsoptions flag in the mkfs command
CMD: trevis-18vm11 mkfs.lustre --mgs --fsname=lustre --mdt --index=0 --param=sys.timeout=20 --param=mdt.identity_upcall=/usr/sbin/l_getidentity --backfstype=ldiskfs --device-size=2400000 --mkfsoptions=\"-O large_dir -i 1048576 -O ea_inode -E lazy_itable_init\" --reformat /dev/mapper/mds1_flakey trevis-18vm11: mkfs.lustre: don't specify multiple -O options trevis-18vm11: trevis-18vm11: mkfs.lustre FATAL: mkfs failed 22 trevis-18vm11: mkfs.lustre: exiting with 22 (Invalid argument) Permanent disk data: Target: lustre:MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x65 (MDT MGS first_time update ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: sys.timeout=20 mdt.identity_upcall=/usr/sbin/l_getidentity device size = 2048MB conf-sanity test_111: @@@@@@ FAIL: add mds1 failed with new params
It looks like Lustre does not like having multiple ‘-O’ flags in the --mkfsoptions. This may be an issue with the test and forming the --mkfsoptions list.
Looking in the client test_log, we also see that test_115 suffers from this same issue, but the test is skipped due to this issue
== conf-sanity test 115: Access large xattr with inodes number over 2TB ============================== 02:17:13 (1559787433) Stopping clients: trevis-18vm4.trevis.whamcloud.com,trevis-18vm5 /mnt/lustre (opts:) CMD: trevis-18vm4.trevis.whamcloud.com,trevis-18vm5 running=\$(grep -c /mnt/lustre' ' /proc/mounts); if [ \$running -ne 0 ] ; then echo Stopping client \$(hostname) /mnt/lustre opts:; … CMD: trevis-18vm11 mkfs.lustre --mgsnode=trevis-18vm11@tcp --fsname=lustre --mdt --index=0 --param=sys.timeout=20 --param=mdt.identity_upcall=/usr/sbin/l_getidentity --backfstype=ldiskfs --device-size=200000 --mkfsoptions=\"-O ea_inode -E lazy_itable_init\" --device-size=3298534883328 --mkfsoptions='-O lazy_itable_init,ea_inode,^resize_inode,meta_bg -i 1024' --mgs --reformat /dev/loop0 trevis-18vm11: trevis-18vm11: mkfs.lustre FATAL: Unable to build fs /dev/loop0 (256) trevis-18vm11: trevis-18vm11: mkfs.lustre FATAL: mkfs failed 256 Permanent disk data: Target: lustre:MDT0000 Index: 0 Lustre FS: lustre Mount type: ldiskfs Flags: 0x65 (MDT MGS first_time update ) Persistent mount opts: user_xattr,errors=remount-ro Parameters: mgsnode=10.9.4.221@tcp sys.timeout=20 mdt.identity_upcall=/usr/sbin/l_getidentity device size = 3145728MB formatting backing filesystem ldiskfs on /dev/loop0 target name lustre:MDT0000 kilobytes 3221225472 options -i 1024 -J size=4096 -I 1024 -q -O lazy_itable_init,ea_inode,^resize_inode,meta_bg,dirdata,uninit_bg,^extents,dir_nlink,quota,huge_file,flex_bg -E lazy_journal_init -F mkfs_cmd = mke2fs -j -b 4096 -L lustre:MDT0000 -i 1024 -J size=4096 -I 1024 -q -O lazy_itable_init,ea_inode,^resize_inode,meta_bg,dirdata,uninit_bg,^extents,dir_nlink,quota,huge_file,flex_bg -E lazy_journal_init -F /dev/loop0 3221225472k Invalid filesystem option set: lazy_itable_init,ea_inode,^resize_inode,meta_bg,dirdata,uninit_bg,^extents,dir_nlink,quota,huge_ file,flex_bg SKIP: conf-sanity test_115 format large MDT failed
It looks like this started failing on 2019-06-03 with Lustre 2.12.53.104. Logs for failed test sessions with this test failure are at
https://testing.whamcloud.com/test_sets/6b65b0f4-8695-11e9-869c-52540065bddc
https://testing.whamcloud.com/test_sets/34c0a1a0-8698-11e9-af1f-52540065bddc
https://testing.whamcloud.com/test_sets/efc1357e-8895-11e9-8c65-52540065bddc