Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-9408

Client fails to mount with ZFS master (0.7.0) and Lustre master (2.9.56)

Details

    • Bug
    • Resolution: Not a Bug
    • Critical
    • None
    • Lustre 2.10.0
    • 3
    • 9223372036854775807

    Description

      ZFS: zfs-0.7.0-rc3-225-g7a25f08
      SPL: spl-0.7.0-rc3-8-g481762f
      Lustre: v2_9_56_0-11-gbfa524f

      This is a straightforward single MDS with MDT/MGT and an OSS with a single OST.
      (also reproduced with split MDT and MGT on a slightly earlier ZFS).

      This setup works without problems with ldiskfs backing.

      Snippet from MDS log on initial mount:

      Apr 26 14:01:18 ieel-mds04 kernel: SPL: Loaded module v0.7.0-rc3_8_g481762f
      Apr 26 14:01:19 ieel-mds04 kernel: ZFS: Loaded module v0.7.0-rc3_225_g7a25f08, ZFS pool version 5000, ZFS filesystem version 5
      ...
      Apr 26 14:20:05 ieel-mds04 kernel: SPL: using hostid 0x7e3a4ec9
      Apr 26 14:20:52 ieel-mds04 kernel: Lustre: MGS: Connection restored to ec5ab9aa-e46a-19dd-47d3-1aa7d07fdc3f (at 0@lo)
      Apr 26 14:20:52 ieel-mds04 kernel: Lustre: srv-scratch-MDT0001: No data found on store. Initialize space
      Apr 26 14:20:52 ieel-mds04 kernel: Lustre: scratch-MDT0001: new disk, initializing
      Apr 26 14:20:52 ieel-mds04 kernel: Lustre: scratch-MDT0001: Imperative Recovery not enabled, recovery window 300-900
      Apr 26 14:20:52 ieel-mds04 kernel: LustreError: 21965:0:(osd_oi.c:497:osd_oid()) unsupported quota oid: 0x16
      Apr 26 14:20:52 ieel-mds04 kernel: LustreError: 22330:0:(fid_handler.c:329:__seq_server_alloc_meta()) srv-scratch-MDT0001: Allocated super-sequence failed: rc = -115
      Apr 26 14:20:52 ieel-mds04 kernel: LustreError: 22330:0:(fid_request.c:227:seq_client_alloc_seq()) cli-scratch-MDT0001: Can't allocate new meta-sequence,rc -115
      Apr 26 14:20:52 ieel-mds04 kernel: LustreError: 22330:0:(fid_request.c:383:seq_client_alloc_fid()) cli-scratch-MDT0001: Can't allocate new sequence: rc = -115
      Apr 26 14:20:52 ieel-mds04 kernel: LustreError: 22330:0:(lod_dev.c:419:lod_sub_recovery_thread()) scratch-MDT0001-osd getting update log failed: rc = -115
      

      OST Log on initial mount:

      Apr 26 14:20:56 ieel-oss03 kernel: SPL: Loaded module v0.7.0-rc3_8_g481762f
      Apr 26 14:20:58 ieel-oss03 kernel: ZFS: Loaded module v0.7.0-rc3_225_g7a25f08, ZFS pool version 5000, ZFS filesystem version 5
      Apr 26 14:21:08 ieel-oss03 kernel: SPL: using hostid 0x5d9bdb4b
      Apr 26 14:25:01 ieel-oss03 kernel: LNet: HW nodes: 1, HW CPU cores: 2, npartitions: 1
      Apr 26 14:25:01 ieel-oss03 kernel: alg: No test for adler32 (adler32-zlib)
      Apr 26 14:25:01 ieel-oss03 kernel: alg: No test for crc32 (crc32-table)
      Apr 26 14:25:01 ieel-oss03 kernel: Lustre: Lustre: Build Version: 2.9.56_11_gbfa524f
      Apr 26 14:25:01 ieel-oss03 kernel: LNet: Added LNI 192.168.56.22@tcp [8/256/0/180]
      Apr 26 14:25:01 ieel-oss03 kernel: LNet: Accept secure, port 988
      Apr 26 14:25:02 ieel-oss03 kernel: Lustre: scratch-OST0000: new disk, initializing
      Apr 26 14:25:02 ieel-oss03 kernel: Lustre: srv-scratch-OST0000: No data found on store. Initialize space
      Apr 26 14:25:02 ieel-oss03 kernel: Lustre: scratch-OST0000: Imperative Recovery not enabled, recovery window 300-900
      Apr 26 14:25:02 ieel-oss03 kernel: LustreError: 13214:0:(osd_oi.c:497:osd_oid()) unsupported quota oid: 0x16
      Apr 26 14:25:07 ieel-oss03 kernel: Lustre: scratch-OST0000: Connection restored to scratch-MDT0001-mdtlov_UUID (at 192.168.56.13@tcp)
      

      Client attempting to mount (messages):

      Apr 26 14:30:44 ieel-c03 kernel: LNet: HW CPU cores: 2, npartitions: 1
      Apr 26 14:30:44 ieel-c03 kernel: alg: No test for adler32 (adler32-zlib)
      Apr 26 14:30:44 ieel-c03 kernel: alg: No test for crc32 (crc32-table)
      Apr 26 14:30:49 ieel-c03 kernel: sha512_ssse3: Using AVX optimized SHA-512 implementation
      Apr 26 14:30:52 ieel-c03 kernel: Lustre: Lustre: Build Version: 2.8.0.51-1-PRISTINE-3.10.0-514.6.1.el7.x86_64
      Apr 26 14:30:52 ieel-c03 kernel: LNet: Added LNI 192.168.56.32@tcp [8/256/0/180]
      Apr 26 14:30:52 ieel-c03 kernel: LNet: Accept secure, port 988
      Apr 26 14:30:52 ieel-c03 kernel: LustreError: 2336:0:(lmv_obd.c:553:lmv_check_connect()) scratch-clilmv-ffff88003b98d800: no target configured for index 0.
      Apr 26 14:30:52 ieel-c03 kernel: LustreError: 2336:0:(llite_lib.c:265:client_common_fill_super()) cannot connect to scratch-clilmv-ffff88003b98d800: rc = -22
      Apr 26 14:30:52 ieel-c03 kernel: LustreError: 2363:0:(lov_obd.c:922:lov_cleanup()) scratch-clilov-ffff88003b98d800: lov tgt 0 not cleaned! deathrow=0, lovrc=1
      Apr 26 14:30:52 ieel-c03 kernel: Lustre: Unmounted scratch-client
      Apr 26 14:30:52 ieel-c03 kernel: LustreError: 2336:0:(obd_mount.c:1426:lustre_fill_super()) Unable to mount  (-22)
      

      Attached logs.tbz2 has debug_kernel dumps.

      Attachments

        Issue Links

          Activity

            [LU-9408] Client fails to mount with ZFS master (0.7.0) and Lustre master (2.9.56)

            The osd_oid() has been totally removed via the patch:
            https://review.whamcloud.com/#/c/27093/

            yong.fan nasf (Inactive) added a comment - The osd_oid() has been totally removed via the patch: https://review.whamcloud.com/#/c/27093/

            Andreas, Fan Yong is working on Project quota for ZFS, i think that messages will be removed too with zfs project quota supported.

            wangshilong Wang Shilong (Inactive) added a comment - Andreas, Fan Yong is working on Project quota for ZFS, i think that messages will be removed too with zfs project quota supported.

            It would be good to get a patch to quiet the spurious "osd_oid()) unsupported quota oid: 0x16" message at startup, since even I find that confusing and wonder whether there is something wrong. We know this is for project quota, which isn't supported in ZFS yet.

            adilger Andreas Dilger added a comment - It would be good to get a patch to quiet the spurious " osd_oid()) unsupported quota oid: 0x16 " message at startup, since even I find that confusing and wonder whether there is something wrong. We know this is for project quota, which isn't supported in ZFS yet.
            pjones Peter Jones added a comment -

            Thanks Nathaniel

            pjones Peter Jones added a comment - Thanks Nathaniel

            If I format mdt00 with --index=0, everything works just fine. This could be closed as "not a bug" i guess, though it's kind of a strange one to debug.

            utopiabound Nathaniel Clark added a comment - If I format mdt00 with --index=0, everything works just fine. This could be closed as "not a bug" i guess, though it's kind of a strange one to debug.
            ihara Shuichi Ihara (Inactive) added a comment - - edited

            Ah, are you sure does "--index=1" work without --index=0?

            Apr 26 14:20:52 ieel-mds04 kernel: Lustre: scratch-MDT0001: new disk, initializing
            Apr 26 14:20:52 ieel-mds04 kernel: Lustre: scratch-MDT0001: Imperative Recovery not enabled, recovery window 300-900

            it seem you've also setup --index=1 without --index=0 in your original description..

            However, you have --index=0 for MDT in your ldiskfs setup.

            8 UP osp scratch-OST0000-osc-MDT0000 scratch-MDT0000-mdtlov_UUID 5
            9 UP osp scratch-OST0001-osc-MDT0000 scratch-MDT0000-mdtlov_UUID 5

            ihara Shuichi Ihara (Inactive) added a comment - - edited Ah, are you sure does "--index=1" work without --index=0? Apr 26 14:20:52 ieel-mds04 kernel: Lustre: scratch-MDT0001: new disk, initializing Apr 26 14:20:52 ieel-mds04 kernel: Lustre: scratch-MDT0001: Imperative Recovery not enabled, recovery window 300-900 it seem you've also setup --index=1 without --index=0 in your original description.. However, you have --index=0 for MDT in your ldiskfs setup. 8 UP osp scratch-OST0000-osc-MDT0000 scratch-MDT0000-mdtlov_UUID 5 9 UP osp scratch-OST0001-osc-MDT0000 scratch-MDT0000-mdtlov_UUID 5

            People

              wangshilong Wang Shilong (Inactive)
              utopiabound Nathaniel Clark
              Votes:
              0 Vote for this issue
              Watchers:
              7 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: