Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-3554

2.4.0<->2.3 interop: sanity test 233: FAIL: cannot access /mnt/lustre using its FID '[0xfc10c:0xed828a32:0x0]'

Details

    • 3
    • 8947

    Description

      sanity test 233 failed as follows:

      == sanity test 233: checking that OBF of the FS root succeeds == 21:44:37 (1372826677)
      stat: cannot stat `/mnt/lustre/.lustre/fid/[0xfc10c:0xed828a32:0x0]': Invalid argument
       sanity test_233: @@@@@@ FAIL: cannot access /mnt/lustre using its FID '[0xfc10c:0xed828a32:0x0]' 
      

      Dmesg on client node showed that:

      Lustre: DEBUG MARKER: == sanity test 233: checking that OBF of the FS root succeeds == 21:44:37 (1372826677)
      LustreError: 73818:0:(dir.c:433:ll_get_dir_page()) read cache page: [0x200000002:0x2:0x0] at 0: rc -1
      LustreError: 73818:0:(dir.c:595:ll_dir_read()) error reading dir [0x200000002:0x2:0x0] at 0: rc -1
      Lustre: DEBUG MARKER: /usr/sbin/lctl mark  sanity test_233: @@@@@@ FAIL: cannot access \/mnt\/lustre using its FID \'[0xfc10c:0xed828a32:0x0]\' 
      

      Dmesg on MDS node showed that:

      Lustre: DEBUG MARKER: == sanity test 233: checking that OBF of the FS root succeeds == 21:44:37 (1372826677)
      Lustre: 126270:0:(mdd_device.c:880:obf_lookup()) mdd_obd-lustre-MDT0000: [0xfc10c:0xed828a32:0x0] is invalid, sequence should be >= 0x200000400
      Lustre: DEBUG MARKER: /usr/sbin/lctl mark  sanity test_233: @@@@@@ FAIL: cannot access \/mnt\/lustre using its FID \'[0xfc10c:0xed828a32:0x0]\' 
      

      Maloo report:
      https://maloo.whamcloud.com/test_sets/9ed1db56-e3a4-11e2-9137-52540035b04c

      The sanity test 233 was added by patch http://review.whamcloud.com/6209 which was not on Lustre b2_3 branch.

      Attachments

        Activity

          [LU-3554] 2.4.0<->2.3 interop: sanity test 233: FAIL: cannot access /mnt/lustre using its FID '[0xfc10c:0xed828a32:0x0]'
          cfaber#1 Colin Faber [X] (Inactive) made changes -
          Comment [ I don't think this bug is fixed yet. Running 2.4.1 stock release on a single node configuration I see this pretty regularly when performing lots of fid2path lookups

          {code}
          Lustre: 25268:0:(mdt_handler.c:5738:mdt_fid2path()) lustre-MDT0000: [0x200000002:0x2:0x0] is invalid, sequence should be >= 0x200000400
          {code}

          Here's the exact RPM install list on my SL6.4 machine

          {code}
          kernel-devel-2.6.32-358.6.2.el6_lustre.g230b174.x86_64
          kernel-firmware-2.6.32-358.6.2.el6_lustre.g230b174.x86_64
          lustre-modules-2.4.0-2.6.32_358.6.2.el6_lustre.g230b174.x86_64_gd3f91c4.x86_64
          lustre-2.4.0-2.6.32_358.6.2.el6_lustre.g230b174.x86_64_gd3f91c4.x86_64
          kernel-debuginfo-2.6.32-358.6.2.el6_lustre.g230b174.x86_64
          python-perf-debuginfo-2.6.32-358.6.2.el6_lustre.g230b174.x86_64
          perf-debuginfo-2.6.32-358.6.2.el6_lustre.g230b174.x86_64
          lustre-source-2.4.0-2.6.32_358.6.2.el6_lustre.g230b174.x86_64_gd3f91c4.x86_64
          lustre-iokit-1.4.0-1.noarch
          kernel-headers-2.6.32-358.6.2.el6_lustre.g230b174.x86_64
          lustre-ldiskfs-4.1.0-2.6.32_358.6.2.el6_lustre.g230b174.x86_64_gd3f91c4.x86_64
          lustre-osd-ldiskfs-2.4.0-2.6.32_358.6.2.el6_lustre.g230b174.x86_64_gd3f91c4.x86_64
          lustre-tests-2.4.0-2.6.32_358.6.2.el6_lustre.g230b174.x86_64_gd3f91c4.x86_64
          lustre-debuginfo-2.4.0-2.6.32_358.6.2.el6_lustre.g230b174.x86_64_gd3f91c4.x86_64
          {code} ]
          pjones Peter Jones made changes -
          Fix Version/s New: Lustre 2.4.1 [ 10294 ]
          Fix Version/s New: Lustre 2.5.0 [ 10295 ]
          Resolution New: Fixed [ 1 ]
          Status Original: Open [ 1 ] New: Resolved [ 5 ]
          pjones Peter Jones added a comment -

          Landed for 2.4.1 and 2.5.0

          pjones Peter Jones added a comment - Landed for 2.4.1 and 2.5.0
          pjones Peter Jones made changes -
          Labels Original: yuc

          A master patch for 233 can be seen here: http://review.whamcloud.com/7324

          Test 185 is fixed in Master by http://review.whamcloud.com/#/c/6887/

          keith Keith Mannthey (Inactive) added a comment - A master patch for 233 can be seen here: http://review.whamcloud.com/7324 Test 185 is fixed in Master by http://review.whamcloud.com/#/c/6887/
          yujian Jian Yu made changes -
          Labels New: yuc

          It seems both test_233 and test_185 failed and this is due to advanced features in the 2.4 server.

          I setup some patches for 2.4 (and they should end up in Master) to do a quick mds server version check before running sanity test 233.

          Patch for 233:
          http://review.whamcloud.com/6890

          Patch for 185:
          http://review.whamcloud.com/6891

          keith Keith Mannthey (Inactive) added a comment - It seems both test_233 and test_185 failed and this is due to advanced features in the 2.4 server. I setup some patches for 2.4 (and they should end up in Master) to do a quick mds server version check before running sanity test 233. Patch for 233: http://review.whamcloud.com/6890 Patch for 185: http://review.whamcloud.com/6891
          pjones Peter Jones made changes -
          Assignee Original: WC Triage [ wc-triage ] New: Keith Mannthey [ keith ]
          yujian Jian Yu created issue -

          People

            keith Keith Mannthey (Inactive)
            yujian Jian Yu
            Votes:
            0 Vote for this issue
            Watchers:
            5 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: