Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-852

Test failure on test suite conf-sanity, subtest test_53b

Details

    • Bug
    • Resolution: Duplicate
    • Major
    • None
    • None
    • None
    • 3
    • 1,048
    • 5210

    Description

      This issue was created by maloo for Chris Gearing <chris@whamcloud.com>

      This issue relates to the following test suite run: https://maloo.whamcloud.com/test_sets/cfce0b7e-0f45-11e1-9051-52540025f9af.

      The sub-test test_53b failed with the following error:

      $'Assertion 23 failed: (($tstarted && $tmin && $tmax)) (expanded: ((0 && 0 && 0)))
      Insane MDT thread counts (PDSH problems?)'

      Info required for matching: conf-sanity 53b

      Attachments

        Issue Links

          Activity

            [LU-852] Test failure on test suite conf-sanity, subtest test_53b
            di.wang Di Wang added a comment -

            It seems this is a duplicate of LU-1048, not by pdir ops patch. Sorry.

            di.wang Di Wang added a comment - It seems this is a duplicate of LU-1048 , not by pdir ops patch. Sorry.
            di.wang Di Wang added a comment -

            Andreas

            I just checked the log carefully, and this seems a different problem.

            The reason of this failure is that MDT can not be start, which seems related with Liang's pdir patch. I will reassign this to Liang.

            17:36:06:kmem_cache_create: duplicate cache dynlock_cache
            17:36:07:Pid: 10192, comm: modprobe Not tainted 2.6.32-220.el6_lustre.x86_64 #1
            17:36:08:Call Trace:
            17:36:08: [<ffffffff81161da8>] ? kmem_cache_create+0x538/0x5a0
            17:36:08: [<ffffffff81161cc6>] ? kmem_cache_create+0x456/0x5a0
            17:36:09: [<ffffffffa06e4eb0>] ? init_once+0x0/0x90 [ldiskfs]
            17:36:09: [<ffffffffa046c30d>] ? dynlock_cache_init+0x1f/0x42 [ldiskfs]
            17:36:09: [<ffffffffa046c2af>] ? init_ldiskfs_fs+0x1d7/0x1e2 [ldiskfs]
            17:36:10: [<ffffffffa046c0d8>] ? init_ldiskfs_fs+0x0/0x1e2 [ldiskfs]
            17:36:12: [<ffffffff8100204c>] ? do_one_initcall+0x3c/0x1d0
            17:36:12: [<ffffffff810af641>] ? sys_init_module+0xe1/0x250
            17:36:12: [<ffffffff8100b0f2>] ? system_call_fastpath+0x16/0x1b
            17:36:13:Not able to create dynlock cache
            17:36:14:LustreError: 10167:0:(obd_mount.c:1445:server_kernel_mount()) premount /dev/mapper/lvm--MDS-P0:0x0 ldiskfs failed: -19 Is the ldiskfs module available?
            17:36:14:LustreError: 10167:0:(obd_mount.c:1770:server_fill_super()) Unable to mount device /dev/mapper/lvm--MDS-P0: -19
            17:36:15:LustreError: 10167:0:(obd_mount.c:2316:lustre_fill_super()) Unable to mount (-19)
            17:36:15:Lustre: DEBUG MARKER: conf-sanity test_53b: @@@@@@ FAIL: MDT start failed

            di.wang Di Wang added a comment - Andreas I just checked the log carefully, and this seems a different problem. The reason of this failure is that MDT can not be start, which seems related with Liang's pdir patch. I will reassign this to Liang. 17:36:06:kmem_cache_create: duplicate cache dynlock_cache 17:36:07:Pid: 10192, comm: modprobe Not tainted 2.6.32-220.el6_lustre.x86_64 #1 17:36:08:Call Trace: 17:36:08: [<ffffffff81161da8>] ? kmem_cache_create+0x538/0x5a0 17:36:08: [<ffffffff81161cc6>] ? kmem_cache_create+0x456/0x5a0 17:36:09: [<ffffffffa06e4eb0>] ? init_once+0x0/0x90 [ldiskfs] 17:36:09: [<ffffffffa046c30d>] ? dynlock_cache_init+0x1f/0x42 [ldiskfs] 17:36:09: [<ffffffffa046c2af>] ? init_ldiskfs_fs+0x1d7/0x1e2 [ldiskfs] 17:36:10: [<ffffffffa046c0d8>] ? init_ldiskfs_fs+0x0/0x1e2 [ldiskfs] 17:36:12: [<ffffffff8100204c>] ? do_one_initcall+0x3c/0x1d0 17:36:12: [<ffffffff810af641>] ? sys_init_module+0xe1/0x250 17:36:12: [<ffffffff8100b0f2>] ? system_call_fastpath+0x16/0x1b 17:36:13:Not able to create dynlock cache 17:36:14:LustreError: 10167:0:(obd_mount.c:1445:server_kernel_mount()) premount /dev/mapper/lvm--MDS-P0:0x0 ldiskfs failed: -19 Is the ldiskfs module available? 17:36:14:LustreError: 10167:0:(obd_mount.c:1770:server_fill_super()) Unable to mount device /dev/mapper/lvm--MDS-P0: -19 17:36:15:LustreError: 10167:0:(obd_mount.c:2316:lustre_fill_super()) Unable to mount (-19) 17:36:15:Lustre: DEBUG MARKER: conf-sanity test_53b: @@@@@@ FAIL: MDT start failed

            Di, since the LU-718 patch will likely not be landed to master for 2.2, since it is more like a feature than a fix at this stage, is it still possible to fix this problem with a smaller patch?

            This failure is being hit via autotest:

            https://maloo.whamcloud.com/test_sets/8029de58-48b2-11e1-9c04-5254004bbbd3
            https://maloo.whamcloud.com/test_sets/8c925a50-48d3-11e1-9c04-5254004bbbd3

            adilger Andreas Dilger added a comment - Di, since the LU-718 patch will likely not be landed to master for 2.2, since it is more like a feature than a fix at this stage, is it still possible to fix this problem with a smaller patch? This failure is being hit via autotest: https://maloo.whamcloud.com/test_sets/8029de58-48b2-11e1-9c04-5254004bbbd3 https://maloo.whamcloud.com/test_sets/8c925a50-48d3-11e1-9c04-5254004bbbd3
            di.wang Di Wang added a comment -

            Will fix it in lu-718

            di.wang Di Wang added a comment - Will fix it in lu-718
            di.wang Di Wang added a comment -

            This is the error running with my patch. And I will fix it in my patch.

            di.wang Di Wang added a comment - This is the error running with my patch. And I will fix it in my patch.

            People

              liang Liang Zhen (Inactive)
              maloo Maloo
              Votes:
              0 Vote for this issue
              Watchers:
              1 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: