Uploaded image for project: 'Lustre'
  1. Lustre
  2. LU-8133

ost-pools are not destroyed if test-case using ost-pools fail.

    XMLWordPrintable

Details

    • Bug
    • Resolution: Duplicate
    • Major
    • None
    • Lustre 2.8.0
    • None
    • 3
    • 9223372036854775807

    Description

      It is found that if the test-case involving ost-pools fails, the pools created in the test-case is not destroyed.
      Following is the result when I purposely made sanity/test_220 to fail and found out that pool test_220 still exists in the system.

      == sanity test 220: preallocated MDS objects still used if ENOSPC from OST == 15:27:51 (1463047071)
      pdsh@Seagate: 169.254.90.7: ssh exited with exit code 3
      UUID                      Inodes       IUsed       IFree IUse% Mounted on
      lustre-MDT0000_UUID       100000         218       99782   0% /mnt/lustre[MDT:0]
      lustre-OST0000_UUID        50016       50016           0 100% /mnt/lustre[OST:0]
      lustre-OST0001_UUID        50016       50016           0 100% /mnt/lustre[OST:1]
      
      filesystem summary:          218         218           0 100% /mnt/lustre
      
      fail_val=-1
      fail_loc=0x229
      169.254.90.6: Pool lustre.test_220 created
      169.254.90.6: poolname is empty
      169.254.90.6: argument lustre. must be <fsname>.<poolname>
      169.254.90.6: pool_add: Invalid argument
      pdsh@Seagate: 169.254.90.6: ssh exited with exit code 22
       sanity test_220: @@@@@@ FAIL: test_220 failed with 2 
        Trace dump:
        = /root/Desktop/code/lustre-wc-rel/lustre/tests/test-framework.sh:4673:error()
        = /root/Desktop/code/lustre-wc-rel/lustre/tests/test-framework.sh:4933:run_one()
        = /root/Desktop/code/lustre-wc-rel/lustre/tests/test-framework.sh:4969:run_one_logged()
        = /root/Desktop/code/lustre-wc-rel/lustre/tests/test-framework.sh:4775:run_test()
        = sanity.sh:11787:main()
      Dumping lctl log to /tmp/test_logs/1463047063/sanity.test_220.*.1463047083.log
      169.254.90.7: ssh: Could not resolve hostname Seagate: Temporary failure in name resolution
      169.254.90.7: rsync: connection unexpectedly closed (0 bytes received so far) [sender]
      169.254.90.7: rsync error: error in rsync protocol data stream (code 12) at io.c(600) [sender=3.0.6]
      pdsh@Seagate: 169.254.90.7: ssh exited with exit code 12
      169.254.90.6: ssh: Could not resolve hostname Seagate: Temporary failure in name resolution
      169.254.90.6: rsync: connection unexpectedly closed (0 bytes received so far) [sender]
      169.254.90.6: rsync error: error in rsync protocol data stream (code 12) at io.c(600) [sender=3.0.6]
      pdsh@Seagate: 169.254.90.6: ssh exited with exit code 12
      169.254.90.8: ssh: Could not resolve hostname Seagate: Temporary failure in name resolution
      169.254.90.8: rsync: connection unexpectedly closed (0 bytes received so far) [sender]
      169.254.90.8: rsync error: error in rsync protocol data stream (code 12) at io.c(600) [sender=3.0.6]
      pdsh@Seagate: 169.254.90.8: ssh exited with exit code 12
      test_220 returned 1
      FAIL 220 (13s)
      == sanity test complete, duration 21 sec == 15:28:04 (1463047084)
      debug=super ioctl neterror warning dlmtrace error emerg ha rpctrace vfstrace config console lfsck
      [root@Seagate tests]# lctl pool_list lustre 
      Pools from lustre:
      lustre.test_220       <==pool still exists
      

      Similar is the case for other tests like replay-single/test_85b, etc.

      Attachments

        Issue Links

          Activity

            People

              wc-triage WC Triage
              529964 Bhagyesh Dudhediya (Inactive)
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: