[LU-13079] ost-pools test_23a: Some errors happened when getting quota info. Some devices may be not working or deactivated. The data in "[]" is inaccurate. Created: 15/Dec/19 Updated: 08/Jan/20 Resolved: 08/Jan/20 |
|
| Status: | Resolved |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 2.14.0 |
| Fix Version/s: | None |
| Type: | Bug | Priority: | Major |
| Reporter: | Maloo | Assignee: | WC Triage |
| Resolution: | Fixed | Votes: | 0 |
| Labels: | None | ||
| Issue Links: |
|
||||||||||||
| Severity: | 3 | ||||||||||||
| Rank (Obsolete): | 9223372036854775807 | ||||||||||||
| Description |
|
This issue was created by maloo for Arshad <arshad.super@gmail.com> This issue relates to the following test suite run: https://testing.whamcloud.com/test_sets/f3b1229e-1ea8-11ea-b1e8-52540065bddc test_23a failed with the following error: test_23a failed with 22
Total allocated inode limit: 0, total allocated block limit: 0
uid 500 is using default file quota setting
Some errors happened when getting quota info. Some devices may be not working or deactivated. The data in "[]" is inaccurate.
ost-pools test_23a: @@@@@@ FAIL: test_23a failed with 22
Trace dump:
= /usr/lib64/lustre/tests/test-framework.sh:6108:error()
= /usr/lib64/lustre/tests/test-framework.sh:6410:run_one()
= /usr/lib64/lustre/tests/test-framework.sh:6449:run_one_logged()
= /usr/lib64/lustre/tests/test-framework.sh:6295:run_test()
= /usr/lib64/lustre/tests/ost-pools.sh:1285:main()
Dumping lctl log to /autotest/at-candidate/2019-12-14/lustre-reviews-el7_6-x86_64--review-dne-part-2--1_5__70456___a2827a62-8be9-4c1a-ac98-c35b83c5f16e/ost-pools.test_23a.*.1576332059.log
CMD: trevis-68vm1.trevis.whamcloud.com,trevis-68vm2,trevis-68vm3,trevis-68vm4,trevis-68vm5 /usr/sbin/lctl dk > /autotest/at-candidate/2019-12-14/lustre-reviews-el7_6-x86_64--review-dne-part-2--1_5__70456___a2827a62-8be9-4c1a-ac98-c35b83c5f16e/ost-pools.test_23a.debug_log.\$(hostname -s).1576332059.log;
dmesg > /autotest/at-candidate/2019-12-14/lustre-reviews-el7_6-x86_64--review-dne-part-2--1_5__70456___a2827a62-8be9-4c1a-ac98-c35b83c5f16e/ost-pools.test_23a.dmesg.\$(hostname -s).1576332059.log
Resetting fail_loc on all nodes...CMD: trevis-68vm1.trevis.whamcloud.com,trevis-68vm2,trevis-68vm3,trevis-68vm4,trevis-68vm5 lctl set_param -n fail_loc=0 fail_val=0 2>/dev/null
done.
Destroy the created pools: testpool
CMD: trevis-68vm4 /usr/sbin/lctl pool_list lustre
lustre.testpool
CMD: trevis-68vm4 /usr/sbin/lctl pool_list lustre.testpool
CMD: trevis-68vm4 lctl pool_remove lustre.testpool lustre-OST0000_UUID
trevis-68vm4: OST lustre-OST0000_UUID removed from pool lustre.testpool
CMD: trevis-68vm4 lctl pool_remove lustre.testpool lustre-OST0003_UUID
trevis-68vm4: OST lustre-OST0003_UUID removed from pool lustre.testpool
CMD: trevis-68vm4 lctl pool_remove lustre.testpool lustre-OST0006_UUID
trevis-68vm4: OST lustre-OST0006_UUID removed from pool lustre.testpool
CMD: trevis-68vm4 lctl pool_list lustre.testpool | wc -l
CMD: trevis-68vm4 lctl pool_list lustre.testpool | wc -l
CMD: trevis-68vm4 lctl pool_destroy lustre.testpool
trevis-68vm4: Pool lustre.testpool destroyed
CMD: trevis-68vm4 lctl get_param -n lod.lustre-MDT0000-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm4 lctl get_param -n lod.lustre-MDT0000-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm5 lctl get_param -n lod.lustre-MDT0001-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm5 lctl get_param -n lod.lustre-MDT0001-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm4 lctl get_param -n lod.lustre-MDT0002-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm4 lctl get_param -n lod.lustre-MDT0002-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm5 lctl get_param -n lod.lustre-MDT0003-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm5 lctl get_param -n lod.lustre-MDT0003-mdtlov.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm1.trevis.whamcloud.com lctl get_param -n lov.lustre-*.pools.testpool 2>/dev/null || echo foo
CMD: trevis-68vm1.trevis.whamcloud.com lctl get_param -n lov.lustre-*.pools.testpool 2>/dev/null || echo foo
|
| Comments |
| Comment by Arshad Hussain [ 15/Dec/19 ] |
|
This looks similar to This is failing |
| Comment by James Nunez (Inactive) [ 16/Dec/19 ] |
|
We're seeing this failure for sanity-quota tests 7b and 27b starting on 14 DEC 2019. Logs for one of these failures is at https://testing.whamcloud.com/test_sets/8b797c1c-1f02-11ea-bb75-52540065bddc . |
| Comment by Andreas Dilger [ 16/Dec/19 ] |
|
I also noticed some spurious error messages being printed: quotactl mdt4 failed. quotactl mdt5 failed. : : quotactl mdt62 failed. quotactl mdt63 failed. This should not be printed to the console at all. I'd assume that this returns a useful error code like -ENODEV for the case where the device is non-existent. |
| Comment by Andreas Dilger [ 16/Dec/19 ] |
|
James Nunez (jnunez@whamcloud.com) uploaded a new patch: https://review.whamcloud.com/37041 This patch is causing an increase in sanity-quota and ost-pools testing tracked under |
| Comment by Wang Shilong (Inactive) [ 17/Dec/19 ] |
|
I think following patch fixed the problem: |
| Comment by Wang Shilong (Inactive) [ 08/Jan/20 ] |
|
|