[LU-7442] conf-sanity test_41c: @@@@@@ FAIL: unexpected concurent MDT mounts rc=17 rc2=0 Created: 17/Nov/15 Updated: 28/Nov/16 Resolved: 06/Aug/16 |
|
| Status: | Resolved |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 2.8.0 |
| Fix Version/s: | Lustre 2.9.0 |
| Type: | Bug | Priority: | Major |
| Reporter: | hemaharish | Assignee: | Bruno Faccini (Inactive) |
| Resolution: | Fixed | Votes: | 0 |
| Labels: | None | ||
| Environment: |
single node setup |
||
| Issue Links: |
|
||||||||
| Severity: | 3 | ||||||||
| Rank (Obsolete): | 9223372036854775807 | ||||||||
| Description |
|
modules unloaded. Starting mds1: -o rw,user_xattr /dev/vdb /mnt/mds1 mount.lustre: set /sys/block/vdb/queue/max_sectors_kb to 2147483647 error: set_param: /proc/{fs,sys} / {lnet,lustre}/fail_loc: Found no match mount.lustre: mount /dev/vdb at /mnt/mds1 failed: File exists |
| Comments |
| Comment by Bruno Faccini (Inactive) [ 17/Nov/15 ] |
|
Looks like conf-sanity/test_41c needs some fixes/cleanup, as for |
| Comment by hemaharish [ 20/Nov/15 ] |
|
Hi,
== conf-sanity test 41c: concurrent mounts of MDT/OST should all fail but one == 10:42:24 (1447996344)
umount lustre on /mnt/lustre.....
stop ost1 service on centos6.6-Upstream-landing
stop mds service on centos6.6-Upstream-landing
modules unloaded.
Loading modules from /home/hema/xyratex/code/lustre-wc-rel/lustre/tests/..
detected 1 online CPUs by sysfs
libcfs will create CPU partition based on online CPUs
debug=-1
subsystem_debug=all -lnet -lnd -pinger
gss/krb5 is not supported
quota/lquota options: 'hash_lqs_cur_bits=3'
fail_loc=0x703
Starting mds1: -o loop /tmp/lustre-mdt1 /mnt/mds1
fail_loc=0x0
Starting mds1: -o loop /tmp/lustre-mdt1 /mnt/mds1
mount.lustre: mount /dev/loop1 at /mnt/mds1 failed: Operation already in progress
The target service is already running. (/dev/loop1)
Start of /tmp/lustre-mdt1 on mds1 failed 114
Started lustre-MDT0000
1st MDT start succeed
2nd MDT start failed with EALREADY
fail_loc=0x703
Starting ost1: -o loop /tmp/lustre-ost1 /mnt/ost1
fail_loc=0x0
Starting ost1: -o loop /tmp/lustre-ost1 /mnt/ost1
mount.lustre: mount /dev/loop2 at /mnt/ost1 failed: Operation already in progress
The target service is already running. (/dev/loop2)
Start of /tmp/lustre-ost1 on ost1 failed 114
Started lustre-OST0000
1st OST start succeed
2nd OST start failed with EALREADY
stop mds service on centos6.6-Upstream-landing
Stopping /mnt/mds1 (opts:-f) on centos6.6-Upstream-landing
Stopping /mnt/ost1 (opts:-f) on centos6.6-Upstream-landing
start mds service on centos6.6-Upstream-landing
Starting mds1: -o loop /tmp/lustre-mdt1 /mnt/mds1
Started lustre-MDT0000
start ost1 service on centos6.6-Upstream-landing
Starting ost1: -o loop /tmp/lustre-ost1 /mnt/ost1
Started lustre-OST0000
mount lustre on /mnt/lustre.....
Starting client: centos6.6-Upstream-landing: -o user_xattr,flock centos6.6-Upstream-landing@tcp:/lustre /mnt/lustre
setup single mount lustre success
umount lustre on /mnt/lustre.....
Stopping client centos6.6-Upstream-landing /mnt/lustre (opts:)
stop ost1 service on centos6.6-Upstream-landing
Stopping /mnt/ost1 (opts:-f) on centos6.6-Upstream-landing
stop mds service on centos6.6-Upstream-landing
Stopping /mnt/mds1 (opts:-f) on centos6.6-Upstream-landing
modules unloaded.
Resetting fail_loc on all nodes...done.
PASS 41c (78s)
|
| Comment by Gerrit Updater [ 20/Nov/15 ] |
|
HemaHarish (hema.yarramilli@seagate.com) uploaded a new patch: http://review.whamcloud.com/17301 |
| Comment by Bruno Faccini (Inactive) [ 20/Nov/15 ] |
|
The reason of the failure (in fact the non-permanent failure!) is still a bit mysterious for me, but patch re-loading of modules after cleanup is harmless and will clear any special cases... |
| Comment by hemaharish [ 23/Nov/15 ] |
|
Yes, failure was permanent on single node setup without patch. |
| Comment by Bruno Faccini (Inactive) [ 14/Jul/16 ] |
|
hemaharish, |
| Comment by nasf (Inactive) [ 15/Jul/16 ] |
|
Thanks Bruno. I have rebased my patch against the patch http://review.whamcloud.com/#/c/17427 to resolve conf-sanity test_41c failure. |
| Comment by Gerrit Updater [ 06/Aug/16 ] |
|
Oleg Drokin (oleg.drokin@intel.com) merged in patch http://review.whamcloud.com/17301/ |
| Comment by Peter Jones [ 06/Aug/16 ] |
|
Landed for 2.9 |