[LU-2140] failure on conf-sanity test_35a: mount /dev/mapper/lvm--MDS-P1 at /mnt/mds1 failed: No such file or directory Created: 10/Oct/12  Updated: 13/Oct/12  Resolved: 13/Oct/12

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: Lustre 2.4.0
Fix Version/s: Lustre 2.4.0

Type: Bug Priority: Minor
Reporter: Maloo Assignee: Niu Yawei (Inactive)
Resolution: Fixed Votes: 0
Labels: None

Severity: 3
Rank (Obsolete): 5148

 Description   

This issue was created by maloo for Andreas Dilger <adilger@whamcloud.com>

This issue relates to the following test suite run: https://maloo.whamcloud.com/test_sets/8bd74142-0d28-11e2-9d1f-52540035b04c.

The sub-test test_35a failed with the following errors in the MDS console log:

Lustre: DEBUG MARKER: test -b /dev/lvm-MDS/P1
Lustre: DEBUG MARKER: mkdir -p /mnt/mds1; mount -t lustre -o user_xattr,acl  		                   /dev/lvm-MDS/P1 /mnt/mds1
LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. quota=on. Opts: 
Lustre: Found index 0 for lustre-MDT0000, updating log
Lustre: Modifying parameter lustre-MDT0000-mdtlov.lov.stripesize in log lustre-MDT0000
LustreError: 458:0:(mgc_request.c:248:do_config_log_add()) failed processing sptlrpc log: -2
Lustre: lustre-MDT0000: used disk, loading
LustreError: 504:0:(sec_config.c:1024:sptlrpc_target_local_copy_conf()) missing llog context
LustreError: 535:0:(obd_mount.c:1195:lustre_osp_add_conn()) lustre-MDT0000-osp-MDT0000: can't add conn: rc = -2
LustreError: 15c-8: MGC192.168.4.20@o2ib: The configuration from log 'lustre-client' failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information.
LustreError: 458:0:(obd_mount.c:1867:server_start_targets()) lustre-MDT0000: failed to start OSP: -2
LustreError: 458:0:(obd_mount.c:2397:server_fill_super()) Unable to start targets: -2
Lustre: Failing over lustre-MDT0000
LustreError: 458:0:(ldlm_request.c:1169:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway
LustreError: 458:0:(ldlm_request.c:1795:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108
Lustre: 458:0:(client.c:1909:ptlrpc_expire_one_request()) @@@ Request  sent has timed out for slow reply: [sent 1349809494/real 1349809494]  req@ffff88030f3de400 x1415377800134763/t0(0) o251->MGC192.168.4.20@o2ib@0@lo:26/25 lens 224/224 e 0 to 1 dl 1349809500 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1
Lustre: server umount lustre-MDT0000 complete
LustreError: 458:0:(obd_mount.c:2985:lustre_fill_super()) Unable to mount (-2)

Info required for matching: conf-sanity 35a



 Comments   
Comment by Niu Yawei (Inactive) [ 11/Oct/12 ]
00000100:00000200:0.0:1349235540.266457:0:2553:0:(events.c:536:ptlrpc_uuid_to_peer()) 127.0.0.2@tcp->12345-ffffffec@<0:0>
00000100:00000100:0.0:1349235540.266458:0:2553:0:(client.c:81:ptlrpc_uuid_to_connection()) cannot find peer 127.0.0.2@tcp!
00010000:00080000:0.0:1349235540.266460:0:2553:0:(ldlm_lib.c:69:import_set_conn()) can't find connection 127.0.0.2@tcp

The log shows ptlrpc_uuid_to_peer() failed, it's usually caused by the peer is in unreachable network. Do we have tcp network on the MDT node? If not, I think it should be a test script defect, maybe we should check the nettype first, then add proper fake nid.

Comment by Niu Yawei (Inactive) [ 11/Oct/12 ]

http://review.whamcloud.com/4247

Comment by James A Simmons [ 11/Oct/12 ]

Patch fixed this test from failing

Comment by Andreas Dilger [ 13/Oct/12 ]

Marking fixed per James' comments.

Generated at Sat Feb 10 01:22:42 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.