[LU-2140] failure on conf-sanity test_35a: mount /dev/mapper/lvm--MDS-P1 at /mnt/mds1 failed: No such file or directory Created: 10/Oct/12 Updated: 13/Oct/12 Resolved: 13/Oct/12 |
|
| Status: | Resolved |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 2.4.0 |
| Fix Version/s: | Lustre 2.4.0 |
| Type: | Bug | Priority: | Minor |
| Reporter: | Maloo | Assignee: | Niu Yawei (Inactive) |
| Resolution: | Fixed | Votes: | 0 |
| Labels: | None | ||
| Severity: | 3 |
| Rank (Obsolete): | 5148 |
| Description |
|
This issue was created by maloo for Andreas Dilger <adilger@whamcloud.com> This issue relates to the following test suite run: https://maloo.whamcloud.com/test_sets/8bd74142-0d28-11e2-9d1f-52540035b04c. The sub-test test_35a failed with the following errors in the MDS console log: Lustre: DEBUG MARKER: test -b /dev/lvm-MDS/P1 Lustre: DEBUG MARKER: mkdir -p /mnt/mds1; mount -t lustre -o user_xattr,acl /dev/lvm-MDS/P1 /mnt/mds1 LDISKFS-fs (dm-0): mounted filesystem with ordered data mode. quota=on. Opts: Lustre: Found index 0 for lustre-MDT0000, updating log Lustre: Modifying parameter lustre-MDT0000-mdtlov.lov.stripesize in log lustre-MDT0000 LustreError: 458:0:(mgc_request.c:248:do_config_log_add()) failed processing sptlrpc log: -2 Lustre: lustre-MDT0000: used disk, loading LustreError: 504:0:(sec_config.c:1024:sptlrpc_target_local_copy_conf()) missing llog context LustreError: 535:0:(obd_mount.c:1195:lustre_osp_add_conn()) lustre-MDT0000-osp-MDT0000: can't add conn: rc = -2 LustreError: 15c-8: MGC192.168.4.20@o2ib: The configuration from log 'lustre-client' failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information. LustreError: 458:0:(obd_mount.c:1867:server_start_targets()) lustre-MDT0000: failed to start OSP: -2 LustreError: 458:0:(obd_mount.c:2397:server_fill_super()) Unable to start targets: -2 Lustre: Failing over lustre-MDT0000 LustreError: 458:0:(ldlm_request.c:1169:ldlm_cli_cancel_req()) Got rc -108 from cancel RPC: canceling anyway LustreError: 458:0:(ldlm_request.c:1795:ldlm_cli_cancel_list()) ldlm_cli_cancel_list: -108 Lustre: 458:0:(client.c:1909:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1349809494/real 1349809494] req@ffff88030f3de400 x1415377800134763/t0(0) o251->MGC192.168.4.20@o2ib@0@lo:26/25 lens 224/224 e 0 to 1 dl 1349809500 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1 Lustre: server umount lustre-MDT0000 complete LustreError: 458:0:(obd_mount.c:2985:lustre_fill_super()) Unable to mount (-2) Info required for matching: conf-sanity 35a |
| Comments |
| Comment by Niu Yawei (Inactive) [ 11/Oct/12 ] |
00000100:00000200:0.0:1349235540.266457:0:2553:0:(events.c:536:ptlrpc_uuid_to_peer()) 127.0.0.2@tcp->12345-ffffffec@<0:0> 00000100:00000100:0.0:1349235540.266458:0:2553:0:(client.c:81:ptlrpc_uuid_to_connection()) cannot find peer 127.0.0.2@tcp! 00010000:00080000:0.0:1349235540.266460:0:2553:0:(ldlm_lib.c:69:import_set_conn()) can't find connection 127.0.0.2@tcp The log shows ptlrpc_uuid_to_peer() failed, it's usually caused by the peer is in unreachable network. Do we have tcp network on the MDT node? If not, I think it should be a test script defect, maybe we should check the nettype first, then add proper fake nid. |
| Comment by Niu Yawei (Inactive) [ 11/Oct/12 ] |
| Comment by James A Simmons [ 11/Oct/12 ] |
|
Patch fixed this test from failing |
| Comment by Andreas Dilger [ 13/Oct/12 ] |
|
Marking fixed per James' comments. |