Details
-
Bug
-
Resolution: Fixed
-
Critical
-
Lustre 2.8.0
-
None
-
3
-
9223372036854775807
Description
[root@testnode tests]# MDSCOUNT=4 sh llmount.sh
Stopping clients: testnode /mnt/lustre (opts:)
Stopping clients: testnode /mnt/lustre2 (opts:)
Loading modules from /work/lustre-release_new/lustre/tests/..
detected 8 online CPUs by sysfs
libcfs will create CPU partition based on online CPUs
debug=vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck
subsystem_debug=all -lnet -lnd -pinger
quota/lquota options: 'hash_lqs_cur_bits=3'
Formatting mgs, mds, osts
Format mds1: /tmp/lustre-mdt1
Format mds2: /tmp/lustre-mdt2
Format mds3: /tmp/lustre-mdt3
Format mds4: /tmp/lustre-mdt4
Format ost1: /tmp/lustre-ost1
Format ost2: /tmp/lustre-ost2
Checking servers environments
Checking clients testnode environments
Loading modules from /work/lustre-release_new/lustre/tests/..
detected 8 online CPUs by sysfs
libcfs will create CPU partition based on online CPUs
debug=vfstrace rpctrace dlmtrace neterror ha config ioctl super lfsck
subsystem_debug=all -lnet -lnd -pinger
Setup mgs, mdt, osts
Starting mds1: -o loop /tmp/lustre-mdt1 /mnt/mds1
Started lustre-MDT0000
Starting mds2: -o loop /tmp/lustre-mdt2 /mnt/mds2
Started lustre-MDT0001
Starting mds3: -o loop /tmp/lustre-mdt3 /mnt/mds3
Started lustre-MDT0002
Starting mds4: -o loop /tmp/lustre-mdt4 /mnt/mds4
Started lustre-MDT0003
Starting ost1: -o loop /tmp/lustre-ost1 /mnt/ost1
Started lustre-OST0000
Starting ost2: -o loop /tmp/lustre-ost2 /mnt/ost2
Started lustre-OST0001
Starting client: testnode: -o user_xattr,flock testnode@tcp:/lustre /mnt/lustre
Using TIMEOUT=20
seting jobstats to procname_uid
Setting lustre.sys.jobid_var from disable to procname_uid
Waiting 90 secs for update
Updated after 3s: wanted 'procname_uid' got 'procname_uid'
disable quota as required
[root@testnode tests]# MDSCOUNT=4 ONLY="90 91" sh sanityn.sh
Logging to shared log directory: /tmp/test_logs/1443440481
Starting client testnode: -o user_xattr,flock testnode@tcp:/lustre /mnt/lustre2
Started clients testnode:
testnode@tcp:/lustre on /mnt/lustre2 type lustre (rw,user_xattr,flock)
testnode: Checking config lustre mounted on /mnt/lustre
Checking servers environments
Checking clients testnode environments
Using TIMEOUT=20
disable quota as required
osd-ldiskfs.track_declares_assert=1
1+0 records in
1+0 records out
1048576 bytes (1.0 MB) copied, 0.073335 s, 14.3 MB/s
running as uid/gid/euid/egid 500/500/500/500, groups:
[touch] [/mnt/lustre/d0_runas_test/f6126]
excepting tests: 14b 18c 19 28 29 35
skipping tests SLOW=no: 33a
192.168.1.61@tcp:/lustre /mnt/lustre2 lustre rw,flock,user_xattr 0 0
== sanityn test 90: open/create and unlink striped directory == 04:41:22 (1443440482)
start pid 7559 to open/create under striped directory
start pid 7560 to unlink striped directory
sanityn.sh: line 3395: 7559 Terminated ( cd $DIR1; while true; do
$LFS mkdir -c$MDSCOUNT $tdir > /dev/null 2>&1; touch $tdir/f{0..3} > /dev/null 2>&1;
done )
sanityn.sh: line 3395: 7560 Terminated ( cd $DIR2; while true; do
rm -rf $tdir > /dev/null 2>&1;
done )
Resetting fail_loc on all nodes...done.
PASS 90 (180s)
== sanityn test 91: chmod and unlink striped directory == 04:44:22 (1443440662)
start pid 83746 to chmod striped directory
start pid 83747 to unlink striped directory
sanityn.sh: line 3432: 83746 Terminated ( cd $DIR1; while true; do
$LFS mkdir -c$MDSCOUNT $tdir > /dev/null 2>&1; chmod go+w $tdir > /dev/null 2>&1;
done )
sanityn.sh: line 3432: 83747 Terminated ( cd $DIR2; while true; do
rm -rf $tdir > /dev/null 2>&1;
done )
Resetting fail_loc on all nodes...done.
PASS 91 (180s)
cleanup: ======================================================
== sanityn test complete, duration 362 sec == 04:47:23 (1443440843)
Stopping clients: testnode /mnt/lustre2 (opts:)
[root@testnode tests]# MDSCOUNT=4 sh llmountcleanup.sh
Stopping clients: testnode /mnt/lustre (opts:-f)
Stopping client testnode /mnt/lustre opts:-f
Stopping clients: testnode /mnt/lustre2 (opts:-f)
Stopping /mnt/mds1 (opts:-f) on testnode
Stopping /mnt/mds2 (opts:-f) on testnode
Stopping /mnt/mds3 (opts:-f) on testnode
Stopping /mnt/mds4 (opts:-f) on testnode
Stopping /mnt/ost1 (opts:-f) on testnode
Stopping /mnt/ost2 (opts:-f) on testnode
waited 0 for 31 ST ost OSS OSS_uuid 0
LustreError: 19915:0:(class_obd.c:638:cleanup_obdclass()) obd_memory max: 947765874, leaked: 4111712
Memory leaks detected
Attachments
Issue Links
- is related to
-
LU-7231 ENOSPC on remote MDT might create a in-consistent striped directory
-
- Resolved
-