Details
-
Bug
-
Resolution: Cannot Reproduce
-
Major
-
None
-
Lustre 2.13.0
-
None
-
3
-
9223372036854775807
Description
Added support for consistent memory leak detection to my scripts and sure enough, conf-sanity test 59 leaks memory in current master:
[ 5916.569153] Lustre: DEBUG MARKER: == conf-sanity test 59: writeconf mount option ======================================================= 15:41:23 (1551472883) [ 5918.077007] Lustre: Setting parameter lustre-MDT0000.mdt.identity_upcall in log lustre-MDT0000 [ 5918.118690] Lustre: ctl-lustre-MDT0000: No data found on store. Initialize space [ 5918.119237] Lustre: Skipped 2 previous similar messages [ 5918.132181] Lustre: lustre-MDT0000: new disk, initializing [ 5918.132596] Lustre: Skipped 1 previous similar message [ 5918.372703] Lustre: ctl-lustre-MDT0000: super-sequence allocation rc = 0 [0x0000000200000400-0x0000000240000400]:0:mdt [ 5919.051560] Lustre: DEBUG MARKER: oleg15-server.localnet: executing set_default_debug -1 all 16 [ 5922.520329] Lustre: lustre-MDT0000: Connection restored to 192.168.10.168@tcp (at 0@lo) [ 5922.521053] Lustre: Skipped 16 previous similar messages [ 5922.540923] Lustre: lustre-OST0000: Imperative Recovery not enabled, recovery window 60-180 [ 5922.541528] Lustre: Skipped 3 previous similar messages [ 5923.201728] Lustre: DEBUG MARKER: oleg15-server.localnet: executing set_default_debug -1 all 16 [ 5938.511155] Lustre: 8603:0:(client.c:2134:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1551472898/real 1551472898] req@ffff8800bab07b40 x1626836967632352/t0(0) o251->MGC192.168.10.168@tcp@0@lo:26/25 lens 224/224 e 0 to 1 dl 1551472904 ref 2 fl Rpc:XN/0/ffffffff rc 0/-1 [ 5938.513054] Lustre: 8603:0:(client.c:2134:ptlrpc_expire_one_request()) Skipped 2 previous similar messages [ 5938.544098] Lustre: server umount lustre-OST0000 complete [ 5938.544517] Lustre: Skipped 3 previous similar messages [ 5940.849292] Lustre: MGS: Logs for fs lustre were removed by user request. All servers must be restarted in order to regenerate the logs: rc = 0 [ 5940.858390] Lustre: Setting parameter lustre-MDT0000.mdt.identity_upcall in log lustre-MDT0000 [ 5941.576567] Lustre: DEBUG MARKER: oleg15-server.localnet: executing set_default_debug -1 all 16 [ 5943.500363] LustreError: 13b-9: lustre-OST0000 claims to have registered, but this MGS does not know about it, preventing registration. [ 5943.508686] LustreError: 13a-8: Failed to get MGS log lustre-OST0000 and no local copy. [ 5943.509328] LustreError: 15c-8: MGC192.168.10.168@tcp: The configuration from log 'lustre-OST0000' failed (-2). This may be the result of communication errors between this node and the MGS, a bad configuration, or other errors. See the syslog for more information. [ 5943.510952] LustreError: 9571:0:(obd_mount_server.c:1386:server_start_targets()) failed to start server lustre-OST0000: -2 [ 5943.529731] LustreError: 9571:0:(obd_mount_server.c:1939:server_fill_super()) Unable to start targets: -2 [ 5943.530468] LustreError: 9571:0:(obd_mount_server.c:1589:server_put_super()) no obd lustre-OST0000 [ 5943.576890] LustreError: 9571:0:(obd_mount.c:1608:lustre_fill_super()) Unable to mount (-2) [ 5944.443255] Lustre: MGS: Regenerating lustre-OST0000 log by user request: rc = 0 [ 5945.111938] Lustre: DEBUG MARKER: oleg15-server.localnet: executing set_default_debug -1 all 16 [ 5947.346104] Lustre: MGS: Regenerating lustre-OST0001 log by user request: rc = 0 [ 5948.013805] Lustre: DEBUG MARKER: oleg15-server.localnet: executing set_default_debug -1 all 16 [ 5952.788610] Lustre: lustre-OST0001-osc-MDT0000: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete [ 5952.790296] Lustre: lustre-OST0001: Not available for connect from 0@lo (stopping) [ 5957.780675] LustreError: 11-0: lustre-OST0000-osc-MDT0000: operation ost_statfs to node 0@lo failed: rc = -107 [ 5957.797623] LustreError: 137-5: lustre-OST0001_UUID: not available for connect from 0@lo (no target). If you are running an HA pair check that the target is mounted on the other server. [ 5962.804735] Lustre: lustre-OST0000: Not available for connect from 0@lo (stopping) [ 5962.805177] LustreError: 137-5: lustre-OST0001_UUID: not available for connect from 0@lo (no target). If you are running an HA pair check that the target is mounted on the other server. [ 5962.807882] Lustre: Skipped 1 previous similar message [ 5973.778356] Lustre: DEBUG MARKER: oleg15-server.localnet: executing /home/green/git/lustre-release/lustre/scripts/lustre_rmmod ldiskfs [ 5974.249337] Key type lgssc unregistered [ 5975.903115] LNet: Removed LNI 192.168.10.168@tcp [ 5975.975610] LustreError: 11726:0:(class_obd.c:824:obdclass_exit()) obd_memory max: 366353522, leaked: 80
Full report: http://testing.linuxhacker.ru:3333/lustre-reports/122/testresults/conf-sanity-zfs-centos7_x86_64-centos7_x86_64/
Attachments
Issue Links
- is related to
-
LU-10193 Index object on-disk layout compatibility
- Resolved