[LU-6607] MDS ( 2 node DNE) running out of memory and crash Created: 15/May/15 Updated: 24/Mar/18 Resolved: 24/Mar/18 |
|
| Status: | Resolved |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 2.7.0 |
| Fix Version/s: | None |
| Type: | Bug | Priority: | Blocker |
| Reporter: | Haisong Cai (Inactive) | Assignee: | Lai Siyao |
| Resolution: | Won't Fix | Votes: | 1 |
| Labels: | sdsc | ||
| Environment: |
Linux panda-mds-19-6.sdsc.edu 3.10.73-1.el6.elrepo.x86_64 #1 SMP Thu Mar 26 16:28:30 EDT 2015 x86_64 x86_64 x86_64 GNU/Linux lustre-2.7.51-3.10.73_1.el6.elrepo.x86_64_gb019b03.x86_64 |
||
| Attachments: |
|
| Severity: | 4 |
| Rank (Obsolete): | 9223372036854775807 |
| Description |
|
2 node DNE MDS A MDS node randomly running out of memory and hang. Clients are generating a ton of Lustre errors with strings "ptlrpc_expire_one_request". The numbers are from several hundred thousands to several millions of such errors from each node. Here are number of error counts from some nodes: comet-12-31 662616 I am attaching logs from both client and server on one such incidence. |
| Comments |
| Comment by Peter Jones [ 15/May/15 ] |
|
Lai Could you please advise on this issue? Thanks Peter |
| Comment by Haisong Cai (Inactive) [ 15/May/15 ] |
|
Just like to highlight these messages on server (should also be in messages-19-6.gz file) May 15 06:35:19 panda-mds-19-6 kernel: Lustre: ldlm_canceld: This server is not able to keep up with request traffic (cpu-bound). |
| Comment by Haisong Cai (Inactive) [ 19/May/15 ] |
|
Hi Lai, Any update? thanks, |
| Comment by Di Wang [ 19/May/15 ] |
|
Hello, Cai I checked the debug log and dmesg, and I can see MDT0001 seems very slow at that moment. though I can not figure out why from these message. So 1. Could you please post these information here stack trace of MDT0001 (panda-mds-19-6), which will help us understand what MDT0001 was busying with. Something like echo t > /proc/sysrq-trigger dmesg > /tmp/dmesg.out 2. Could you please post "cat /proc/slabinfo" here when OOM happens? Thanks |
| Comment by Haisong Cai (Inactive) [ 19/May/15 ] |
|
Hi WangDi, I understand when to run 2). Haisong |
| Comment by Di Wang [ 19/May/15 ] |
|
Hello, Cai Oh, I only need output of 1) when MDT1 is busy. But if you can get both at the same time, that would be great. Thanks |
| Comment by Haisong Cai (Inactive) [ 13/Jul/15 ] |
|
WangDi, We had 2 incidences recently and both time I failed to collect need info. Haisong |
| Comment by Haisong Cai (Inactive) [ 01/Sep/15 ] |
|
WangDi, We ran into this problem on one of MDS (mdt0, the master again today) echo t > /proc/sysrq-trigger dmesg.out & slabinfo.txt will be uploaded separately. Haisong |
| Comment by Haisong Cai (Inactive) [ 01/Sep/15 ] |
|
Files collected between 2 time MDS crashes. |
| Comment by Di Wang [ 01/Sep/15 ] |
|
Ah, it is a ZFS environment (ZFS + DNE)? A few questions here 1. I saw this on your MDS console message(dmesg_mds.gz), the kernel version is definitely not EL6? EL7? But we do not support EL7 server on MDS yet. could you please confirm what kernel did you use on MDS? Linux version 3.10.73-1.el6.elrepo.x86_64 (mockbuild@Build64R6) (gcc version 4.4.7 20120313 (Red Hat 4.4.7-11) (GCC) ) #1 SMP Thu Mar 26 16:28:30 EDT 2015 2. In the slab info kmalloc-8192 9033431 9033431 8192 1 2 : tunables 8 4 0 : slabdata 9033431 9033431 0 8192 size slab costs too much memory, 941G! that is too much. Btw: how much OSTs for each OSS? |
| Comment by Haisong Cai (Inactive) [ 01/Sep/15 ] |
|
Hi WangDi, We are running CentOS 6.6 with Linux kernel 3.10.73 from elrepo. Filesystem has 16 OSS and each has 6 OSTs. Haisong |
| Comment by Haisong Cai (Inactive) [ 01/Sep/15 ] |
|
On one of the 2 MDS servers: [root@panda-mds-19-6 panda-mds-19-6]# sysctl -a | grep slab |
| Comment by Di Wang [ 01/Sep/15 ] |
|
Is that possible you can upgrade MDS to 2.7.58 ? there are quite a few fix on these area since 2.7.51. Btw: we are currently testing ZFS on DNE at |
| Comment by Haisong Cai (Inactive) [ 01/Sep/15 ] |
|
We are about to apply a new patch related to Will it be satisfy your recommendation? Haisong |
| Comment by Di Wang [ 01/Sep/15 ] |
|
Hmm, I think |
| Comment by Haisong Cai (Inactive) [ 01/Sep/15 ] |
|
Hi Wang Di, I understand What I said earlier was, to work on Was that 2.7.58 equivalent? Haisong |
| Comment by Di Wang [ 02/Sep/15 ] |
|
Ah, it is. you can use that build. Thanks |
| Comment by Haisong Cai (Inactive) [ 02/Sep/15 ] |
|
Hi WangDi, You stated that 2.7.58 has a lot fixes. But it may still not fix our problem, correct? thanks, |
| Comment by Di Wang [ 03/Sep/15 ] |
|
Hello, Haisong Yes, I do not know the exact reason why for this 8192_size slab caused so much memory here. No, I do not think this is related with any default setting. Did you do a lot cross-MDT operation here, like creating remote directory or striped directory? (unfortunately, there are not enough stack trace information here) Btw: this stack trace is collected when OOM happens ? or before? or about to happen? Right now, I would suggest 1. Use 2.7.58 plus that patch (http://review.whamcloud.com/#/c/14926/) you need, maybe also include http://review.whamcloud.com/#/c/16161/. Even though 2.7.58 might not help you on this issue, but it is way better than 2.7.51 on DNE. |
| Comment by Peter Jones [ 24/Mar/18 ] |
|
SDSC have moved onto more current releases so I do not think any further work is needed here |