[LU-1150] 1.8 client using excessive slab when mounting 2.1 server Created: 29/Feb/12 Updated: 29/Feb/12 Resolved: 29/Feb/12 |
|
| Status: | Closed |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 1.8.x (1.8.0 - 1.8.5) |
| Fix Version/s: | None |
| Type: | Bug | Priority: | Major |
| Reporter: | Christopher Morrone | Assignee: | WC Triage |
| Resolution: | Duplicate | Votes: | 0 |
| Labels: | None | ||
| Environment: |
Lustre 1.8.5.0-6chaos |
||
| Issue Links: |
|
||||||||
| Severity: | 3 | ||||||||
| Rank (Obsolete): | 6443 | ||||||||
| Description |
|
We have found that our 1.8 clients are using excessive ammounts of slab now that they are mounting 2.1 server filesystems. The memory is not lost, per say, because a umount of the offending filesystem will result in the memory being freed and the node returning to normal operation. But we have seen slab usage grow to as much as 20GB on a node with only 24GB of ram. This is all in the generic slabs (mostly 16k, and 8k), not in a lustre named slab. Buffers and cache are necessarily nearly non-existant when the slab usage reaches these numbers. Our 1.8 is a bit old, so if you know of a fix that is in newer 1.8 versions let me know. I tried searching a bit, but didn't find anything that looked promising. Really, I am not too concerned about fixing 1.8, but I think we need to know enough about the problem to figure out if 2.1 will have the same issue or not. I enabled full lustre debugging, and used the debug daemon to collect logs while unmounting. Something like this: lctl debug_daemon enable; umount /p/lscratchd; lctl debug_daemon disable The log isn't as long as I would have liked, so I probably need to let it run longer before disabling. But |
| Comments |
| Comment by Peter Jones [ 29/Feb/12 ] |
|
It looks like this is a duplicate of |