[LU-3447] Client RDMA too fragmented: 128/255 src 128/256 dst frags Created: 10/Jun/13  Updated: 15/Mar/14  Resolved: 15/Mar/14

Status: Resolved
Project: Lustre
Component/s: None
Affects Version/s: Lustre 2.1.5
Fix Version/s: None

Type: Bug Priority: Major
Reporter: Erich Focht Assignee: Bruno Faccini (Inactive)
Resolution: Fixed Votes: 0
Labels: client
Environment:

Lustre servers running 2.1.5, Lustre clients with 1.8.9.


Attachments: File client_log_messages    
Severity: 3
Rank (Obsolete): 8618

 Description   

During an IOR-like benchmark doing directIO from multiple clients (16, 64) clients get disconnected and evicted. The MPI process dies in misery and some of it's processes aren't even killable.

We've seen that there was a similar bug a while ago that was marked as solved, it was occuring on lnet routers (https://bugzilla.lustre.org/show_bug.cgi?id=13607). This one is on clients.

What can lead to the "RDMA too fragmented" issue? Any hint or suggestion? Client log messages are in the attached file.

Regards,
Erich



 Comments   
Comment by Erich Focht [ 11/Jun/13 ]

Increasing the MTT size on the client nodes seems to solve the problem. For instructions: http://community.mellanox.com/docs/DOC-1120
We've set log_num_mtt to 24.

Having a more meaningful error message would be nice.

This bug can be closed.

Comment by Bruno Faccini (Inactive) [ 12/Jun/13 ]

Hello Eric,
Thank's for the hint that solved the issue on your side.
But to be complete on this it would be nice to give a try to the "map_on_demand" dynamic feature (o2iblnd proc/module parameter, but this has to be set on all nodes) that may also be a way to fix such problem.

Comment by Peter Jones [ 12/Jun/13 ]

Bruno

Can you please advise?

Thanks

Peter

Comment by Erich Focht [ 13/Jun/13 ]

Hi Bruno,

is that option available on 1.8.9 as well as on 2.X? Thanks for pointing me to it!

It is difficult to do that in the customer's environment if we need to set this on both clients and servers, he has 3-4 Lustre filesystems (not all from us), a mix of versions, and 3.5k clients. But I'll try to find an opportunity to do it and discuss with the customer.

Best regards,
Erich

Comment by Bruno Faccini (Inactive) [ 14/Jun/13 ]

Hello Eric,
Working more on this very un-frequent problem, it seems highly possible that it is caused by upper-layer/application doing big and un-aligned I/Os. Since you indicated that your customer got it when running some MPI application doing Direct-IOs, can you also check on his side about the fact that these I/Os could be unaligned (page boundaries) and about their size ??

Comment by Bruno Faccini (Inactive) [ 12/Jul/13 ]

Hello Eric,
Any news on your side ??

Comment by Erich Focht [ 25/Jul/13 ]

Hi Bruno,

unfortunately we cannot use the module option there. It is a huge enironment with several Lustre setups and the customer is not willing to switch that option over everywhere. Which we'd need to do (as far as I understand) on clients as well as on servers. So we can't switch the clients selectively over. But we will test it as soon as we can on another (upcoming) installation.

Regards,
Erich

Comment by John Fuchs-Chesney (Inactive) [ 08/Mar/14 ]

Erich,
Do you want us to keep this ticket open?
Maybe you have had a chance to test the issue on a later installation?
Thanks,
~ jfc.

Comment by John Fuchs-Chesney (Inactive) [ 15/Mar/14 ]

Customer was able to resolve problem. No more required here.

Generated at Sat Feb 10 01:33:59 UTC 2024 using Jira 9.4.14#940014-sha1:734e6822bbf0d45eff9af51f82432957f73aa32c.