[LU-6580] Poor read performance with many ptlrpcd threads on the client Created: 07/May/15 Updated: 13/Oct/21 Resolved: 13/Oct/21 |
|
| Status: | Resolved |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | Lustre 2.7.0 |
| Fix Version/s: | None |
| Type: | Bug | Priority: | Major |
| Reporter: | Sebastien Buisson (Inactive) | Assignee: | Dmitry Eremin (Inactive) |
| Resolution: | Duplicate | Votes: | 0 |
| Labels: | None | ||
| Attachments: |
|
||||||||
| Issue Links: |
|
||||||||
| Severity: | 3 | ||||||||
| Rank (Obsolete): | 9223372036854775807 | ||||||||
| Description |
|
Hi, With Lustre 2.7.0, we have noticed a very bad impact of the number of ptlrpcd threads on the client over the data read performance. For instance see the results attached to this ticket. We use IOR with 24 tasks on a Lustre client node with 24 cores and one Infiniband FDR interface. What can explain this phenomenon in Lustre 2.7.0, as we do not see it with Lustre 2.5 and 2.6? Thanks in advance, |
| Comments |
| Comment by Andreas Dilger [ 07/May/15 ] |
|
Can you please take a look at the patch in That doesn't explain why there is a regression in 2.7.0, which still needs to be looked into, but may help fix the problem. |
| Comment by Sebastien Buisson (Inactive) [ 18/May/15 ] |
|
Hi, I gave a try to patch from I can see two interesting phenomena:
These results are obtained with 'ptlrpcd_partner_group_size=1', because using the default value (2) gives slightly lower performance. So the ability to restrict ptlrpcd threads to a specific cpt (thanks to patch http://review.whamcloud.com/13972) is very helpful. But this brings a new question: why read performance is so poor when putting ptlrpcd threads on all cpts? Thanks, |