[LU-8280] Bad Lustre Read Performance with Master Build 3371 Created: 15/Jun/16 Updated: 23/Jul/16 Resolved: 23/Jul/16 |
|
| Status: | Resolved |
| Project: | Lustre |
| Component/s: | None |
| Affects Version/s: | None |
| Fix Version/s: | None |
| Type: | Bug | Priority: | Critical |
| Reporter: | Cong Xu (Inactive) | Assignee: | Cliff White (Inactive) |
| Resolution: | Not a Bug | Votes: | 0 |
| Labels: | None | ||
| Attachments: |
|
| Severity: | 3 |
| Rank (Obsolete): | 9223372036854775807 |
| Description |
|
*Important notes: Please do not rebuild Lustre on wolf-[33-45], thanks! The read performance of Lustre master build 3371 is bad. Following are the detailed configurations for my evaluation over Lustre file system using IOR benchmark: [Lustre Configuration] MDS/MDT: wolf-37 Stripe Size: 4194304 Lustre Build Version: 3371
[IOR Benchmark] IOR Command:
IOR Results: Share File case Began: Tue Jun 14 20:16:18 2016 Test 0 started: Tue Jun 14 20:16:18 2016 access bw(MiB/s) block(KiB) xfer(KiB) open(s) wr/rd(s) close(s) total(s) iter Max Write: 2546.35 MiB/sec (2670.05 MB/sec) Summary of all tests: Finished: Tue Jun 14 20:28:01 2016 *************************************On the contrary, "lustre-b2_8 -b 12" performs well********************************** [Lustre Configuration] MDS/MDT: wolf-37 Stripe Size: 4194304 Lustre Build Version: 12
[IOR Benchmark] IOR Command:
IOR Results: Share File case Began: Wed Jun 15 01:17:36 2016 Test 0 started: Wed Jun 15 01:17:36 2016 access bw(MiB/s) block(KiB) xfer(KiB) open(s) wr/rd(s) close(s) total(s) iter Max Write: 2559.64 MiB/sec (2683.98 MB/sec) Summary of all tests: Finished: Wed Jun 15 01:24:15 2016 |
| Comments |
| Comment by Oleg Drokin [ 15/Jun/16 ] |
|
I moved this into Lustre project sine it seems that newer Lustre versions are slower and 2.8 is still good, so it's unlikely to do anything with the infrastructure. |
| Comment by Cliff White (Inactive) [ 27/Jun/16 ] |
|
Note: we normally run IOR with -i 5 at least - a single iteration does not produce a consistent result. Also, our normal performance tests use -b 4G, not -b 4m. |
| Comment by Cong Xu (Inactive) [ 27/Jun/16 ] |
|
In this evaluation, we configure Lustre stripe size, IOR Transfer size and Block size to be 4MB, and the number of Clients equals to number of OSTs. Thus, everything is perfect matched and we expect to deliver the maximum bandwidth of Lustre file system. |
| Comment by Cliff White (Inactive) [ 28/Jun/16 ] |
|
I ran a comparison of 2.8.0 and 2.8.53.38 on Spirit. |
| Comment by Cliff White (Inactive) [ 28/Jun/16 ] |
|
In my experience a single client cannot generate enough IO request to saturate a single OSS, >1 client per OST is normally required. For this reason we run performance across a range of client sizes, and use multiple threads per client. |
| Comment by Cliff White (Inactive) [ 28/Jun/16 ] |
|
This spreadsheet contains the relevant runs from Hyperion. 2.8.55 is below 2.8.0 performance @ 16 clients for both reads and writes, writes are also below standard at larger client counts. |
| Comment by Peter Jones [ 23/Jul/16 ] |
|
IIRC Cliff had reported that this was an issue with the way the tests were being run |