Details
-
Improvement
-
Resolution: Unresolved
-
Minor
-
None
-
Lustre 2.5.0
-
None
-
MDS Servers(2nos) - 12cores, 64GB Memory, QDR Single port
OSS Servers (2nos) - 16cores, 64GB Memory, FDR multi rail(2ports from each OSS)
Clients(8nos) - 12cores, 64GB memory, QDR single port
Lustre Version - 2.5
CentOS version - 6.4
No. of OSTs - 16(Configured in RAID-6(8+2)) and load balanced between OSS servers with 8 OSTs on each
MDS Servers(2nos) - 12cores, 64GB Memory, QDR Single port OSS Servers (2nos) - 16cores, 64GB Memory, FDR multi rail(2ports from each OSS) Clients(8nos) - 12cores, 64GB memory, QDR single port Lustre Version - 2.5 CentOS version - 6.4 No. of OSTs - 16(Configured in RAID-6(8+2)) and load balanced between OSS servers with 8 OSTs on each
-
12974
Description
Hi,
With the above environment, need suggestion on performance at the clients, since i'm struck. Appreciate your help.
I'm getting the block device write performance of 9.6GB/s with 16 LUNs which is measured through XDD. I ran obdfilter survey at the OSS machines and getting around 8.4GB/s as write performance. I've measured the LNET performance and getting 9.6GB/s between OSS machines and 8 clients. which But when I run IOR in the clients, i'm getting around 2.6GB/s for write performance with one client. While I run it across two nodes, getting 4.4GB/s of write throughput. But while I scale beyond 2nodes, getting the same performance of 4GB/s only. Could you please help to find the root cause for this performance availability issue.