White Papers
Dell HPC Lustre Storage solution with Intel Omni-Path
sync
echo 3 > /proc/sys/vm/drop_caches
In addition, to simulate a cold cache on the server, a “sync” was performed on all the active servers
(OSS and MDS) before each test and the kernel was instructed to drop caches with the same commands
used on the client.
In measuring the performance of the Dell Storage for HPC with Intel EE for Lustre solution, all tests
were performed with similar initial conditions. The file system was configured to be fully functional
and the targets tested were emptied of files and directories prior to each test.
4.1 N-to-N Sequential Reads / Writes
The sequential testing was done with the IOzone testing tool version 3.444. The throughput results
presented in Figure 10 are converted to MB/s. The file size selected for this testing was such that the
aggregate sample size from all threads was consistently 2TB. That is, sequential reads and writes had
an aggregate sample size of 2TB divided equally among the number of threads within that test. The
block size for IOzone was set to 1MB to match the 1MB Lustre request size.
Each file written was large enough to minimize cache effects from OSS and clients. In addition, the
other techniques to prevent cache effects helped to avoid them as well. The files written were
distributed evenly across the OSTs (Round Robin). This was to prevent uneven I/O loads on any single
SAS connection or OST, in the same way that a user would expect to balance a workload.
Figure 10: Sequential Reads / Writes Dell HPC Lustre Storage with Intel Omni-Path
Figure 10 shows the sequential performance of the 960TB test configuration. With the test bed used,
write performance peaks slightly less than 14.9GB/sec while read performance peaks near 16.8GB/sec.
0
2
4
6
8
10
12
14
16
18
1 4 8 16 24 32 48 64 72 96 128 256
Throughput in GB/s
Number of concurrent threads
IOzone Sequential - Dell HPC Lustre Storage with Intel Omni-Path
Write Read