SPEC SFS®2014_vdi ResultCopyright © 2016-2019 Standard Performance Evaluation Corporation |
WekaIO | SPEC SFS2014_vdi = 1600 Desktops |
---|---|
WekaIO Matrix 3.1.9 with Supermicro BigTwin Servers | Overall Response Time = 0.48 msec |
|
WekaIO Matrix 3.1.9 with Supermicro BigTwin Servers | |
---|---|
Tested by | WekaIO | Hardware Available | July 2017 | Software Available | December 2018 | Date Tested | January 2019 | License Number | 4553 | Licensee Locations | San Jose, California |
WekaIO Matrix is a flash native parallel and distributed, scale out file system designed to solve the challenges of the most demanding workloads, including AI and machine learning, genomic sequencing, real-time analytics, media rendering, EDA, software development and technical compute. Matrix software is a POSIX compliant parallel file system that delivers industry leading performance and scale at a fraction of traditional storage products price. The software can support billions of files and scales to hundreds of petabytes in a single namespace. Matrix can be deployed on commodity servers as a dedicated storage appliance or in a hyperconverged mode with zero additional storage footprint. The same software runs on-premises and in the public cloud. WekaIO Matrix is a software only solution that runs on any standard X86 hardware infrastructure delivering huge savings compared to proprietary all-flash based appliances. This test platform is deployed as a dedicated storage implementation on Supermicro BigTwin servers.
Item No | Qty | Type | Vendor | Model/Name | Description |
---|---|---|---|---|---|
1 | 1 | Parallel File System | WekaIO | Matrix Software V3.1.9 | WekaIO Matrix is a parallel and distributed POSIX file system that scales across compute nodes and distributes data and metadata evenly across the nodes for parallel access. |
2 | 6 | Storage Server Chassis | Supermicro | SYS-2029BT-HNR | Supermicro BigTwin chassis, each with 4 nodes per 2U chassis, populated with 6 NVMe drives per node. A total of 23 nodes were used in the testing. |
3 | 138 | 3.84TB U.2 NVMe SSD | Micron | MTFDHAL3T8TCT1AR | Micron 9200 Pro U.2 NVMe Enterprise Class Drives. |
4 | 46 | Processor | Intel | SR3B3 | Intel Xeon Gold 6126 12 Cores 2.6GHz Processor. |
5 | 23 | Network Interface Card | Mellanox | MCX456A-ECAT | 100Gbit ConnectX-4 Ethernet dual port PCI-E adapters, one per node. |
6 | 276 | DIMM | Supermicro | DIMM 16GB 2667MHz 2Rx8 ECC | System Memory DDR4 2667MHz ECC. |
7 | 23 | Boot Drive | Micron | MTFDDAV240TCB1AR | Micron Pro 5100 SATA M.2, 240GB. |
8 | 23 | Network Interface Card | Supermicro | AOC-MHIBE-M1CGM-O | SIOM Single Port InfiniBand EDR QSFP28 VPI running in Ethernet mode. |
9 | 23 | BIOS Module | Supermicro | SFT-OOB-LIC | Out of Band Firmware Management BIOS-Flash. |
10 | 10 | Switch | Mellanox | MSN2700-CS2FC | 32-port 100GbE Switch. |
11 | 5 | Clients | Supermicro | SYS-2029BT-HNR | Clients are built-to-order from Supermicro. The base build is a BigTwin SYS-2029BT-HNR 2U/4-node chassis with X11DPT-B motherboards. Each node in the SYS-2020BT-HNR represents one client. The built-to-order components in each client includes 2 Intel(R) Xeon(R) Gold 6126 12-core CPUs, 24 DDR4-2666 16GB ECC RDIMMs, 1 100GbE connection to the switch fabric via 1 Mellanox ConnectX-4 PCIe Ethernet adapter, 1 AOC-MHIBE-M1CGM-O SIOM Single Port InfiniBand EDR QSFP28 VPI that is not used/connected. Out of the 20 clients, 1 was used as prime and 19 were used to generate the workload. |
Item No | Component | Type | Name and Version | Description |
---|---|---|---|---|
1 | Storage Node | MatrixFS File System | 3.1.9 | WekaIO Matrix is a distributed and parallel POSIX file system that runs on any NVMe, SAS or SATA enabled commodity server or cloud compute instance and forms a single storage cluster. The file system presents a POSIX compliant, high performance, scalable global namespace to the applications. |
2 | Storage Node | Operating System | CentOS 7.4 | The operating system on each storage node was 64-bit CENTOS Version 7.4. |
3 | Client | Operating System | CentOS 7.4 | The operating system on the load generator client was 64-bit CENTOS Version 7.4. |
4 | Client | MatrixFS Client | 3.1.9 | MatrixFS Client software is mounted on the load generator clients and presents a POSIX compliant file system. |
Storage Node | Parameter Name | Value | Description |
---|---|---|
SR-IOV | Enabled | Enables CPU virtualization technology |
HyperThreading | Disabled | HyperThreading |
None
Storage Node | Parameter Name | Value | Description |
---|---|---|
Jumbo Frames | 4190 | Enables up to 4190 bytes of Ethernet frames | Client | Parameter Name | Value | Description |
nofile | 500000 | Client side Linux kernel /etc/security/limits.conf nofile setting |
MTU | 4190 | Client OS NIC setting, MTU |
The MTU is set to 4190 and is required and valid for all environments and workloads.
Not applicable.
Item No | Description | Data Protection | Stable Storage | Qty |
---|---|---|---|---|
1 | 3.84TB U.2 Micron 9200 Pro NVMe SSD in the Supermicro BigTwin chassis | 16+2 | Yes | 138 |
2 | 240GB M.2 Micron 5100 SATA SSD in the Supermicro BigTwin to store and boot OS | None | Yes | 23 |
Number of Filesystems | 1 | Total Capacity | 342.73 TiB | Filesystem Type | MatrixFS |
---|
A single WekaIO Matrix file system was created and distributed evenly across all 138 NVMe drives in the cluster (23 storage nodes x 6 drives/node). Data was protected to a 16+2 failure level. The file system overprovisions an additional 20% of capacity for performance quality of service at high water mark.
WekaIO MatrixFS was created and distributed evenly across all 23 storage nodes in the cluster. The deployment model is as a dedicated server protected with Matrix Distributed Data Coding schema of 16+2. All data and metadata is distributed evenly across the 23 storage nodes.
Item No | Transport Type | Number of Ports Used | Notes |
---|---|---|---|
1 | 100GbE NIC | 46 | The solution used a total of 46 100GbE ports from the storage nodes to the network switch. |
2 | 100GbE NIC | 19 | The solution used a total of 19 100GbE ports from the clients to the network switch. |
3 | 100GbE NIC | 1 | The solution used a total of 1 100GbE ports from the prime to the network switch. |
The solution under test had a total of 320 100GbE ports from the 10 Mellanox MSN 2700 switches. The switches were configured in a leaf-spine topology where 4 were spines and 6 were leaves. Each spine switch utilized 24 ports from the connections to the leaf switches. At the leaf switches, the storage nodes consumed a total of 46 100GbE ports, while the clients and prime utilized 20 100GbE ports. The leaf-to-spine connections consumed a total of 96 100GbE ports. Combined, the storage nodes and clients utilized a total of 162 100GbE ports.
Item No | Switch Name | Switch Type | Total Port Count | Used Port Count | Notes |
---|---|---|---|---|---|
1 | Qty. 10, Mellanox MSN 2700 | 100Gb Ethernet | 320 | 162 | Switches have Jumbo Frames enabled with MTU set to 4190 |
Item No | Qty | Type | Location | Description | Processing Function |
---|---|---|---|---|---|
1 | 46 | CPU | SYS-2029BT-HNR | Intel(R) Xeon(R) Gold 6126, 12 Cores, 2.6GHz CPU | WekaIO MatrixFS, Data Protection, device driver |
2 | 38 | CPU | SYS-2029BT-HNR | Intel(R) Xeon(R) Gold 6126, 12 Cores, 2.6GHz CPU | WekaIO MatrixFS client |
3 | 2 | CPU | SYS-2029BT-HNR | Intel(R) Xeon(R) Gold 6126, 12 Cores, 2.6GHz CPU | SPEC SFS2014 Prime |
Each storage node has 2 processors, each processor has 12 cores at 2.6Ghz. Each client has 2 processors, each processor has 12 cores. WekaIO Matrix utilized 3 of the 24 available cores on the client to run Matrix functions.
Description | Size in GiB | Number of Instances | Nonvolatile | Total GiB |
---|---|---|---|---|
Storage node memory | 192 | 23 | V | 4416 |
Client memory | 384 | 19 | V | 7296 |
Prime memory | 384 | 1 | V | 384 | Grand Total Memory Gibibytes | 12096 |
Each storage node has 192GBytes of memory for a total of 4,416GBytes. Each client has 384GBytes of memory for a total of 7,296GBytes. The prime has 384GBytes of memory for a total of 384GBytes.
WekaIO does not use any internal memory to temporarily cache write data to the underlying storage system. All writes are committed directly to the storage disk, therefore there is no need for any RAM battery protection. Data is protected on the storage media using WekaIO Matrix Distributed Data Protection (16+2). In the event of a power failure a write in transit would not be acknowledged.
The solution under test was a standard WekaIO Matrix enabled cluster in dedicated server mode. The solution will handle both large file I/O as well as small file random I/O and metadata intensive applications. No specialized tuning is required for different or mixed use workloads. None of the components used to perform the test were patched with Spectre or Meltdown patches (CVE-2017-5754,CVE-2017-5753,CVE-2017-5715).
None.
5 x SYS-2029BT-HNR storage Chassis (19 clients) were used to generate the benchmark workload. Each client had 1 x 100GbE network connection to a Mellanox MSN 2700 switch. 6 x Supermicro BigTwin 2029BT-HNR storage chassis (23 nodes) were benchmarked. Each storage node had 2 x 100GbE network connection to a Mellanox MSN 2700 switch. The clients (Supermicro) had the MatrixFS native NVMe POSIX Client mounted and had direct and parallel access to all 23 storage nodes.
None
None
Generated on Wed Mar 13 16:19:31 2019 by SpecReport
Copyright © 2016-2019 Standard Performance Evaluation Corporation