It used to be that the storage media was the determining factor of an application’s performance. However, with non-volatile memory express (NVMe) solid-state drives (SSDs) now having reached price parity with Serial-Attached SCSI (SAS) SSDs, the potential to achieve hundreds of thousands of input/output operations per second (IOPS) has practically become commoditized. The problem is that these drives are now exposing new application bottlenecks – primary being the system architecture, the latency of the storage network and the efficiency of the storage software stack.
Why the Storage System Architecture Matters
Fully capitalizing on very fast storage media requires the components of the storage system architecture around the media to come together in an optimized manner. How the various drives are interconnected, how those drives connect with internal storage switches, and how this network of drives connects to the CPU, will all impact the application’s performance levels. For example, the CPU must have enough PCI lanes to avoid being overwhelmed by the speeds at which the NVMe SSD feeds it data.
Why the Storage Network Matters
NVMe-over Fabrics (NVMe-oF) is on a path to production deployments within the next year, and with it the promise of performance on par with direct-attached storage implementations, but with the superior efficiency of networked storage. However, the transition to NVMe-oF will not be rip-and-replace. NVMe-oF implementations will need to be able to plug into existing infrastructure, in order to support the mixed node environments that will exist.
Why the Storage Software Matters
Legacy storage software algorithms are plagued with inefficiencies, because storage media used to be so slow that, within reason, application performance was not impacted by the efficiency, or lack thereof, of the storage software. That is no longer the case with NVMe SSDs and continued development of CPUs. For example, storage software needs to be written to take advantage of the higher-core count processors that have hit the market. Data protection and reduction services like compression, deduplication and snapshots are becoming table stakes with high-core count processors. Customers do not want to compromise between these capabilities and performance. These data services must be computationally efficient, and they must be used intelligently, where they will have a substantial enough impact to justify any performance overhead.
Violin Systems recently joined Storage Switzerland for a discussion on how to architect the flash storage system to optimize performance. Access our on demand webinar, “Flash Storage – Deciding Between High Performance and EXTREME Performance”, to learn more.
Sign up for our Newsletter. Get updates on our latest articles and webinars, plus EXCLUSIVE subscriber only content.