Nvme™ and Nvme-Of™ in Enterprise Arrays
Nvme™ and Nvme-Of™ in Enterprise Arrays
Mike Kieran
2
NVM Express Sponsored Track for Flash Memory Summit 2018
Track Title Speakers
NVMe-101-1
8/7/18 NVM Express: NVM Express roadmaps and market data for NVMe, NVMe-oF, Janene Ellefson, Micron Amber Huffman, Intel
8:30-9:35 and NVMe-MI - what you need to know the next year. J Metz, Cisco David Allen, Segate
8/7/18 NVMe architectures for in Hyperscale Data Centers, Enterprise Data Centers, Janene Ellefson, Micron Andy Yang, Toshiba
9:45-10:50 and in the Client and Laptop space. Chris Peterson, Jonmichael Hands, Intel
Facebook
3:40-4:45 NVMe Drivers and Software: This session will cover the software and drivers Uma Parepalli, Cavium Suds Jain, VMware
NVMe-102-1
8/7/18 required for NVMe-MI, NVMe, NVMe-oF and support from the top operating Austin Bolen, Dell EMC David Minturn, Intel
systems. Myron Loewen, Intel James Harris, Intel
Lee Prewitt, Microsoft
4:55-6:00 NVMe-oF Transports: We will cover for NVMe over Fibre Channel, NVMe over Brandon Hoff, Emulex Curt Beckmann, Brocade
8/7/18 RDMA, and NVMe over TCP. Fazil Osman, Broadcom Praveen Midha, Marvell
J Metz, Cisco
8/8/18 NVMe-oF Enterprise Arrays: NVMe-oF and NVMe is improving the Brandon Hoff, Emulex Mike Kieran, NetApp
NVMe-201-1
8:30-9:35 performance of classic storage arrays, a multi-billion dollar market. Clod Barrera, IBM Brent Yardley, IBM
8/8/18 NVMe-oF Appliances: We will discuss solutions that deliver high-performance Jeremy Warner, Toshiba Nishant Lodha, Marvell
9:45-10:50 and low-latency NVMe storage to automated orchestration-managed clouds. Manoj Wadekar, eBay Lior Gal, Excelero
Kamal Hyder, Toshiba
NVMe-202-1
8/8/18 NVMe-oF JBOFs: Replacing DAS storage with Composable Infrastructure Bryan Cowger, Praveen Midha, Marvell
3:20-4:25 (disaggregated storage), based on JBOFs as the storage target. Kazan Networks Fazil Osman, Broadcom
8/8/18 Testing and Interoperability: This session will cover testing for Conformance, Brandon Hoff, Emulex Jason Rusch, Viavi
4:40-6:45 Interoperability, Resilience/error injection testing to ensure interoperable Tim Sheehan, IOL Nick Kriczky, Teledyne
solutions base on NVM Express solutions. Mark Jones, FCIA 3
Abstract and Agenda
• Abstract:
• Enterprise Arrays: NVMe-oF™ and NVMe™ is improving the performance of classic
storage arrays, a multi-billion dollar market.
• NVMe-oF Panel
• Storage Segmentation – Brandon Hoff, Emulex
• NVMe over Fabrics Overview – Clod Barrera, IBM
• NVMe over Fabrics on Enterprise Arrays, ANA, and more – Mike Kieran, NetApp
• Performance Improvements at the Storage Array
• Performance improvements in NVMe over Fabrics at the initiator and end-to-end –
Brandon Hoff, Emulex
• Performance Improvements in the Sever and End-to-End
• Q&A
4
NVMe™ over Fabrics – Storage Architectures
Enterprise Arrays - Traditional SAN Benefits:
• Storage services (dedup, compression, thin
APPs provisioning)
• High availability at the array
APPs NVMe-oF Enterprise • Fully supported from the array vendor
Arrays • Example: NetApp/IBM
APPs
Benefits:
Server SAN/Storage Appliances • High performance storage
• Lower cost that storage arrays, minimal storage
APPs
services
APPs e.g. • Roll-your-own support model
NVMe-oF
Rows of • Ex. SUSE on Servers configured to be storage
APPs servers
targets
Benefits:
JBOF/Composable Storage • Very low latency
• Low cost
APPs
• Great for a single rack/single switch
APPs NVMe-oF Blocks of • Leverages NICs, smart NICs, and HBAs for
Storage NVMe-oF to PCIe®/NVMe™ translation
APPs
5
Enterprise Storage Market Block Storage Capacity Shipped
60,000
• Fibre Channel storage shows strong
Fibre Channel
Petabytes Shipped
growth in capacity 50,000
– Fibre Channel Storage capacity shipped
is larger than all other types of external 40,000
storage combined
30,000
• The adoption of All Flash Arrays and
iSCSI
NVMe™ storage will drive the need 20,000
for faster networks
10,000
• iSCSI is the dominate technology Others
block over Ethernet -
• The only RDMA market for block 2017 2018 2019 2020
storage is Infiniband Other Includes: FICON, FCoE,
Infiniband, External SAS
IDC WW Capacity Shipped, 2016
6
Three Areas of Performance Improvement
End to End Performance Improvements
7
Flash Memory Summit 2018
Santa Clara, CA
8
All-Flash Arrays
Hybrid Arrays
Server Server
Direct Attached Flash Server Server
Hyperconverged
Server Server
Flash Flash
Flash Flash
Flash Flash Flash Flash
Flash Flash
Flash Flash
Flash Flash
$20
(now on Gen 6 at 32Gbps
$B $15
• Expect gradual decline in FC SAN share of
$10
storage attachment
$5
• Storage fabrics for new workloads, CSPs, Cold $0
storage all favor IP storage attach – iSCSI, 2012 2013 2014 2015 2016 2017 2018
10
NVMe™ and NVMe-oF™
• NVMe protocol enables native parallelism within SSDs and All Flash Arrays (AFA)
• NVMe allows more efficient host software stacks for lower latency at application
• User-space drivers for selected software (e.g. In-memory DB) for maximum benefit
AFA / SSD
“IBM Storage and the NVM Express
Revolution” Koltsidas & Hsu 2017
– IBM Redpaper
11
NVMe-oF™ Performance Benefits
AFA SSD
NVM PCIe
e oF NVMe
AFA / SSD
• NVMe™ and NVMe-oF have new kernel driver • From IBM Research – Spark application with
stacks in hosts to reduce lock contention and RDMA connection to storage from user space
increase parallelism. Improved throughput and showed up to 5X improvement in performance.
lower latency.
• Requires complete re-structure of I/O system
• For I/O-bound workloads, NVMe-oF lowers server and application awareness/modification
I/O load and wait times.
13
The Benefits of Continuity
client investment
NVMe-oF™
• Management of full storage path (NVMe™ between
SAN Fabric
• Performance and availability management hosts and storage)
• Audit controls
• Upgrade migration process Storage Array
14
Value of NVMe™ and NVMe-oF™
Host … Host
• Optimized for Flash
NVMe-oF
• Fast and Getting Faster (NVMe between
SAN Fabric
hosts and storage)
• Reduce Application License costs
• Future proof investment Storage Array
15
Architected for Performance
17 © 2018 NetApp, Inc. All rights reserved. --- NETAPP CONFIDENTIAL --- 17
Impact of NVMe™ For Media Access
NVMe useful for SSDs but required for the next generation of solid state
Drive
~ 10 ms Latency
IO Controller
Latency
Software
Latency
~ 80 µs
~ 80 µs
~ 25 µs ~ 20 µs
~ 10 µs ~ 2 µs ~ 6 µs
~ 10 µs ~ 5 µs ~ 5 µs ~ 5 µs
SAS TLC NVME TLC NVME SCM NVME-oF
HDD
NAND SSD NAND SSD SSD SCM SSD
(Local) (Remote)
18 © 2018 NetApp, Inc. All rights reserved. --- NETAPP CONFIDENTIAL --- 18
NextGen Blocks - NVMe™
Server Storage
CPU Persistent Memory CPU Storage Class Memory
(PMEM) (SCM) as Cache
HBA/ HBA/
HCA HCA
–OR– –OR–
FC, Infiniband, Ethernet
RNIC RNIC
20 © 2018 NetApp, Inc. All rights reserved. --- NETAPP CONFIDENTIAL --- 20
FCP (SCSI) vs. NVMe™/FC Performance and Latency
21
NVMe™ Vocabulary Update
FC FC-NVMe
LUN Namespace
WWPN NQN
igroup Subsystem
* Asymmetric Namespace Access (NetApp defined multipathing
ALUA ANA* protocol for NVMe. Currently out for ratification by NVM Express®
organization.
22 © 2018 NetApp, Inc. All rights reserved. --- NETAPP CONFIDENTIAL --- 22
Ratified: Asymmetric Namespace Access
23
NVMe™ over Fibre Channel Performance Test
24
NVMe™ over Fibre Channel Performance on a A700s single node
25
Architected for Performance
IOPs
transactions on the same storage 600000
SCSI over FC
footprint” 400000
NVMe over FC
CPU %
SCSI over FC
– At 4k, we see 2x the IOPs at 50% of the 20%
NVMe over FC
CPU consumption 10%
0%
512 1k 2k 4k 8k
28
NVMe-oF™: Just runs faster
Application Latency: Response time Average Application Latency
as seen by the server application 4000
us
2000
NVMe over FC
– For this example, 32 (QD) x 32 1000
29
Performance Improvement of NVMe™ over Fabrics – End to End
NVMe/FC Vs. SCSI/FC Performance Improvement on the same hardware
Simulated OLTP Data Warehouse Batch Transaction
Workload IOPS1 IO Throughput2 Latency Test3
IO Average Latency
200 4000 200
IOPS
MB/s
100 2000 100
50 1000 50
0 0 0
NVMe/FC SCSI FCP NVMe/FC SCSI FCP NVMe/FC SCSI FCP
14K 264K Random Read IOs, 16 34K Random Read IOs, 8 Threads,
Random Read IOs, 16
Threads, Queue Depth of 16 Threads, Queue Depth of 16 Queue Depth of 1
30
31
Contact Information
For more information please contact the following:
32
Architected for Performance