1210 ST Eg RAID
1210 ST Eg RAID
STORAGE
SEARCHSTORAGE.CO.UK
ESSENTIAL GUIDE TO
RAID Configuration
Evaluate the myriad options for protecting your data with RAID.
1
,
0 0
Implementation Choices
Alternatives to RAID
Sponsor Resources
DATA LIES AT the heart of the organisation, and data protection takes many forms backup, archiving and the subject of this SearchStorage.co.UK Essential Guide to RAID Configuration. The Redundant Array of Independent Disks was possibly the first-ever form of storage virtualisation. By taking many disks and aggregating them to present as virtual volumes, storage arrays were able to boost I/O performance and increase levels of data protection against drive failure. In fact, by means of striping, mirroring and parity (all explained in this Essential Guide), the varying levels of RAID are able to multiply by many times the resilience and performance of a collection of single hard drives, adding up to much more than a sum of the parts. Located at the heart of all data Located at the heart of all storagethe drive arrayRAID is the data storagethe drive first line of defence against mechaniarrayRAID is the first line of cal disk failure, and for this reason knowledge of RAID is a fundamental defence against mechanical for all storage professionals. disk failure, and for this At the same time the many RAID levels provide numerous possible perreason knowledge of RAID mutations of performance, data prois a fundamental for all tection and cost, and the challenge for storage professionals is to know which storage professionals. is best-suited to their environment. The various RAID levels offer multiple choices when configuring storage arrays: striping data across many drives, dividing files by block or even at byte level; mirroring sets of drives; mirroring striped sets; striping mirrored sets; and using parity dataon a dedicated disk or disks or striped across many to enable drive rebuilds in case of a disk failure. The profusion of possible combinations of RAID characteristics, which all have different impacts on performance, resilience and cost, plus the fundamental importance of RAID to storage array management, is the reason weve compiled this Essential Guide for SearchStorage.co.UK readers. Our RAID content is perennially popular. Here you can learn the fundamentals of storages most basic form of data protection in one handy guide. 2
Antony Adshead is the bureau chief for SearchStorage.co.UK.
Copyright 2010, TechTarget. No part of this publication may be transmitted or reproduced in any form, or by any means, without permission in writing from the publisher. For permissions or reprint information, please contact Mike Kelly, VP and Group Publisher (mkelly@storagemagazine.com).
choices
RAID: Basic Data Protection Implementation Choices Alternatives to RAID Rating RAID Levels 0 to 6 Sponsor Resources
RAID implementation
Learn about the key decisions that youll confront when implementing RAID, including whether a software or hardware approach makes more sense for your organisation, as well as which level is best.
By ANTONY ADSHEAD
When implementing RAID, storage professionals face a lot of choices. In this article we examine key decisions that need to be made: whether to opt for hardware or software RAID; how to select RAID levels based on cost and performance; and, by way of a more detailed example, how to decide whether RAID 6 or RAID 10 is best suited to your needs.
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
STORAGE
RAID 1 and 10 win on data protection but lose in terms of disk costs. RAID 10 offers the best performance and data protection, but at a cost. RAID 5 offers the best tradeoff in terms of price and performance and includes data protection for database use.
RAID 6 VS RAID 10
When comparing RAID 6 vs RAID 10, many storage managers have a hard time deciding between the dual parity of RAID 6 and the mirrored data of RAID 10. RAID 6 stripes data across disks and calculates dual distributed parity. Distributed parity provides fault tolerance against two drive failures. Dual parity means that while a failed disk is being rebuilt the array is still protected by the remaining parity data. RAID 1+0 (RAID 10) is mirrored sets in a striped set. RAID 1+0 creates a striped set from subsets of mirrored drives. If disks fail, RAID 1+0 allows all the remaining disks to continue in use. The array can suffer multiple drive failures as long as no mirror set loses all of its drives. To help you choose the RAID level that best meets the needs of your organisation, lets take a look at some of the advantages of RAID 6 vs RAID 10. RAID 6 gives more useable capacity the more disks you add. Because RAID 10 mirrors everything, an array requires Because RAID 10 mirrors everything, an array requires double the disk cadouble the disk capacity of pacity of the data to be stored. The remainder of the capacity constitutes the the data to be stored. mirror. If a RAID 6 array comprises four disks, only 50% of that space is available as useable capacity, but the proportion of useable space increases as you add more drives. That means half the total capacity of a RAID 10 array will always be dedicated to protection, but with a RAID 6 array the useable capacity grows as the number of drives increases. For example, if you increased the number of disks in a RAID 6 array from four to eight, the space consumed by parity data would decrease from 50% to 25%. RAID 6 requires more processing power. RAID 6 makes two parity calculations for each write operation, so its slower to write than most other RAID levels.
Implementation Choices
Alternatives to RAID
Sponsor Resources
STORAGE
Implementation Choices
Alternatives to RAID
RAID 6 can always protect against two simultaneous disk failures. Because RAID 6 doubles up its parity data, it can withstand two disks failing at the same time. Whether RAID 10 can handle two disk failures simultaneously depends on where they occur. If both the disks that fail are located in the same mirror, the other set can step in. You will lose all data if the same disks in both mirrors fail within the rebuild window (which should be relatively short, however). RAID 10 rebuild times are faster. RAID 10 has among the fastest rebuild times possible because it only has to copy from the surviving mirror to rebuild a drive, which can take as little as 30 minutes for drives of approximately 1 TB. The key drawback of RAID 6 (vs RAID 10) is that the time it takes to rebuild the array after a disk failure is lengthy because of the parity calculations required, often up to 24 hours with even a medium-sized array. RAID 10 doesnt need special hardware. Most controller hardware will support RAID 10 with good performance. Because RAID 6 doubles the parity calculations for every write, it requires specially designed controller hardware. 2
Antony Adshead is the bureau chief for SearchStorage.co.UK.
Sponsor Resources
Memorising RAID level definitions and knowing which level does what can be:
Confusing Hard to Remember Useful All of the above
So how much do you think you know about RAID? Find Out For Yourself Read this Essential Guidethen Test Your Knowledge with Our Exclusive RAID Quiz! And dont forget to bookmark this page for future RAID-level reference.
ALTERNATIVES
TO RAID
RAID: Basic Data Protection
Implementation Choices
Alternatives to RAID
The various forms of RAID have been around for a long time and have done a good job of protecting data. But high-capacity drives and new performance demands have spurred development of RAID alternatives.
By MARC STAIMER
for disk-based data protection since 1989 and is a proven and reliable method thats considered a basic data storage building block. Basic storage principles tend to change very slowly, and, despite its popularity and track record, change is coming to RAID. To gain more insight into why alternatives to RAID might be appealing requires some understanding about RAID and the growing problems with the technology.
Sponsor Resources
11
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
Each RAID type has tradeoffs in write performance, read performance, level of data protection, speed of data rebuilds and the usable storage on each hard disk drive. For example, if guaranteeing data availability is the top priority, then some variation of mirroring or multiple mirrors (RAID 1, 10, triple mirror, etc.) will be required. Having full copies of the data on other HDDs or RAID sets simplifies protection and recovery of the data but at a severe and tangible cost because each mirror reduces usable storage by the same amount of the original data. In addition, system resources are required for every copy, which can impair I/O performance. Realistically, most organisations arent this overprotective; most use RAID 5 and/or RAID 6. When a HDD fails in a RAID 5 set, the system rebuilds the data on a spare drive that replaces the one that failed. The storage system then analyses every sector on every HDD in the RAID set to reconstruct the data. Such heavy utilisation of the other HDDs in the RAID set increases the likelihood of another HDD failure (usually a non-recoverable read error) by an order of magnitude, and this significantly increases the likelihood of data loss. Ten or 20 years ago, when disk Ten or 20 years ago, when capacities were much lower, rebuilds disk capacities were much were measured in minutes. But with disk capacities in the terabytes, relower, rebuilds were measbuilds can take hours, days or even ured in minutes. But with disk weeks. If application users cant tolercapacities in the terabytes, ate the system performance degradation that rebuilds cause, the rebuild is rebuilds can take hours, given a lower priority and rebuild times days or even weeks. increase dramatically. Longer data reconstruction times typically equate to a significantly higher risk of data loss. Because of this, many storage shops are stepping up their use of RAID 6. RAID 6 provides a second set of striped parity blocks that protect the data even if two HDDs fail or suffer a non-recoverable read error in the RAID set. The risk of data loss drops dramatically, but the extra stripe consumes additional usable capacity and system performance takes a bigger hit if two drives must be reconstructed simultaneously from the same RAID group. More disturbing is the increased risk of data loss if a third HDD fails or a non-recoverable read error occurs during the rebuild.
12
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
There are other RAID issues such as bit rot (when HDDs acquire latent defects over time from background radiation, wear, dust, etc.) that can cause a data reconstruction to fail. Most storage systems include some type of background scrubbing that reads, verifies and corrects bit rot before it becomes non-recoverable, but scrubbing consumes system resources. And higher capacities mean more time is needed to scrub. Another onerous RAID issue is that of documenting the chain of ownership when replacing a failed HDD. This comprises the documented trail (who, what, where, when) of the failed HDD from the time it was pulled to the time it was destroyed or reconditioned. Its a tedious, manually intensive task thats a bit less stringent if the HDD is encrypted. Even more frustrating is that the vast majority of failed HDDs sent back to the factory for analysis or reconditioning (somewhere between 67% and 90%) are found to be good or no failure is found. Regrettably, the discovery happens after the system failed the HDD, the HDD was pulled, the data was reconstructed and the chain of ownership documented. Thats a lot of operational pain for no failure found. Solid-state storage devices actually exacerbate the aforementioned RAID problems. Because solid-state drives (SSDs) can handle high-performance applications, they allow for storage systems with fewer high-performance HDDs and more high-density, low-performance hard disk drives. Tom Georgens, NetApps CEO, recently noted, Fast access data will come to be Fast access data will come stored in flash with the rest in SATA to be stored in flash with drives. Lower cap-ex and op-ex for the system can end up translating into the rest in SATA drives. Tom Georgens, CEO, NetApp higher op-ex because of the increase in RAID problems. These RAID issues have inspired numerous vendors, academics and entrepreneurs to come up with alternatives to RAID. We can categorise those innovative alternatives into three groups: RAID + innovation, RAID + transformation and paradigm shift.
RAID + INNOVATION
Several vendors have addressed traditional RAID problems by taking an incremental approach to RAID that leverages its reliability while diminishing some of its shortcomings. IBMs EVENODD (implemented
13
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
by EMC on Symmetrix DMX) and NetApps RAID-DP (implemented on NetApps FAS and V-series) have enhanced RAID 6 by reducing algorithm overhead while increasing performance. NECs RAID-TM, or triple mirror (implemented in its D-Series systems), aims to solve RAID 1 data loss risk if both the primary and mirror drive fail or if theres a non-recoverable read error. RAID-TM writes data simultaneously to three separate HDDs so if two HDDs fail or there are unrecoverable read errors in the same mirror, the app still has access to its data with no degradation in performance even while the drives are rebuilt. The advantage is performance; the disadvantage is far less usable capacity. RAID-X is an IBM XIV Storage System RAID-X is an IBM XIV Storage innovation that uses a wide stripe to System innovation that uses reduce RAID tradeoffs of performance and data loss risk. Its basically a variaa wide stripe to reduce RAID tion of RAID 10 that uses intelligent risk tradeoffs of performance algorithms to randomly distribute block mirrors throughout the entire array. This and data loss risk. approach allows XIV to reconstruct the data on very large, 2 TB HDDs in less than 30 minutes. As with all mirroring technology, the tradeoff is reduced usable capacity. Hewlett-Packards LeftHand Networks and Pivot3 provide similar variations of Network RAID for their x86-based clustered iSCSI storage. Network RAID leverages the concept of RAID but uses storage nodes as its lowest component level instead of disk drives. This allows it to distribute a logical volumes data blocks across the cluster with one to four data mirrors depending on the Network RAID level. Ongoing block-level, self-healing nodal health checks allow Network RAID to copy and repair the data to another node before a failure occurs. This decreases the probability of a hard disk drive fault or nonrecoverable read error causing a performance-sapping rebuild; but like all mirroring technology, it reduces the amount of usable storage. These are just some of the RAID + innovation technologies. Others are currently incubating, including proposals for RAID 7 (triple parity and more) or TSHOVER (triple parity).
RAID + TRANSFORMATION
There are also RAID alternatives that attempt to re-invent RAID. They typically use RAID and are layered on top of it in some way. The concept is to keep whats good about RAID and fix the rest.
14
STORAGE
RAID ENHANCEMENTS
Max. HDD failure protection
RAID 6 EVENODD An IBM innovation that uses only two additional redundant HDDs and consists of simple exclusive-OR computations. The advantage of EVENODD is that it only requires parity hardware, which is typically present in standard RAID 5 controllers. This reduces the number of exclusiveOR operations over the more common Reed-Solomon computations at approximately 50% (based on 15 drives). EVENODD has similar performance issues of RAID 6 when dual hard disk drive (HDD) rebuilds are occurring. One advantage of EVENODD is that its based on XOR. A disadvantage is that it has a few hot spots in certain diagonal blocks that cause very poor short write performance. Stores row parity across the HDDs in a RAID 4 group; the additional parity HDD stores diagonal parity across the HDD in a RDP group. The two RDP parity stripes provide data protection in the event of two HDD failures occurring within the RAID group. Performance is nearly equal to single-parity RAID 4 or RAID 5. Higher performance than standard RAID 6 but with similar performance issues when concurrently rebuilding two HDDs. RAID-TM delivers the high speed of RAID 1 while providing the high reliability and double HDD fault protection of RAID 6. RAID-TM writes data simultaneously to three separate HDDs. Even with two HDD faults or unrecoverable read errors in the same mirror, the application still has access to its data with no degradation in performance even while drives are rebuilt. RAID-X doesnt require a spare HDD, just spare capacity on existing HDDs in the storage system. The objects can be mirrored between any two types of HDDs (no need to match drive size or speed). Rebuild performance is extremely fast because data is mirrored. This is a variation of RAID 1 or RAID 10, but with the added protection of random distribution. A second drive failure can result in data loss that can only be mitigated with additional mirrorings. Useable storage can be restricted depending on the number of mirrors (minimally half). Lays out a logical volumes blocks across the cluster, providing reliability configurable on a per-volume basis to best meet the needs of each applications data. Depending on a logical volumes Network RAID level, one to four copies of each block are striped across the cluster. A volumes RAID level can be changed (auto-restriping) without data availability interruption. Also provides proactive block-level self-healing to decrease probability of a non-recoverable read error. However, each block copy reduces the amount of useable storage. Also know as heal-in-place storage. Uses series of automated repair sequences designed to eliminate or mitigate the majority of HDD failures and unnecessary RAID data rebuilds. Isolates HDD sectors it cant fix and rebuilds only the data lost on those sectors. More expensive upfront than traditional RAID, but with a much lower total cost of ownership. Heal-in-place storage requires a pool of unused HDDs for the fail-in-place capability. BeyondRAID is essentially a virtualization engine on top of RAID that chooses the correct RAID algorithm based on the data protection required. It writes blocks that can actually alternate between data protection methodologies. If more storage capacity is required, additional HDDs can be inserted or small HDDs can be replaced with larger ones. Simple administration allows switching from single- to dual-disk redundancy with a single click. Protects against dual drive failures and adds transparent automatic data healing. Its data-aware, allows for mixed drive sizes, drive reordering, proportional rebuild time and self-management. Only available for small systems of up to eight drives. Also known as a form of forward error correction (FEC), erasure coding adds additional information to a stored object that allows any data set to be completely resurrected from a subset of the total information. Multiple slices (storage objects) or subsets of a data set are distributed across multiple storage or server nodes. Additional information attached to a stored object equals greater resiliency of the data set, protecting against larger numbers of components (disk drives, storage objects or server nodes) that can be lost and still recover the complete data set. The additional information on each storage object also reduces the amount of useable storage. The biggest issue with erasure code-based storage is reduced write performance, especially small writes. 2
Vendors
EMC
NetApp
RAID + Innovation
NEC
Implementation Choices
IBM (XIV)
Alternatives to RAID
Network RAID
Rating RAID Levels 0 to 6
Sponsor Resources
Self-healing storage
Atrato, DataDirect Networks, Panasas, LSI, NEC, 3PAR and Xiotech Data Robotics
RAID + Transformation
BeyondRAID
Erasure codes
15
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
Examples of transformation technologies include self-healing storage and BeyondRAID. Self-healing storage. Xiotechs Intelligent Storage Elements (ISE) is a good example of self-healing storage. ISE tightly integrates RAID and HDDs and combines them into a single storage element. Xiotech engineered ISE to resolve most RAID rebuild issues by eliminating 67% to 90% of the rebuilds. It starts by reducing HDD faults by proactively healing hard disk drives before a fault occurs using similar HDD reconditioning algorithms employed by the factory. It also uses advanced vibration conXiotech engineered ISE to trols and sealed systems called Dataresolve most RAID rebuild Pacs to reduce the possibility of outside influences that cause HDD faults. issues by eliminating 67% When a fault does occur, it reacts by to 90% of the rebuilds. providing remedial component repair within the sealed DataPac using methods similar to those the original manufacturer uses. It analyses power cycles, recalibrates components, remanufactures the HDD and migrates data when required to other sectors or HDDs. If the fault persists, ISE isolates only the non-recoverable sectors and then initiates data reconstruction only for the faulty HDD sectors. So, there are far fewer rebuilds, and when one is required theres much less to reconstruct. In addition, its all automated so no manual intervention to pull failed drives is required. The result is equivalent to a factory-remanufactured HDD with only the components that are beyond repair taken out of service. The downside to this transformational technology is that it has higher up-front costs, although it lowers the total cost of ownership (Xiotech provides a five-year warranty). Atratos Velocity1000 (V1000) uses a self-healing technology called Fault Detection, Isolation Recovery (FDIR) in combination with Atratos Virtualization Engine (AVE). FDIR watches component and system health and adds self-diagnostics and autonomic self-healing, but it doesnt attempt to remanufacture or recondition HDDs in place as Xiotech does. Atrato puts 160 2.5-inch SATA drives in a 3U system called SAID (self-maintaining array of independent disks). The company uses its extensive SATA drive performance database of operational reliability testing (ORT) to monitor the installed drives actual performance to detect SATA HDD deviations.
16
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
Atrato also deals with HDD faults by first attempting to repair the faulting HDD sectors (although not with manufacturer-level reconditioning, remanufacturing or component recalibration). If the fault or non-recoverable read error cant be repaired, the sector is isolated and only the affected data is reconstructed and remapped to virtual spare capacity. If a disk drive completely fails, it is reconstructed and remapped to the virtual spare capacity. Atrato reduces the number of rebuilds and rebuild times by reconstructing only affected data on virtual drives. Atrato backs its technology with a three-year warranty. DataDirect Networks DDN S2A techDataDirect Networks DDN nology heal-in-place approach to disk S2A technology heal-infailure attempts several levels of HDD place approach to disk failure recovery before a hard disk drive is removed from service. It begins by keepattempts several levels of ing a journal of all writes to each HDD HDD recovery before a hard that show behavior aberrations and then attempts recovery operations. disk drive is removed from When recovery operations succeed, service. only a small portion of the HDD requires rebuilding using the journaled information so rebuild times are reduced and a service call may be avoided. Panasas ActiveScan technology continuously monitors HDDs and their contents to detect problems. ActiveScan monitors data objects, RAID parity, disk media and disk drive attributes. When a potential problem is detected, data is moved to spare blocks on the same disk. Future HDD failure is predicted through the use of HDD self-monitoring analysis and reporting technology (SMART), which permits action to be taken to protect data before a failure occurs. When a HDD failure is predicted, user-set policies pre-emptively migrate data to other HDDs, which eliminates or mitigates the need for reconstruction. LSI and NEC both detect HDD sector errors while allowing operations to continue with the other drives in the RAID group. If an alternative sector can be assigned, the HDD is allowed to return to operation, avoiding a complete rebuild. Performance is maintained throughout the detection and repair process. This is a limited selfhealing technology that reduces the number of rebuilds and helps maintain performance. 3PARs InSpire Architecture is engineered to sustain high performance levels by leveraging advanced HDD error isolation to reduce the
17
STORAGE
Implementation Choices
Alternatives to RAID
amount of data that requires reconstruction and by taking advantage of its massive parallelism to provide rapid rebuilds (typically fewer than 30 minutes). The system uses chunklets in their many-tomany drive relationships. That same massive parallelism allows 3PAR to isolate RAID sets across multiple drive chassis to minimize the risk of data loss if a chassis is lost. BeyondRAID. Data Robotics BeyondRAID sits on top of RAID and makes it completely transparent to the administrator. It transforms RAID from a deterministic offline process into one thats online and dynamic. Essentially self-managing, BeyondRAID chooses RAID sets based on the required data protection at any given point in time. But its how BeyondRAID addresses Data Robotics Inc.s RAID issues that truly makes it stand BeyondRAID sits on top out. It protects against one or two HDD failures and has built-in automatic of RAID and makes it data self-healing (not storage selfcompletely transparent healing). Data blocks are spread to the administrator. across all drives so data reconstruction is very fast. Because the system is data aware, it allows for different drive sizes, drive re-ordering and proportional rebuild times. Because it tops out at eight SATA drives, it is most appealing for small and medium-sized businesses (SMBs), but it is a true fireand-forget storage system.
Sponsor Resources
18
STORAGE
reconstitute the original data. In actuality, with erasure codes there can be multiple simultaneous failures across a string of hosting devices, servers, storage elements, HDDs or networks, and the data will still be accessible in real time. Also known as forward error correction (FEC), erasure coding storage is a completely different approach to RAID. Erasure codes eliminate all the RAID issues described here. Its a new approach and at this time only three vendors have erasure code-based products: Cleversafes dsNet, EMCs Atmos and NECs HYDRAstor. Erasure codes appear to be better suited for large data sets than smaller ones. Its especially appropriate for cloud or distributed storage because it never has to replicate a data set and can distribute the data over multiple geographic locations.
Implementation Choices
RAIDS EVOLUTION
The issues with traditional RAID are well known and are escalating with higher disk capacities. The RAID alternatives described here address many of those problems, and more new approaches are on the way. Selecting the best fit for a particular environment requires research, testing, pilot programmes, patience and a willingness to take a risk with a non-traditional approach. 2
Marc Staimer is founder and senior analyst at Dragon Slayer Consulting in Beaverton, Ore., a consulting practice that focuses on strategic planning, product development and market development for technology products. You can contact him at marcstaimer@mac.com.
Alternatives to RAID
Sponsor Resources
19
Alternatives to RAID
Sponsor Resources
Compare RAID levels and learn about each ones tradeoffs in performance and data protection.
By ARUN TANEJA and ANTONY ADSHEAD
single hard drive to the operating system, RAID technology allows storage pros to store the same data in different places on multiple disks. I/O operations can therefore be shared between drives, which means performance can increase and data storage protection can improve. For organisations considering a RAID deployment, there are multiple factors that need to be looked at, particularly the available levels of RAID technology and the specific needs of their data storage infrastructure. Lets look at the functions of and differences between the various RAID levels.
21
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
22
STORAGE
In RAID 0+1, the array will continue to operate if one or more drives in the same mirror set fail. But, but if drives on both mirror sets fail, all the data is lost. If disks fail, RAID 1+0 performs better than RAID 0+1 because all the remaining disks continue in use. The array can suffer multiple drive failures as long as no mirror set loses all drives.
Implementation Choices
Alternatives to RAID
Sponsor Resources
23
STORAGE
Implementation Choices
Alternatives to RAID
Sponsor Resources
24