Академический Документы
Профессиональный Документы
Культура Документы
Flash memory can be used in many different ways to solve real-world issues in the enterprise. This white paper examines the various options available today and in the near future with a focus on use cases for enterprise storage requirements and application acceleration.
Executive Summary
The truth today for many enterprise IT organizations is that they're still adding physical capacity to disk arrays that have long since run out of "I/O capacity." For many applications, I/O performance is far more important than total capacity. The practice of meeting a specific I/O objective by adding disk spindles to a storage subsystem has many potential pitfalls; excessive power and cooling requirements a large physical rack space required to house a sufficient number of disk drives to meet the performance need an increase in the potential for drive failure as the result of having an extended quantity of disk drives high acquisition costs without necessarily solving the problem at hand additional effort is required to effect optimal data placement for I/O performance a majority of the disk capacity can go unused in order to obtain the I/O objective
There have been many different approaches to providing solutions for these issues. A significant change in the storage industry has been driven by the increased use of SSD (Solid State Disk), also sometimes described as EFD (Enterprise Flash Disk) but in either case based on flash memory technology. For the purposes of this discussion we will refer to SSDs in the context of flash, but generically the term SSD can refer to other technologies such as those based on RAM memory. Flash memory has the potential to provide a number of benefits to resolving the trade-off between I/O and capacity that is a current challenge for many storage deployments. Flash provides extremely fast random I/O in comparison with standard disks, so its most obvious applicability is in the areas where small-block random I/O is impacted. These include databases, OLTP, Virtualization (hypervisorbased) environments and VDI (virtual desktop infrastructure). The intent of this paper is to examine some of the different ways that enterprise storage suppliers are incorporating flash into existing or new products, the advantages and limitations of the different options and some of the upcoming developments in this exciting area. The underlying focus will be to examine the ways in which the products described can be used in enterprise storage deployments and to provide some guidance in selecting the best solutions for specific requirements. What this paper does not address is the SSD market as a whole; the focus is on enterprise requirements that are characterized by a given products applicability to an enterprises needs rather than those of a consumer or a small / medium business. Some of the criteria that can apply to products intended for enterprise class use include extremely high levels of availability, non-disruptive upgrade capability, scalability in terms of performance and connectivity, and a comprehensive support program to ensure continuous access to and utilization of the products features.
Will Alger
Brief Background
The concept of using some form of semiconductor-based memory as a replacement for spinning disk drives or to augment the performance of a disk subsystem has been around for quite some time. We may think of this technology as being relatively new, however it has existed in one form or another for several decades. Products using DRAM or SRAM to mimic the functions of a disk drive have been in use since the 1970s. At one time bubble memory was seen as a legitimate competitor to disk drives. Companies such as Texas Memory Systems (TMS), Dataram and Imperial Technology were early innovators in this market. The high cost and low-density of the early generation products limited their deployment to relatively niche applications. Many factors have reinforced the dominance of disk drives as the mainstay of primary storage over time. Data stored on disk is non-volatile; it does not require that a disk be powered on to be retained. Development of new technologies that increase the areal density on disk platters has allowed for huge amounts of capacity to be provided. Disks provide relatively good performance and are a well-characterized building block that has plenty of standardization so multiple sources of supply can be interchanged relatively easily. Above all, disk drives are extremely cost effective in comparison with other types of storage devices and fulfill their role fairly well for many different applications. Disk drives have some negative characteristics however; since they are electromechanical devices they inevitably wear out. Over time, the capacity increase achieved with disks has not been matched with a corollary increase in performance so storage design with disks alone has not always kept pace with application requirements. Disk-based storage subsystems can be heavy consumers of power and cooling, a factor of concern in modern data centers. And finally, in most current disk subsystems there is a discrepancy between maximum capacity and maximum performance; very few if any can be said to provide 100% utilization, which would be characterized as being able to maintain a linear relationship between capacity and I/O. Disks at a device level and a storage subsystem level have incorporated cache to accelerate some aspects of performance. Caching helps, however it does not have universal benefit for all I/O workloads and at a subsystem level is an expensive proposition. An interesting application in some storage subsystems is the ability to treat a portion of the cache memory as a virtual disk; this can be a challenge as far as subsystem tuning since the isolation of part of the cache detracts from overall storage performance. Nevertheless, the concept of cache as virtual disk has some applicability for uses such as providing extremely high performance repositories for database re-do logs, as one example. This concept has been expanded to utilize SSDs instead of SDRAM for a similar purpose in some implementations.
Will Alger
All of these methods can potentially provide benefits to application performance, however there are some inherent limitations to flash memory itself that have to be addressed. These include finite lifespans that are limited to an average number of writes per cell of the flash device(s). The types of flash devices in general use today include SLC (single-level cell) and MLC (multi-level cell). SLC-based devices are faster, have greater longevity (100,000 writes per cell on average vs. 10,000 for MLC / 30,000 for E-[enterprise]MLC) but are more expensive and provide less density than MLC-based devices. The flash market today has some interesting dynamics, to say the least. Traditional storage manufacturers face several challenges in how they address the incorporation of flash technology in their existing products; this is because their disk array architectures are the result of many years of development that at this point are primarily oriented to managing and maximizing disk performance. Because of this it is difficult to integrate SSDs in a manner that optimizes their use in existing products. This has created a huge opportunity for smaller companies to start with a clean slate approach to the development of interesting flash-based architectures because they have no legacy product portfolio to protect. The flash market is populated with literally hundreds of start-ups with interesting concepts but little breadth in their product lines. Are these suppliers viable for enterprise applications? We need to apply the criteria of what constitutes an enterprise level solution to the various approaches and see if they really can deliver. Lets take a look at some of the representative product offerings in each of the categories we identified above.
Will Alger
Will Alger
Interestingly, other companies have incorporated Fusion-IO products into their own offerings. Some of these include Kaminario, combining their high-speed DRAM-based K2 product with Fusion-IO cards to create hybrid offerings, as well as Nutanix, a start-up company that has developed a three-tier storage device for VMware environments that includes the MLC version of the Fusion-IO ioDrive in addition to Intel SSDs and SATA disk drives. Fusion-IO is of course not the only company providing products of this type. Additional suppliers include OCZ, with some interesting board-level technology, TMS, a longtime supplier of storage-oriented devices but a relatively new player in the board market with what appear to be some extremely high-performance products, the memory manufacturer Micron Technology, and specialty provider Virident who focus on SLC-based PCIe cards. EMC, the largest supplier of storage products, announced its intent to introduce PCI-e flash products in 2012. It will be interesting to see what approach EMC takes in this market given that there have been a number of discussions about how, for example, EMC might use its VMAX storage platform to run applications under VMware, the implications of its RAPID-IO high speed interconnect technology in the VMAX and the possibility of cooperative processing between EMC storage subsystems and server-resident flash cards to cache applications.
Will Alger
Will Alger
So why hasnt EMC, or HP or IBM or any of the big existing storage providers come out with an entirely new flash-based storage architecture that really addresses this issue? Its a reasonable question but there is no simple answer. In some instances it could be that such a move would tend to cannibalize existing revenues. It could be that the engineering expertise needed to create entirely new storage architectures is not available to those companies. Perhaps the big players are waiting to see what the market really wants based on the offerings available from the dozens of start-ups and smaller companies developing all-flash or hybrid storage systems and then to develop products of a similar scope or acquire the smaller players. SSDs are now available with fibre channel, SAS (serial attached SCSI) and SATA interfaces, so they can easily be incorporated into existing storage subsystems that use those same interconnects. Thats the good news; the bad news is that the use of those disk-oriented interfaces limits the potential transfer rate and I/O performance of the flash device. That is the message of the PCI-e flash manufacturers and it is a valid point. It is an extremely good idea to take a look at the entire scope of the applications that you want to accelerate if that is your goal, and see if there are more creative ways to address those requirements rather than simply adding SSDs to an existing storage platform which may limit the value of the entire exercise.
Will Alger
Caching Appliances
The basic concept of using a caching front-end to accelerate I/O requests is simple enough; after all, its what weve been doing in our storage subsystems for a long time. The potential appeal of a flash-based caching approach is that whereas most cache architectures in storage subsystems use expensive DRAM or SRAM in relatively small amounts, a flash approach could provide significantly more cache capacity and potentially accelerate workloads transparently and without changes to our underlying storage and fabric architectures. Another possibility to consider is that a cache appliance of this sort could also be used to sit in front of more than one storage system or heterogeneous mixes of subsystems and would thereby maximize and extend the investment in those systems. Products in this category follow the traditional delineation of SAN (block level) or NAS (file level) interconnects and data handling paradigms, which makes perfect sense. There have been some failed attempts at creating products of this type and a couple of companies working in this area have already come and gone, but there are some credible players as well. Perhaps the best known products in the NAS caching / acceleration market are those from Avere Systems. Avere manufactures the FXT series of appliances that incorporate DRAM, NVRAM and flash SSDs in different model configurations in conjunction with 10Gbit and 1Gbit Ethernet ports to provide caching for NAS storage subsystems such as those from NetApp. The FXT modules can be combined to create clusters of up to 50 modules, so they are very scalable. In the SAN market, there are similar products from Dataram, the XcelaSAN series, as well as from Gridiron Systems with what it markets as the Turbocharger. Conceptually these are very similar; the caching focuses on servicing read I/O and writes are passed through to the disk back-end. This makes good use of what flash is good at, random reads with a relatively low locality of reference and bypasses the issue of write integrity by committing writes directly to disk. Generalized benefit can certainly be obtained from caching appliances given appropriate workloads and expectations. One thing to consider is that these are single purpose devices; they cannot be redeployed to address changing requirements or conditions as could a standalone flash-based storage device. As a result it would be wise to compare the TCO and ROI metrics that can be accomplished with a caching approach vs. other possible solutions.
Will Alger
Will Alger
Will Alger
January 2012 10 | P a g e
Conclusions
There is a dizzying variety of possibility in the flash market as it relates to enterprise applications today, and we can be sure of only one thing- there will be more options, more claims and more confusion in the short term. It is best to remember what we originally set out to solve and not get sidetracked by what may appear to be the latest and greatest solution; there will lots of those, and of course not all of them will live up to the hype.
Some things that can be recommended in the evaluation of flash-based products; Ignore vendor claims of performance; do your own comparisons Look for real-world examples that closely correspond to your own needs Realize that support and commitment from your selected provider of storage are more important than other factors that may be temporarily desirable such as apparent low cost Scalability is important; if the upgrade path is buy another one then you need to look at other options Reliability is paramount; dont accept design compromises that can affect the integrity of your data Look for account references running applications similar to your own and gauge their experiences
Will Alger
January 2012 11 | P a g e