Академический Документы
Профессиональный Документы
Культура Документы
White Paper
HP Thin Technologies
A Competitive Comparison
Printed in the United States of America
Copyright 2012 Edison Group, Inc. New York. Edison Group offers no warranty either expressed
or implied on the information contained herein and shall be held harmless for errors resulting
from its use.
NetApp ___________________________________________________________________ 8
Background _____________________________________________________________________ 8
HP 3PAR StoreServ ComparisonStart Thin ________________________________________ 8
HP 3PAR StoreServ ComparisonGet Thin _________________________________________ 9
HP 3PAR StoreServ ComparisonStay Thin _________________________________________ 9
As the drive to "do more with less" becomes a mantra for many organizations,
optimizing space utilization is a key goal of many IT departments. Storage continues to
be one of the major cost components of today's infrastructure deployments. Thin
technology, including thin provisioning, offers efficiency benefits that can significantly
reduce both capital and operational costs. However implementations of thin
technologies differ with the storage vendors.
HP 3PAR StoreServ is seen as a leader in thin technology, with three key aims:
1. Start Thinensure thin provisioned storage occurs with minimum overhead.
2. Get Thinensure data moved to HP 3PAR StoreServ remains thin on migration.
3. Stay Thinensure data is kept at optimal efficiency over time.
To validate this statement, a literature review, extensive customer interviews, and two
tests were performed:
1. Zero-Page-Reclaim Performancevalidation of the ability to reclaim freed resources
as part of normal operations.
2. Large Pre-allocationtest of the ability to create new storage volumes with minimal
overhead.
Overall, HP 3PAR StoreServ was the best performer in achieving the goals of "start
thin," "get thin," and "stay thin."
Objective
This report looks at thin provisioning technology from the major storage vendors in
today's marketplace. It compares the thin implementations from seven storage array
platforms, including Hewlett Packard's 3PAR storage arrays. In particular this white
paper highlights three important differentiating aspects of HP 3PAR StoreServ's thin
technology:
1. The ability to "start thin"provisioned storage is thin at deployment time.
2. Getting thinthe ability to move data from thick to thin.
3. Staying thinmaintaining thin LUNs.
Audience
Terminology
In recent years storage has become one of the major cost components within the data
center. Although the price of storage continues to fall, the rate of data growth in many
organizations continues to rise steeply, resulting in increasing costs for managing the
storage systems. Every year there is a requirement to "do more with less," using storage
more efficiently without increasing the operational budget. There are a number of key
initiatives being undertaken by organizations in order to reduce their storage
consumption. These relate directly to the use of thin technology.
Reducing WasteStorage utilization never reaches 100 percent of the physical space
provisioned from an array, as each level of configurationfrom the array to the
hostintroduces some inefficiency. Reducing waste increases utilization and allows
the deferral of additional capital expenditure.
Reducing OverheadDeploying storage isnt a quick task; from purchase order to
deployment on the data center floor, the process can take months to achieve. Storage
administrators usually keep storage in reserve in order to manage the purchase
process.
FlexibilityEnd users want the minimum disruption to their applications and as a
result, many over-order storage resources, in many cases up to 36 months ahead of
when the space is actually needed. Ideally, end users should be able to lay out their
storage needs based on growth plans and then allocate that storage on-demand.
Improving Cost EfficiencyStorage Tiering (placing data on the most cost-effective
media for the I/O profile required) is a key technology in reducing storage costs.
Dynamic tiering can be used to automate the process of data placement, based on the
use of storage pools for LUN creation. Thin provisioned LUNs directly aid the
deployment of a tiered storage model. A thin LUN is built from blocks of physical
disk capacity from within a pool of storage with metadata to associate the logical
LUN to the physical space. The physical blocks can therefore be taken from multiple
pools, where each pool represents a different tier.
For many reasons, storage utilization on hosts never reaches 100 percent utilization.
However with "thick" LUNs, physical space is reserved out on an array for the entire
size of a volume. Thin provisioned deployments can take advantage of all physical
storage available by creating more logical storage capacity than is physically available.
So called "over-provisioning" enables the utilization of physical space to be pushed to
levels higher than can be achieved in normal deployments.
The ability to over-provision storage does come with a few drawbacks. Should physical
space be completely exhausted, hosts will receive write errors, indicating a physical
problem on the array. Write failures are not usually handled gracefully by the host
operating system and can lead to system crashes. Therefore, physical versus logical
space capacity needs to be carefully managed.
Over time as files are created and deleted, thin LUNs become less efficient. This is due to
the way in which the file system on the LUN manages file allocations, free space and
metadata. Some file system implementations are more "thin friendly" than others and
are designed to re-use released space. However, housekeeping of thin provisioned
storage is required in order to maintain optimal levels of efficiency. Storage vendors
have introduced features that enable the array to recover unused storage resources:
Zero-Page-Reclaim (ZPR)A storage array identifies an entire block of storage
consisting of binary zeros, the block will be assumed to be unused and is released
back to the free pool. The ability to find unused blocks depends on a number of
factors, including the file system and array block-size and the level of file
fragmentation. Smaller array block-sizes are better for ZPR operations.
SCSI UNMAPThe UNMAP command is a low-level I/O operation that can be
used by the host to signal to the storage array that a block of storage is no longer in
use and can be released to the free pool. Unfortunately very few operating systems
currently support this feature.
The HP 3PAR StoreServ storage platform has a "thin by design" architecture that places
no restrictions on the use of either thin or thick LUNs. This applies to LUN performance,
and capacity, removing the need for the storage administrator to design the layout of the
array to cater for thin technology. HP 3PAR StoreServ is specifically optimized for thin
provisioning and contains many unique design features that enable "starting thin,"
"getting thin," and "staying thin."
RAIDHP 3PAR StoreServ arrays offer a unique RAID technology that provides
chassis high availability, and divides physical disks into "chunklets" of either 256MB
or 1GB in size. Chunklets are combined to form Logical Volumes (LVs) and
Common Provisioning Groups (CPGs) from which Virtual Volumes are created.
Thin Provisioning Virtual Volumes use a block size increment of 16KB, which is the
minimum reclaimable unit of storage within the array.
Hardware ASICNow at Generation Four, HP 3PAR StoreServ uses dedicated
custom ASIC (Application Specific Integrated Circuit) processors to perform the
identification and recovery of unused resources that can be reclaimed from thin
provisioned virtual volumes. An ASIC enables processor-intensive tasks to be
offloaded to dedicated hardware, removing the performance impact of features such
as space reclamation from the array and ensures consistent host I/O response times.
The HP 3PAR StoreServ ASIC provides a range of functions, including inline ZPR.
Thin PersistenceAn operating system task that identifies and recovers freed
resources.
Thin ConversionPerforms the migration of thick to thin volumes through a
process of inline zero detection. As data is copied to the array, zeroed blocks of data
are identified and logically mapped rather than physically written to disk.
Thin Copy ReclamationPerforms space recovery on volume copies within the
array.
Thin Reclamation APIHP 3PAR StoreServ developed the Thin Reclamation API
in partnership with Symantec. This feature allows the file system to signal when
freed resources can be released on the array. It is supported by Symantec Veritas
Storage Foundation from Version 5 onwards.
Virtualization SupportHP 3PAR StoreServ supports the VMware VAAI API,
including the "block zeroing" command.
ManagementHP 3PAR StoreServ arrays provide alerts for specific thin
provisioning space issues. Alerts are issued based on pre-defined thresholds and
enable efficient monitoring of capacity in thin environments.
EMC VMAX
Symmetrix VMAX, EMC's flagship enterprise storage platform, is the first enterprise-
class storage product to move away from custom hardware design. It uses commodity-
based Intel processors with customized hardware managing the interconnect between
storage modules. The VMAX operating systemEnginuityis an evolution of the code
developed for the first Symmetrix ICDA (Integrated Cache Disk Arrays) in 1991, and it
still retains many of the original architectural design features and constraints. The
discussion of VMAX in this section covers the latest 10K, 20K and 40K models.
Background
Thin provisioning in VMAX is implemented as a feature called Virtual Provisioning
EMC's brand name for their thin provisioning technology. Thin provisioned LUNs are
known as thin devices and take physical storage from thin pools. A thin pool is created
using standard "thick" LUNs (termed data devices), which are subdivided into allocation
units called thin device extents. Thin pools must use the same emulation and RAID
protection type and EMC recommend building them from disks of the same rotational
speed and data device size.
A thin extent is 12 tracks or 768KB in size and represents both the initial minimum
assigned to all thin devices when they are bound to a thin pool and also the lowest
increment of granularity when the capacity of a thin device is extended. HP 3PAR
StoreServ thin technology uses the much smaller increment of 16KB, which results in
less wastage, particularly with fragmented and thin-hostile file systems. Thin devices are
effectively cache-based objects that simply reference the underlying physical pool of
standard LUNs in array. These LUNs in turn, map to physical disks. A single VMAX
system supports up to 512 pools and 64,000 thin devices.
When VMAX thin devices are bound to a thin storage pool, a minimum allocation of one
thin extent (768KB) is reserved. As thin devices are effectively cache objects, each device
consumed an additional 148KB of cache, plus 8KB per 1GB based on the size of the thin
device. With HP 3PAR StoreServ thin technology, no initial space reservations are made.
VMAX Restrictions
EMC recommends a utilization level of between 60-80 percent per thin pool in order to
prevent "out of space" issues. With multiple pools (which are required for different
RAID data protection types) this can result in significant waste. HP 3PAR StoreServ thin
technology does not require separate pools for multiple protection types. When using
Synchronous SRDF with VMAX, only one active write is permitted per thin device.
Where thin devices are created into meta-devices, this can result in a performance
impact. There is also a limit of eight read-requests-per-paths for each thin device, which
can result in slow performance with high read miss rates.
Background
NetApp filers implement block-based storage within Data ONTAP by emulating LUNs
within volumes known as FlexVols. FlexVols are then created on aggregates (pools of
physical storage) and physical disk RAID groups. The underlying architecture uses a
data layout called WAFL (Write Anywhere File Layout) that operates a "write-new"
policy for both new data and updates; no block or file data is ever updated in place.
WAFL uses a page size of 4KB, storing updates in non-volatile RAM before writing an
entire "stripe" of data to disk. In this way, writes are optimized on commit-to-disk using
a RAID-4 physical disk configuration. NetApp LUNs are emulated through files on
volumes; therefore, both block and file data can be mixed within the same storage pool.
LUN creation is a simple process to achieve, however the use of block-based LUNs
involves significant complexity.
Hitachi VSP
The VSP is Hitachi's current enterprise-level storage array and is the evolution of
previous Lightning and USP-V models. The VSP retains the use of custom ASIC
technology, in which the management of storage processes is handled by Virtual Storage
Directors connected to the back-end switch matrix. Custom ASIC usage has been a
feature of all of the Hitachi storage platforms; however it isn't used directly in the thin
provisioning approach or in managing the efficiency of thin provisioned storage.
Background
VSP thin provisioning technology is known as HDP--Hitachi Dynamic Provisioning.
HDP thin LUNs (called LDEVs or logical devices) are created from a HDP pool that
comprises standard LDEV devices. In turn, LDEVs are created from RAID groups, built
from up to 16 disks in one of seven RAID-5 or RAID-6 variations. At the physical level,
data is written in tracks of 256KB per physical disk, which results in a standard logical
On VSP, thin LDEVs allocate a minimum of one 42MB page on assignment to an HDP
pool. With HP 3PAR StoreServ thin technology, no initial space reservations are made.
EMC's VNX platform is an evolution of the previous CLARiiON and Celerra products
(serving block and file protocols respectively). The two platforms were brought together
and marketed as a single platform, using one management tool, called Unisphere. Block-
based storage LUNs are presented from the base hardware unit, with file access
implemented on x-blade modules. Thin provisioning technology is implemented using
the Virtual Provisioning (VP) feature. VP extends the capabilities of LUN configuration
to include both thick and thin LUNs on the same disk pool. VP disk pools can be
comprised of large numbers of disks (greater than the standard disk pool which is
limited to 16 devices), but still configures disks in RAID groups for resiliency.
IBM XIV
The IBM XIV storage array platform was acquired from an Israeli startup, founded by
the inventor of the EMC Symmetrix, Moshe Yanai. The platform takes a radical
departure from traditional arrays and uses only high-capacity SATA or SAS hard drives,
although the configuration has recently been expanded to accelerate I/O using an SSD
cache layer. XIV is now at the third generation of hardware, utilizing either 2TB or 3TB
drives, with 6TB of SSD cache. Each array is comprised of between six and 15 server
nodes, which hold 12 hard drives each, resulting in a maximum configuration of 180
drives. Each node subdivides disks into 1MB chunks, which are then distributed across
all disks as a single large pool of mirrored data. XIV uses the terms "soft size" and "hard
size" to refer to the logical and physical size of a LUN respectively. These terms also
apply equally to pools that can be allocated physical capacity. It is possible for a pool to
deplete hard (physical capacity) and lock access to a volume, despite there being free
physical space in other pools. The overall capacity of an XIV array is referred to as the
"system hard size." A "system soft size"the degree of over-provisioning permitted at
the array levelis also defined, but can't be modified by the system administrator. This
value has to be set by an IBM engineer and requires the customer to indemnify IBM
against any issues that occur as a result of the change.
XIV supports the Symantec Thin Reclamation API for instant space reclamation,
enabling hosts running Symantec Storage Foundation, version 5 and above, to directly
signal to the array when storage is released. This requires the deployment of the Veritas
File System on each host connected to the array. Instant space reclamation does not
support mirrored volumes, volumes that have snapshots or snapshots themselves,
making the recovery process limited.
Dell Compellent
Compellent Technologies, Inc. was founded in 2002 and subsequently acquired by Dell
in 2011, from which time it was marketed under the Dell Compellent brand name. Based
on commodity components, Compellents unique offering is called Data Progression, an
automated tiered storage feature enabling migration of data between storage tiers at the
block level.
The aim of performing vendor comparison tests is to show how HP 3PAR StoreServ
compares to other vendors in terms of performance and efficiency. Although the
implementations from each vendor appear to offer similar features, the implementations
differ greatly in their performance and efficiency. The following tests were performed in
the competitive summary list.
The storage systems tested were not directly comparable for performance targets or
specifications. They varied in the number, size and types of drives, the number of
controllers and other physical specifications. Therefore, the data generated in Edison's
zero-page-reclaim performance test should only be compared for the differences for each
tested array from the test baseline. The exception is the Large Pre-Allocation results,
which demonstrate the effects of the different thin provisioning and storage
architectures on capacity utilization, rather than a change in performance for the
systems.
Details of the hardware tested can be found in the Appendix, at the end of this
document.
This test aims to show the impact of zero-page-reclaim functionality on each array. The
reclaim function is an essential property of "stay thin," ensuring that ongoing allocations
don't turn thin volumes into thick ones over time. Ideally this test should not impact I/O
performance. The test process performed the following steps:
1. Create a single large "thick" 200GB LUN and assign to a Windows host.
2. Quick format the LUN with the NTFS file system.
3. Perform load test with IOMETER, writing binary zeros to the LUN, recording IOPS
and latency figures.
4. Repeat the test with a 200GB thin LUN.
Prior to the test, the zero-page-reclaim task was enabled on the VSP system. For the
EMC platforms, the zero-page-reclaim feature was enabled by performing a LUN
migration, the method recommended by EMC.
This test aims to show the overhead at the initial creation of thin LUNs and addresses
the requirement to "start thin." Ideally the creation of thin LUNs should reserve the
minimum amount of storage possible on the array. The test process performed the
following steps:
1. Create five 200GB thin LUNs and assign to a Windows host.
2. Quick format the LUNs with the NTFS file system.
3. Measure the amount of space consumed as indicated by the array.
The data in these tables represents the performance for each array capable of zero-page-
reclaim within the test parameters. NetApp FAS was not included in this test, as the
system has no native support for zero-page-reclaim. Data for IBM XIV was not included,
because that system performs reclaim over a very long period of time that was outside
the test parameters.1
HP 3PAR 0.00%
1According to an IBM Redbook, IBM XIV Storage System: Copy Services and Migration, it,
"could take up to three weeks for used space value to decrease This is because recovery of
empty space runs as a background task."(Page 264). Not only is the time required for ZPR outside
the parameters of our research, enabling over-provisioning is, "not within the scope of the
administrator role."(Page 30) This suggests that an IBM engineer must perform an
overprovisioned configuration.
HP 3PAR 1.23%
The results of this test show that ZPR activity has an impact on both the latency and
throughput of each platform except HP 3PAR StoreServ. The greatest effect was seen on
EMC VMAX performance and the latency increase with Dell Compellent.
Edison was able to determine that performance impact shown on EMC VMAX was
because the platform needs to read each thin device extent into cache in order to
perform ZPR processing. This cache load clearly has a direct impact on array
performance.
Edison was unable to diagnose the causes of the increase latency on the Dell Compellent
system.
The HP 3PAR StoreServ array has dedicated ASICs to handle the ZPR workload without
impacting on delivering I/O to hosts.
The results from this test are shown in the following table and graph.
EMC VMAX 17
HP 3PAR 3,125
Thin provisioning is a great space optimization feature that can be used to increase
levels of utilization on storage arrays. The direct benefit is reducing capital expenditure
on hardware and operational expenditure on management. However as we have seen
from the tests, not all thin provisioning implementations are equal in terms of their
ability to optimize space with minimal impact on performance.
Best Practices
The testing and research in this white paper highlights a number of best practice
considerations:
1. Implement Zero-Page-ReclaimThis feature should be used to ensure LUNs stay
thin, however on most platforms (except 3PAR because of its custom ASIC and XIV
The following documents were referenced during the production of this white paper.
The following equipment was used to perform the testing documented in this white
paper.
Arrays
Hitachi VSP
NetApp FAS3140, running Data ONTAP 8.0.2, RAID-DP across 28 drives.
Dell CompellentRAID-5 across 72x 600GB SAS drives
EMC VNX5700 running microcode 5.31, RAID-5 across 24x 300GB 15K SAS drives.
HP 3PAR F400 InForm OS 3.1.1 (MU1)
EMC VMAX-20K
IBM XIV Gen2, 72x 1TB SATA drives
Servers
4AA4-4079ENW