Вы находитесь на странице: 1из 20

EMC Support Matrix

Interoperability Results
P/N 300-000-166 Rev 41

Table of Contents
Copyright EMC Corporation 2012...................................................................................................................................................................................................1 EMC's Policies and Requirements for EMC Support Matrix ........................................................................................................................................................2 Selections .........................................................................................................................................................................................................................................3 Base Connectivity ............................................................................................................................................................................................................................4 EMC Virtual Provisioning Support.................................................................................................................................................................................................5 HBA - Cable Compatibility..............................................................................................................................................................................................................6 Director Bit/Flag Information..........................................................................................................................................................................................................7 Fibre Connectivity: Switch ..............................................................................................................................................................................................................8 Switched Fabric Topology Parameters.........................................................................................................................................................................................9 WAN Optimization Controllers.....................................................................................................................................................................................................12 Path Management Software ..........................................................................................................................................................................................................13 AppSync Virtualization Server Solutions....................................................................................................................................................................................14 Replication Manager Block Storage............................................................................................................................................................................................15 Replication Manager Virtualization Server Solutions................................................................................................................................................................16 NAS Gateway Connectivity ...........................................................................................................................................................................................................17 Storage Array Properties..............................................................................................................................................................................................................18

Copyright EMC Corporation 2012


Copyright 2001, 2012 EMC Corporation. All Rights Reserved.

EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice.

THE INFORMATION IN THIS PUBLICATION IS PROVIDED "AS IS". EMC CORPORATION MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.

Use, copying, and distribution of any EMC software described in this publication requires an applicable software license.

Trademark Information

EMC2, EMC, EMC ControlCenter, AlphaStor, ApplicationXtender, Catalog Solution, Celerra, CentraStar, CLARalert, CLARiiON, ClientPak, Connectrix, Co-StandbyServer, Dantz, Direct Matrix Architecture, DiskXtender, Documentum, EmailXtender, EmailXtract, HighRoad, Legato, Legato NetWorker, Navisphere, OpenScale, PowerPath, RepliStor, ResourcePak, Retrospect, Smarts, SnapShotServer, SnapView/IP, SRDF, Symmetrix, TimeFinder, VisualSAN, VSAM Assist, Xtender, Xtender Solutions, and where information lives are registered trademarks and EMC Developers Program, EMC OnCourse, EMC Proven, EMC Snap, EMC Storage Administrator, Access Logix, ArchiveXtender, Authentic Problems, Automated Resource Manager, AutoStart, AutoSwap, AVALONidm, C-Clip, Celerra Replicator, Centera, CLARevent, Codebook Correlation Technology, Common Information Model, CopyCross, CopyPoint, DatabaseXtender, Direct Matrix, DiskXtender 2000, EDM, E-Lab, EmailXaminer, Enginuity, eRoom, FarPoint, FLARE, FullTime, Global File Virtualization, Graphic Visualization, InfoMover, Invista, MirrorView, NetWin, NetWorker, OnAlert, Powerlink, PowerSnap, Rainfinity, RecoverPoint, RepliCare, SafeLine, SAN Advisor, SAN Copy, SAN Manager, SDMS, SnapImage, SnapSure, SnapView, StorageScope, SupportMate, SymmAPI, SymmEnabler, Symmetrix DMX, UltraPoint, Viewlets, VisualSRM, and WebXtender are trademarks of EMC Corporation. All other trademarks used herein are the property of their respective owners.

Copyright EMC Corporation 2012

EMC's Policies and Requirements for EMC Support Matrix


EMC's Support Matrix Policies and Requirements

This section describes EMC's policies and requirements for the EMC Support Matrix document.

Purpose and Limitations of this Document

This document is being provided for informational purposes only and may change at any time. This version supercedes and replaces all previous versions. The information is to serve only as a guide for those configurations/products which EMC has qualified. This document identifies and lists various vendor host systems and integral components that have been tested and qualified by EMC for use with EMC products. Vendor components include, but are not limited to, host operating systems, HBAs (host bus adapters) and associated drivers, firmware and BIOS; Fibre Channel switches, hubs and bridges. This document also lists various vendors, cluster software, tape hardware, storage system boot procedures, heterogeneous information storage, and configurations that EMC has tested and/or qualified for use with EMC products.

The information included in this document is intended as a guide in the configuration of systems for EMC's products. It is not intended to be the sole resource for system configuration. For more information or answers to questions not found in this document, please see EMC's website, other EMC documentation, such as EMC's Host Connectivity Guides, EMC Networked Storage Topology Guide, EMC CLARiiON Open Systems Configuration Guide, relevant vendor documentation, or contact your EMC Sales or EMC Customer Service representative. For Symmetrix 4.x support, please look at the Symmetrix 8000 Series section and check the footnotes for model/microcode compatibility. EMC employees can obtain older revisions of ESM containing more Symmetrix 4.x information from the Symmetrix Service Readiness group. For a complete listing of any known issues with the current support matrix content, please review the ESM Errata on E-Lab Interoperability Navigator. To report any issues with the content contained within the support matrix, follow the steps documented on the Contacts link on E-Lab Interoperability Navigator. For E-Lab Interoperability Navigator access through Powerlink: http://powerlink.emc.com/community/tools/index.jhtml?community_id=2&categoryId=esn_builder&categoryName=E-Lab+Navigator. For access through Avatar: http://avatar.eng.emc.com/Avatar.asp.

Policies for Qualifying Systems

Policies and procedures for support of EMC products are set forth in the customer's applicable agreements. EMC's publication of information relating to system configurations covers only those outlined in this document or by approval from EMC Engineering. Other system configurations not found in this document are not qualified without EMC Engineering approval. EMC has qualified hardware and software provided only by the vendors listed in this document, and host systems, hardware and software from other vendors are not qualified, and may never be. Please contact your EMC Sales or EMC Customer Service representative for updates or information not included in this document.

EMC maintains a large collection of the products listed above as well as third-party application software for qualification with EMC's storage systems and to simulate customer environments, but you must consult the vendors for information about their system internals, such as hardware and associated drivers.

Policy for Future Qualification of Operating System Software Releases

The information in this document is maintained by EMC and EMC strives to update this document with new releases of hardware, operating systems, firmware, BIOS, switch software, etc., as they become available from the vendors. EMC works with the vendors during their development and release processes in order to be fully informed at the time the vendors release new versions. EMC does not announce qualification prior to vendors' General Availability. In some cases, EMC may choose to perform regression testing following the vendors' GA; in these cases and in cases where advance preparation is not possible, EMC may test and qualify vendors' products as appropriate after they have been released. For early support of such products, including vendor beta participation, contact your EMC Sales or EMC Customer Service representative.

Policy for End-of-Life Support

EMC strives to continue support for any installed platforms, but may remove support from this host matrix for new installations within three months after the vendor has announced that this platform has reached the end of life. EMC will continue support for existing installations of hardware, operating systems, and components that the vendor has officially declared to have reached end of life as long as support is available from the vendor.

EMC's Policies and Requirements for EMC Support Matrix

Selections
Your Selections: Host Bus Adapter Operating System Storage Array Switch HPQ Emulex LPe1105-HP 4Gb FC HBA for HP c-Class Bladesystem (403621-B21), HPQ QLogic QMH2462 4Gb FC HBA for HP c-Class Bladesystem (403619-B21), HPQ QLogic QMH2562 8Gb FC HBA for HP c-Class Bladesystem (451871-B21) VMware ESXi 5.1(vSphere 5.1) EMC Symmetrix VMAX 20K HPQ Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Relevant Documentation or Links Symmetrix VMAXe series delivers a multi-controller, scale-up and scale-out architecture at an affordable price. VMAXe uses the same Virtual Matrix Architecture as the Symmetrix VMAX series, breaking through the physical boundaries of fixed backplane storage architectures to deliver the highest levels of reliability, availability, and serviceability. Designed to be 100 percent virtually provisioned, Symmetrix VMAXe storage is automatically configured and allocated. With Fully Automated Storage Tiering for Virtual Pools (FAST VP), Symmetrix VMAXe automatically moves data to the right place, at the right time. Symmetrix VMAXe includes bundled software for intelligent and easy operation, as well as local and remote disaster recovery capabilities with EMC TimeFinder for VMAXe and an integrated splitter for EMC RecoverPoint. Symmetrix DMX series continues to deliver innovation and advanced capabilities. It provides highly available, high-performance, full mainframe support; and flexible connectivity options, including Fibre Channel, FICON, ESCON, Gigabit Ethernet, and iSCSI. Organizations of all sizes rely on Symmetrix for constant data access, business continuity, and massive consolidation. Available under Products > Hardware/Platforms > Symmetrix.

Symmetrix VMAXe Series

Symmetrix DMX Series Product Information

Symmetrix VMAX, the world's most powerful, trusted, and intelligent storage system, is built for the virtual data center. Based on the Virtual Matrix Architecture, Symmetrix VMAX scales performance and capacity to unprecedented levels, delivers nondisruptive Symmetrix VMAX operations, and greatly simplifies and automates the management and protection of information. Fully Automated Storage Tiering for Series Product Virtual Pools (FAST VP) automatically maximizes performance and reduces costs by placing active information on Flash drives and Information inactive information on SATA drives. FAST VP enables organizations to use a combination of Flash, Fibre Channel, SAS, and SATA drives together in Tier 1 environments to optimize performance and reduce costs. E-Lab Advisor (ELA) Storage Target Revisions and Adoption Rates EMC Powerlink E-Lab Issue Tracker The E-Lab Advisor (ELA) tool has integrated the functionality of HEAT, SWAT, Celerra Health Check, SANsummary and SYMAPI Log Analyzer tools. These standalone tools were decommissioned on September 30, 2011. Unlike those tools, E-Lab Advisor features a single upload page. You upload the file, and it will run the corresponding report. To ensure stable and reliable environments, EMC recommends target codes for each EMC Storage Operating Environment (SOE). This report lists the target codes and customer adoption rates for SOEs: Celerra DART, Centera CentraStar, CLARiiON FLARE, Invista, and Symmetrix Enginuity. Connectrix target codes are included. Up-to-date technical documents about EMC's products and services, including microcode release notes, product manuals, software release notes and more. See the "Quicklinks" section of the Powerlink home page. Up-to-date information on EMC product issues that could have an impact on your operations.

VMware ESX Host Connectivity Complete, authoritative information on how to configure VMware ESX environments. Guide

Selections

Base Connectivity
Starting with Symmetrix VMAX, Unified and CLARiiON CX4, minimum link speed supported for Fibre Channel direct connection is 2Gb/s. EMC supports Symmetrix and CLARiiON configured as a boot device for the servers where listed below, provided that these requirements are followed. This will ensure the best possible response times for boot/root/swap volumes. EMC recommends shutting down the host server during maintenance procedures that could cause the boot disk to become unavailable to the host. If the paging file is unavailable to the operating system, a crash can occur. EMC supports online microcode upgrades for Windows systems booting from Symmetrix and non-disruptive upgrades (NDU) for Windows systems booting from CLARiiON arrays. Refer to the specific OS Host Connectivity Guide for best practices. Any issues that arise when trying to use the boot LUN on a host that is not identical to the source host is outside the scope of what EMC can support. In case of environment failure where I/O is not written back to the source LUN, or if the source LUN is not left in a stable state by the source host, the source LUN, or any replicated LUN of the source, is not guaranteed to be usable. All external plug-in adapter cards (FC HBAs/FCoE CNAs) listed in the "Base Connectivity" table below, are supported for connectivity to the external switches listed in the "Switched Fabric Topology Parameters" table. All mezzanine internal adapter cards (FC HBAs/FCoE CNAs/NICs) that are supported by the respective blade server vendors and are listed in the "Base Connectivity" table, are supported for connectivity to the respective embedded blade server switch modules listed in the Blade Server ESSM. (The HBA mezzanine card can connect to the FC switch and pass-thru modules, the CNA mezzanine card to the FIP capable convergence modules and the NIC mezzanine cards to the Ethernet pass-thru modules). Configurations found here or in the ESM are supported and no RPQ is required.

No.
1

Host System

Operating System

Host Bus
PCI Express

Host Bus Adapter


Hewlett Packard Emulex LPe1105-HP 4Gb FC HBA for HP c-Class Bladesystem (403621-B21)

Firmware Driver Revision Topology


In-box driver7 2.82a410, 2.82x4 FC-SW

Storage Array

External Boot

SNIA HBA API Supported

BIOS
3.11a510

VMware ESXi Hewlett Packard 5.1(vSphere Proliant :BL280c 2,BL2x220 5.1) 3, 4, 5, 6 G6 G62,BL460c G62,BL460c G7,BL460c G8,BL465c G52,BL465c G6,BL465c G7,BL490c G62,BL490c G7,BL620c G7,BL680c G5,BL680c G7,BL685c G62,BL685c G7 VMware ESXi Hewlett Packard 5.1(vSphere Proliant :BL280c 5.1) 3, 4, 5, 6 G62,BL2x220 G62,BL460c 2,BL460c G7,BL460c G6 G8,BL465c G52,BL465c G6,BL465c G7,BL490c G62,BL490c G7,BL620c G7,BL680c G5,BL680c G7,BL685c G62,BL685c G7 VMware ESXi Hewlett Packard Proliant :BL280c 5.1(vSphere 2,BL460c 5.1) 3, 4, 5, 6 G6 G62,BL460c G7,BL460c G8,BL465c G52,BL465c G6,BL465c G7,BL490c G62,BL490c G7,BL620c G7,BL680c G5,BL680c G7,BL685c G62,BL685c G7

Y8, 9 EMC Symmetrix 1 VMAX 20K

PCI Express

Hewlett Packard QLogic QMH2462 4Gb FC HBA for HP c-Class Bladesystem (403619-B21)

In-box driver7

FC-SW

Y8, 9 EMC Symmetrix VMAX 20K 1

3.13

PCI Express

Hewlett Packard QLogic QMH2562 8Gb FC HBA for HP c-Class Bladesystem (451871-B21)

In-box driver7

FC-SW

Y8, 9 EMC Symmetrix 1 VMAX 20K

Yes

3.13

Requires EMC Enginuity Operating Environment release 5875.198.148 with Pro-Active Enginuity Pack. Please see the HP website for supported BL c-Class model numbers. The supported model numbers are available in the quickspecs documents available for each BL c-Class blade. 3. EMC supports VMware's recommendations for number of nodes in native ESX cluster configurations. 4. For a list of supported guest operating systems, refer to http://www.vmware.com/resources/compatibility/search.php?deviceCategory=software. For additional information refer to the Guest Operating System Installation Guide at http://www.vmware.com/pdf/GuestOS_guide.pdf 5. For information on VMware VAAI (vStorage APIs for Array Integration) or VASA (vStorage APIs for Storage Awareness) support with EMC storage please refer to the vSphere 5 Simple Support Matrix on elabnavigator.emc.com. 6. All released updates to ESX version are supported unless otherwise specified by a footnote. Please contact server vendor for the update level required for a specific server. 7. This driver is included in the VMware ESX kernel. Driver upgrades not supported. 8. VMWare ESX Server itself is supported booting from external array. 9. Virtual machines running on VMWare ESX are supported booting from the array. 10. Available in the EMC-approved section of the Emulex website. http://www.emulex.com

1. 2.

Return To Top

Base Connectivity

EMC Virtual Provisioning Support


This table describes the supported host environment for Virtual Provisioning. Symmetrix supports the use of virtual provisioned devices as the boot/root/swap/dump volumes. Please refer to the Symmetrix Virtual Provision section of each platform's Host Connectivity Guide in the EMC Powerlink. Symmetrix Virtual Provisioning devices are supported with cluster configurations identified as supported in the EMC Support Matrix unless otherwise noted for the specific cluster software product implemented. Please read Virtual provisioning considerations for cluster configurations in the relevant EMC Host Connectivity Guide. Please refer to each respective OS host connectivity guide for the notes on the limitations of implementing Virtual Provisioning. Please refer to the corresponding EMC Symmetrix Enginuity Release Notes and The EMC Virtual Provisioning and Best Practices Technical Note. All of the components in this table are supported with condition that these components are managed not to hit the out-of-pool (space) events, please refer to Virtual provisioning section in each respective OS host connectivity guide on the practices to prevent the out-of-pool (space) condition. EMC Virtual Provisioning and the industry's thin provisioning are new technologies. Relevant industry specifications have not yet been drafted. Virtual Provisioning, like thin provisioning, has the potential to introduce events into the environment which would not otherwise occur. The unavailability of relevant industry standards results in deviations with the host-based handling of these events and the possibility of undesirable implications when these events occur. However, with the proper precautions these exposures can be minimized or eliminated.

No.
1

Operating System
VMware ESXi 5.1(vSphere 5.1)1, 2, 3, 4 VMware ESXi 5.1(vSphere 5.1)1, 2, 3, 4

Infrastructure Software

Storage Operating Environment

EMC Enginuity 5875:.135.916,.139.936,.198.148.e,.198.1486,.231.172.e,.231.1726,.249.188.e,.249.1886,.267.201.e,.267.2016,.269.201.e,.269.2016;EMC Enginuity 5876:.159.102,.82.57,.85.59;EMC VNX Block OE:05.31,05.32 EMC PowerPath/VE 5.7 P02;VMware NMP5

EMC Enginuity 5875:.135.916,.139.936,.198.148.e,.198.1486,.231.172.e,.231.1726,.249.188.e,.249.1886,.267.201.e,.267.2016,.269.201.e,.269.2016;EMC Enginuity 5876:.159.102,.82.57,.85.59;EMC VNX Block OE:05.31,05.32

1. EMC supports VMware's recommendations for number of nodes in native ESX cluster configurations. 2. For a list of supported guest operating systems, refer to http://www.vmware.com/resources/compatibility/search.php?deviceCategory=software. For additional information refer to the Guest Operating System Installation Guide at http://www.vmware.com/pdf/GuestOS_guide.pdf 3. For information on VMware VAAI (vStorage APIs for Array Integration) or VASA (vStorage APIs for Storage Awareness) support with EMC storage please refer to the vSphere 5 Simple Support Matrix on elabnavigator.emc.com. 4. All released updates to ESX version are supported unless otherwise specified by a footnote. Please contact server vendor for the update level required for a specific server. 5. VMware NMP or Native Multipathing is included in the VMware ESX distribution. 6. 1.Thin devices utilized for boot / root volumes must be fully allocated or if not fully allocated must not be oversubscribed. 2. Reclaimation is prohibited for thin devices utilized for boot / root volumes.

Return To Top

EMC Virtual Provisioning Support

HBA - Cable Compatibility

No.
1

Host Bus Adapter


Hewlett Packard:Emulex LPe1105-HP 4Gb FC HBA for HP c-Class Bladesystem (403621-B21),QLogic QMH2462 4Gb FC HBA for HP c-Class Bladesystem (403619-B21)

Cable
Generic Vendor MM-LC/xx 2

Number of Ports

Return To Top

HBA - Cable Compatibility

Director Bit/Flag Information


Operating Environments that only require Base Topologies are not listed. All the Operating Environments listed in the table require the Base Topologies and the additional settings specified for the Operating Environment. Optional Director Bit Settings include settings for special use or to enable optional functionality (refer to footnotes for uses). Optional Director Bits may also be used to identify tolerant values when an Operating System is used in a heterogeneous environment having the same Required Director Bits. In situations that require multiple combinations of director bits on the same port, the preferred method is to utilize Solutions Enabler where supported to apply settings on an initiator WWN basis. For FC Loop Base Topology (FC-AL), please enable the EAN (Enable Auto Negotiation) and UWN (Unique WWN) bits. For FC Switch Base Topology (FC-SW), please enable the EAN (Enable Auto Negotiation), PP (Point-to-Point) and UWN (Unique WWN) bits. No specific bit is required for iSCSI environment, with the following exception: VCM bit in DMX and ACLX bit in Symmetrix VMAX is required to enable the iSCSI initiator access login and, optionally, the iSCSI Challenge Handshake Authentication Protocol (CHAP) security configuration. For Symmetrix DMX volume masking (VCM), please enable the VCM bit. For Symmetrix VMAX volume masking, please enable ACLX bit. For all Symmetrix arrays, Cluster configurations require the same underlying Operating System director bits noted here in addition to any noted in the Cluster Software table.

No.
1

Operating System
VMware ESXi 5.1(vSphere 5.1)1, 2,
3, 4

Topology
FC-AL5;FC-SW

Required Director Bit Setting


(C) Common Serial Number6;(SC3) SCSI-3 compliance;(SPC-2) SPC-2 compliance

Optional Director Bit Setting


(ACLX) Initiator Group;(OS2007) Host SCSI Compliance 2007

Storage Operating Environment


EMC Enginuity 5876:.82.57,.85.59,.159.102

Storage Array
EMC Symmetrix VMAX 20K

1. EMC supports VMware's recommendations for number of nodes in native ESX cluster configurations. 2. For a list of supported guest operating systems, refer to http://www.vmware.com/resources/compatibility/search.php?deviceCategory=software. For additional information refer to the Guest Operating System Installation Guide at http://www.vmware.com/pdf/GuestOS_guide.pdf 3. For information on VMware VAAI (vStorage APIs for Array Integration) or VASA (vStorage APIs for Storage Awareness) support with EMC storage please refer to the vSphere 5 Simple Support Matrix on elabnavigator.emc.com. 4. All released updates to ESX version are supported unless otherwise specified by a footnote. Please contact server vendor for the update level required for a specific server. 5. Direct connect only, no hubs. 6. C Bit required for VERITAS VxVM DMP functionality.

Return To Top

Director Bit/Flag Information

Fibre Connectivity: Switch


For Symmetrix using switches: "Fanout" represents the maximum initiators (host adapters) login per array port. "Fanin" represents the number of array ports visible to a single initiator port (host adapter). In arbitrated loop environments, these numbers represent the maximum initiators per loop. In switch environments, these numbers are achieved through Zoning. For further details on ratio for fanout, see the Storage Array Properties table in ESM. For ratio for fanin for CLARiiON please see CLARiiON Open Systems Configuration Guide. "Port Sharing" allows any OS/HBA designated with a "Y" (for Yes) to safely share an array port with any other OS/HBA designated with a "Y" provided all of the following requirements are met (Note: "N" for No means this Operating System/HBA cannot share an array port with any dissimilar OS/HBA combination): 1. EMC LUN Masking Software (ESN Manager, SANManager, or AccessLogix) is required to prevent sharing LUNs between Operating Systems on a single array port. 2. Fanout/Fanin is restricted to the lowest value associated with Operating System/HBA combinations sharing the array port. 3. For Symmetrix 5x66 and later, reference the Heterogeneous FA Port Sharing section of Director Bit Information. 4. For Symmetrix 5568 and later, the director bits should be set according to each initiator's required settings by using the Heterogeneous Host Configuration feature in the EMC Solutions Enabler, SYMCLI Device masking component Version 5.x. For Switch Firmware levels and other fabric parameters, see the Switched Fabric Topology Parameters table.

No.
1

Host Bus Adapter


Hewlett Packard QLogic:QMH2462 4Gb FC HBA for HP c-Class Bladesystem (403619-B21),QMH2562 8Gb FC HBA for HP c-Class Bladesystem (451871-B21) Hewlett Packard:Emulex LPe1105-HP 4Gb FC HBA for HP c-Class Bladesystem (403621-B21),QLogic QMH2462 4Gb FC HBA for HP c-Class Bladesystem (403619-B21) Hewlett Packard:Emulex LPe1105-HP 4Gb FC HBA for HP c-Class Bladesystem (403621-B21),QLogic QMH2462 4Gb FC HBA for HP c-Class Bladesystem (403619-B21),QLogic QMH2562 8Gb FC HBA for HP c-Class Bladesystem (451871-B21)

Switch
Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Luns/HBA
256, 512

Port Sharing
Y

Storage Array
EMC Symmetrix VMAX 20K EMC Symmetrix VMAX 20K EMC Symmetrix VMAX 20K

512

256

Return To Top

Fibre Connectivity: Switch

Switched Fabric Topology Parameters


For SAN Router support with the switches listed on this table, please see the SAN Router tables on E-Lab Interoperability Navigator. For single fabrics consisting of McData and Brocade directors/switches & for single fabrics consisting of Cisco MDS, and Brocade, and McDATA directors/switches, see the EMC Networked Storage Topology Guide on E-Lab Interoperability Navigator (http://elabnavigator.emc.com). All switches listed under the "Switch" column in this table are supported for connectivity with the adapter ports listed under the "Host Bus Adapter" column in the "Base connectivity" table. The following connectivity support restrictions apply: (1) Older 1 Gbps HBAs are not supported with the 8 Gbps capable switches (2) Older 2 Gbps HBAs are not supported with the 16 Gbps capable switches (3) HP HBAs that require FC-AL assist, will not be supported by the newer Brocade switches. It is also important to note that the embedded blade server switch modules in this table can only be supported for connectivity with the respective mezzanine cards supported for that blade server vendor. Please check with the blade server vendor for specific connectivity restrictions that may apply to these configurations. Switch Interoperability describes the limitations for a mixed Fibre Channel switched fabric topology. The fabric topology envelope limitations and the switch management applications are defined in the attribute columns. The following is an example showing how to utilize the table: 1. The objective of this example are to construct a SAN consisting of three switches, A, B, and C, and to verify firmware and management application compatibility. In this case you need to verify that each of these components interoperates with the other two. 2. As the data shows, devices listed in the Switch column can be the same SAN with devices listed in the Interoperable Switch column. For example: switch A running firmware X can interoperate with switch B running firmware Y, and switch A running firmware X can interoperate with switch C running firmware Z. 3. In order to verify that switch B running firmware Y can also interoperate with switch C running firmware Z, find switch B in the Switch column and switch C in the Interoperable Switch column. These two steps verify that all three firmware levels are compatible with each other. For McData switches and directors, Single Mode optics support 10km, 20km, and 35km distance switch to switch. For Brocade switches and directors, Single Mode optics support 10km, 20km, and 35km distance switch to switch. OEM equivalents of switches listed in this table are supported by EMC as long as the installed firmware and all other configuration parameters are in accordance with switches specified in this table. Unless otherwise noted, all ESM supported switch ports can accommodate a mix of EMC supplied/supported Long Wave and Short Wave optics, provided that each link segment utilizes the same optical technology. All currently supported EMC storage array interfaces and corresponding storage array microcode levels are supported with listed switch firmware revisions unless otherwise noted. Target Releases are the recommended firmware versions to deploy because they are b uilt upon mature and stable code streams with the most recently rsolved defects. Target Releases are denoted by "(T)" immediately following the firmware revision. At this time, Target Releases are only specified for Connectrix B-series directors and switches.

No.
1

Switch
Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Switch Firmware Revision


3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2

Switch Management Application Revision


Cisco Fabric Manager

Interoperable Switch
Cisco MDS 9124

Interoperable Switch Firmware

Max # Domains Max # per Fabric Hops


5

ISL Domain to Domain


8

Comments Topology
See4 FC-SW

3.3.1c, 3.3.2, 3.3.3, 56 3.3.4, 3.3.4a, 3.3.5b, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.1b, 4.2.3, 4.2.7e, 5.0.1a, 5.0.4b, 5.0.4d, 5.0.7, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 56 3.3.4, 3.3.4a, 3.3.5b, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.1b, 4.2.3, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.0.7, 5.2.1, 5.2.2 9.07.02, 9.08.01, 9.09.02, 9.09.03, 9.09.05, 9.09.06, 9.09.07, 9.09.08 31

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager

Cisco MDS 9222i

See4

FC-SW

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager

EMC Connectrix DS-24M2 , EMC Connectrix DS-32M2 , EMC Connectrix DS-4400M , EMC Connectrix DS-4700M , EMC Connectrix ED-140M 9, EMC Connectrix ED-64M 10

FC-SW

Switched Fabric Topology Parameters

Switched Fabric Topology Parameters


4 Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 3.3.4, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.3, 4.2.7b, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.2.2 Cisco Fabric Manager EMC Connectrix ED-10000M , McDATA Intrepid 10000 8 9.09.03, 9.09.05, 9.09.06, 9.09.08 31 3 8 FC-SW6, 7

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager

McDATA ED-606410, McDATA ED-61409, McDATA ES-3232, McDATA ES-43005, McDATA ES-4400, McDATA ES-4500, McDATA ES-4700 Cisco MDS 9120

9.07.02, 9.08.01, 9.09.02, 9.09.03, 9.09.05, 9.09.06, 9.09.07, 9.09.08, 9.09.09

31

FC-SW

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

3.3.1c, 3.3.2, 3.3.3

244

FC-SW

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9134

3.3.1c, 3.3.2, 3.3.3, 56 3.3.4, 3.3.4a, 3.3.5b, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.1b, 4.2.3, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.0.7, 5.2.2 3.3.1c, 3.3.2 24

See4

FC-SW

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9140

See4

FC-SW

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9148

5.0.1a, 5.0.4, 5.0.4b, 56 5.0.4d, 5.2.1, 5.2.2

See4

FC-SW

10

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9216

3.3.1c, 3.3.2, 3.3.3, 3.3.4, 3.3.4a

24

See4

FC-SW

11

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9216A

3.3.1c, 3.3.2, 3.3.3, 3.3.4, 3.3.4a

244

FC-SW

12

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9216i

3.3.1c, 3.3.2, 3.3.3, 244 3.3.4, 3.3.4a, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.1b, 4.2.3

FC-SW

13

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9506, Cisco MDS 9509

3.3.1c, 3.3.2, 3.3.3, 56 3.3.4, 3.3.4a, 3.3.5b, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.1b, 4.2.3, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.0.7, 5.2.1, 5.2.2 3.3.1c, 3.3.2, 3.3.3, 56 3.3.4, 3.3.4a, 3.3.5b, 4.1.1b, 4.1.3a, 4.2.1a, 4.2.1b, 4.2.3, 4.2.7e, 5.0.1a, 5.0.4, 5.0.4b, 5.0.4d, 5.0.7, 5.2.1, 5.2.2

See4

FC-SW

14

Hewlett Packard Cisco MDS9124e SAN Switch for HP c-Class BladeSystem

Cisco Fabric Manager 1, 2, 3

Cisco MDS 9513

FC-SW

Switched Fabric Topology Parameters

10

Switched Fabric Topology Parameters


All MDS 4/12/24/48 port line cards and Sup2's must be run at SAN OS 3.0 or higher, regardless of the switch frame (9216A, 9216i, 9506, 9509, 9513). MDS 18/4 card requires 3.2.1a or higher. Switch Management Application Revision: Switch Management versions are backward compatible. You may manage any lower-level switch with a higher level Management Application Revision. 3. Switch Management Application Revision: The use of mixed code revisions in the same fabric should be limited to code upgrade processes only. You must always use the management code rev. associated with the highest level of switch code. 4. 20 VSANs maximum . 56 domains maximum per VSAN. 1024 maximum domains in a switch across all VSANs. 5. The minimum firmware revision for the ES-4300 is 5.05.00 (This firmware revision is only to be used on ES-4300) 6. McDATA's "Open Fabric Mode 1.0" and Cisco's "Interopmode 1" are fully supported for McDATA and Cisco interop. Configurations consisting of McDATA's "McDATA Fabric Mode 1.0" and Cisco's "Interopmode 4" are supported when using EOS 9.7.2 and SAN-OS v3.3.1c, and also when using EOS 9.8.1 and NX-OS 4.1.1b. For support of interopmode with previous versions, submit an RPQ. For a complete list of known issues with McDATA and Cisco interop, see Primus case EMC147089. 7. No Application Blades are currently supported in the 48000 when running in a heterogeneous switch interoperability environment. 8. This switch is the same as the IBM SAN256M 9. Do not attempt to manage an ED-140M containing XPM (10 Gb/sec) or QPM (4 Gb/sec) port modules with Connectrix Manager 7.x. The XPM requires Connectrix Manager 8.7.1 and the QPM requires Connectrix Manager 8.9.0. 10. Do not attempt to manage an ED-64M containing XPM (10 Gb/sec) port modules with Connectrix Manager 7.x. The XPM requires Connectrix Manager 8.7.1 at a minimum. 2. 1.

Return To Top

Switched Fabric Topology Parameters

11

WAN Optimization Controllers


Most devices in the IP cloud that follow IEEE standards may be used for LAN/MAN/WAN connectivity within EMC replication networks. WAN optimization controllers, however, must be explicitly listed as supported in the EMC Support Matrix table below if used in EMC replication networks. WAN optimization controllers are devices that employ protocol optimization techniques with the TCP/IP protocol to improve latency and performance over IP networks. All EMC replication solutions over IP networks must also be validated using the Solution Qualifier (SQ) process and must undergo a Network Assessment. EDL/CDL, Centera Replicator and Open Replicator are not currently supported by EMC for use with WAN optimization controllers. For these and any other applications not referenced by this EMC Support Matrix table, support is the responsibility of the customer and WAN optimization controller vendor. All EMC replication solutions over IP networks must also be validated using the Solution Qualifier (SQ) process.

No.
1

WAN Optimization Controller


Riverbed Steelhead:0501-H,0501-M,5050,5050-H,5050-L,5050-M,550

Supported Link Speed Code (Gbps) Level Topology Network


1 7.0.3a Network:IP IP

Distance Protocol
SRDF Adaptive Copy;SRDF/A;SRDFe/A;SRDFe/Adaptive Copy SRDF Adaptive Copy;SRDF/A;SRDFe/A;SRDFe/Adaptive Copy RecoverPoint

Storage Array
EMC Symmetrix VMAX 20K EMC Symmetrix VMAX 20K EMC Symmetrix VMAX 20K EMC Symmetrix VMAX 20K EMC Symmetrix VMAX 20K

Riverbed Steelhead:0601,6050,701,7050

1;10

7.0.3a

Network:IP

IP

Riverbed Steelhead:701,7050

1;10

7.0.3a

Network:IP

IP

Riverbed:CX1555H,CX1555L,CX1555M,CX755H,CX755L,CX755M

6.5.4

Network:IP

IP

RecoverPoint

Riverbed:CX1555H,CX1555L,CX1555M,CX755H,CX755L,CX755M,Steelhead 1 0501-H,Steelhead 0501-M,Steelhead 0601,Steelhead 1050,Steelhead 2050,Steelhead 5050,Steelhead 5050-H,Steelhead 5050-L,Steelhead 5050-M,Steelhead 550,Steelhead 6050,Virtual Steelhead V1050-H,Virtual Steelhead V1050-L,Virtual Steelhead V1050-M,Virtual Steelhead V2050-H,Virtual Steelhead V2050-L,Virtual Steelhead V2050-M

7.0.3a

Network:IP

IP

RecoverPoint

Return To Top

WAN Optimization Controllers

12

Path Management Software


This table describes the support statements for path management software such as EMC PowerPath and native offerings built into the Operating System in EMC Symmetrix and CLARiiON environments. Veritas DMP support statements are found on the Veritas Volume Management table. For HBA and driver support, see the Base Connectivity table. For PowerPath support with non-EMC storage, see the Path Management Software - Third Party Arrays table. When referencing PowerPath information from the ELN posted PDFs, PowerPath and PowerPath/VE information may be listed in separate documents that are available in the same download section. The Limited Functionality Level differentiates certain older or reduced functionality configurations for Path Management Software that are still supported. All configurations listed are at EMC Recommended levels except where identified as Allowed which indicates that some functionality or future compatibility may be reduced. Refer to the product release notes of newer releases for details of what functionality may not be available for older Allowed configurations. Configurations found here or in the ESM are supported and no RPQ is required.

No.
1

Storage Array

Operating System

Multipath Software (Infrastructure Software)

Storage Operating Environment


EMC Enginuity 5876 .82.57, EMC Enginuity 5876 .85.59, EMC Enginuity 5876 .159.102

EMC Symmetrix VMAX VMware ESXi 5.1(vSphere 5.1)1, EMC PowerPath/VE:5.7 P02,5.8; 2, 3, 4 20K VMware NMP5; Veritas VxDMP 6.0.1 EMC Symmetrix VMAX VMware ESXi 5.1(vSphere 5.1)1, Veritas VxDMP 6.0.1 2, 3, 4 20K

EMC Enginuity 5876 .159.102

1. EMC supports VMware's recommendations for number of nodes in native ESX cluster configurations. 2. For a list of supported guest operating systems, refer to http://www.vmware.com/resources/compatibility/search.php?deviceCategory=software. For additional information refer to the Guest Operating System Installation Guide at http://www.vmware.com/pdf/GuestOS_guide.pdf 3. For information on VMware VAAI (vStorage APIs for Array Integration) or VASA (vStorage APIs for Storage Awareness) support with EMC storage please refer to the vSphere 5 Simple Support Matrix on elabnavigator.emc.com. 4. All released updates to ESX version are supported unless otherwise specified by a footnote. Please contact server vendor for the update level required for a specific server. 5. VMware NMP or Native Multipathing is included in the VMware ESX distribution.

Return To Top

Path Management Software

13

AppSync Virtualization Server Solutions


This table describes the support for AppSync Virtualization Server Solutions. -- EMC VSI Management provides an AppSync management plug in. For supported virtual operating environments refer to the EMC VSI for VMware vSphere AppSync Management Product Guide. -- When using the VSI Plug in, for AppSync 1.0, AppSync_server_1.0.0.1_B1_092512.zip must be installed on the AppSync server. This can be found on EMC Powerlink / EMC Support Zone.

No.
1

Information Management Software


EMC AppSync 1.0

Virtualization Solution
VMware ESX Server

Operating System
VMware ESXi 5.1(vSphere 5.1)1, 2,
3, 4

Hosted OS
Microsoft Windows 2008 R2 (x64):DataCenter,Enterprise Edition,Standard Edition;Microsoft Windows 2008 R2 SP1 (x64):DataCenter,Enterprise Edition,Standard Edition;Microsoft Windows [x64] 2008:DataCenter Edition SP2,Enterprise Edition SP2,Standard Edition SP2

1. EMC supports VMware's recommendations for number of nodes in native ESX cluster configurations. 2. For a list of supported guest operating systems, refer to http://www.vmware.com/resources/compatibility/search.php?deviceCategory=software. For additional information refer to the Guest Operating System Installation Guide at http://www.vmware.com/pdf/GuestOS_guide.pdf 3. For information on VMware VAAI (vStorage APIs for Array Integration) or VASA (vStorage APIs for Storage Awareness) support with EMC storage please refer to the vSphere 5 Simple Support Matrix on elabnavigator.emc.com. 4. All released updates to ESX version are supported unless otherwise specified by a footnote. Please contact server vendor for the update level required for a specific server.

Return To Top

AppSync Virtualization Server Solutions

14

Replication Manager Block Storage


For Replication Manager (RM) with VNX and VNXe Storage: -RM supports VNXe iSCSI (Windows only). -RM supports Microsoft iSCSI initiator 2.06, 2.07 and 2.08 on Windows 2003. For Replication Manager (RM) with CLARiiON/VNX Storage: - RM supports iSCSI. - RM supports Microsoft iSCSI initiator 2.06, 2.07 and 2.08 on Windows 2003. For Replication Manager (RM) with Symmetrix Storage: -Windows 2003: SPC2 bit is required on production and mount hosts. (Refer to EMC knowledge base solutions emc177815, emc117300 and emc134969 for more details). A microcode hotfix may be needed in some cases, refer to emc113107. -Windows 2008, the following director bits should be set on the FA: Common Serial Number (C) Enable Auto Negotiation (EAN) Enable Pointtopoint (PP) Host SCSI Compliance 2007 (OS2007) SCSI3 SPC2 Compliance (SPC2) Unique World Wide Name (UWN) SCSI3 compliance (SC3) -RM can coexist with SRDF/CE. For details on how to use RM with SRDF/CE, refer to the "EMC Replication Manager and SRDF/CE" whitepaper on Powerlink -RM supports creation of crash consistent pointintime remote clones when the SAP customer is using SRDF/A. Details are contained in EMC Internal -Technical Note "Create Remote SAP TimeFinder/Clones in an SRDF/A Environment with Replication Manager Technical Note (P/N 300008993 Rev A01 - RM supports iSCSI. - RM supports Microsoft iSCSI initiator 2.06, 2.07 and 2.08 on Windows 2003.

No.
1

Information Management Software


EMC Replication Manager:5.4,5.4.1,5.4.2

Storage Array
EMC Symmetrix VMAX 20K

Storage Operating Environment


EMC Enginuity 5876 .82.57

Return To Top

Replication Manager Block Storage

15

Replication Manager Virtualization Server Solutions


This table shows the support of Replication Manager (RM) Server Virtualization Solutions: VMware Solutions: -- Supported on Symmetrix, CLARiiON, and Celerra arrays only. -- Supported with RecoverPoint. -- Replication Manager requires that Windows virtual machines are configured with the LSI Logic parallel SCSI controller. -- Replication Manager requires minimum FLARE version R22 with VMware. -- Replication Manager supports MSCS on version 3.5.1 and above, and only on RDM devices. Hyper-V Solutions: Physical disk support -- iSCSI supported on Celerra and Clariion arrays. Pass-through disk support (requires Windows 2008 R2) -- iSCSI supported on Celerra and Clariion arrays. -- FC supported on Clariion and Symmetrix. -- Supported with RecoverPoint. Virtual IO Solutions: -- LPARs with VIO configurations are supported for Symmetrix storage -- LPARs with physical HBAs and NPIV configurations are supported for Symmetrix and CLARiiON/VNX storage

No.
1

Information Management Software


EMC Replication Manager:5.4.1,5.4.2

Virtualization Solution
VMware ESX Server

Operating System

Hosted OS

Microsoft Windows 2008 R2 (x64):DataCenter,Enterprise Edition,Standard VMware ESXi 5.1(vSphere 5.1)1, 2, Edition;Microsoft Windows 2008 R2 SP1 (x64):DataCenter,Enterprise Edition,Standard 3, 4, 5 Edition;Microsoft Windows [x64] 2003 DataCenter:Edition,Edition R2,Edition R2 SP2,Edition SP2;Microsoft Windows [x64] 2003 Enterprise:Edition,Edition R2,Edition R2 SP2,Edition SP2;Microsoft Windows [x64] 2003 Standard:Edition,Edition R2,Edition R2 SP2,Edition SP2;Microsoft Windows [x64] 2008 DataCenter:Edition SP26,Edition6;Microsoft Windows [x64] 2008 Enterprise:Edition SP26,Edition6;Microsoft Windows [x64] 2008 Standard:Edition SP26,Edition6;Microsoft Windows [x86] 2003 DataCenter:Edition,Edition SP1,Edition SP2,Editon R2,Editon R2 SP2;Microsoft Windows [x86] 2003:Enterprise Edition,R2 Enterprise Edition,R2 SP2 Enterprise Edition,R2 SP2 Standard Edition,R2 Standard Edition,SP1 Enterprise Edition,SP1 Standard Edition,SP2 Enterprise Edition,SP2 Standard Edition,Standard Edition;Microsoft Windows [x86] 2008 DataCenter:Edition SP26,Edition6;Microsoft Windows [x86] 2008 Enterprise:Edition SP26,Edition6;Microsoft Windows [x86] 2008 Standard:Edition SP26,Edition6;Oracle Enterprise Linux [32-Bit] 4:U4,U5,U6,U7,U8;Oracle Enterprise Linux [32-Bit]:5,5 U1,5 U2,5 U3,5 U4,5 U5,5 U6;Oracle Enterprise Linux [64-Bit] 4:U4,U5,U6,U7,U8;Oracle Enterprise Linux [64-Bit]:5,5 U1,5 U2,5 U3,5 U4,5 U5,5 U6;Red Hat RHEL [32-Bit] 4.0 AS/ES:U3,U4;Red Hat RHEL [32-Bit]:4.5 AS/ES,4.6 AS/ES,4.7 AS/ES,4.8 AS/ES,5.0 AS/ES,5.1 AS/ES,5.2 AS/ES,5.3 AS/ES,5.4 AS/ES,5.5 AS/ES,5.6 AS/ES;Red Hat RHEL [64-Bit] 4.0 AS/ES:U3,U4;Red Hat RHEL [64-Bit]:4.5 AS/ES,4.6 AS/ES,4.7 AS/ES,4.8 AS/ES,5.0 AS/ES,5.1 AS/ES,5.2 AS/ES,5.3 AS/ES,5.4 AS/ES,5.5 AS/ES,5.6 AS/ES;SUSE SLES [32-bit] 10:SP1,SP2;SUSE SLES [32-bit]:11,11 SP1;SUSE SLES [64-bit] 10:SP1,SP2;SUSE SLES [64-bit]:11,11 SP1

1. EMC supports VMware's recommendations for number of nodes in native ESX cluster configurations. 2. For a list of supported guest operating systems, refer to http://www.vmware.com/resources/compatibility/search.php?deviceCategory=software. For additional information refer to the Guest Operating System Installation Guide at http://www.vmware.com/pdf/GuestOS_guide.pdf 3. For information on VMware VAAI (vStorage APIs for Array Integration) or VASA (vStorage APIs for Storage Awareness) support with EMC storage please refer to the vSphere 5 Simple Support Matrix on elabnavigator.emc.com. 4. All released updates to ESX version are supported unless otherwise specified by a footnote. Please contact server vendor for the update level required for a specific server. 5. PowerPath/VE 5.7 and native multi-pathing is supported. 6. Requires minimum RM 5.2.1 for virtual machine support on VMware.

Return To Top

Replication Manager Virtualization Server Solutions

16

NAS Gateway Connectivity


This table describes the support statements for NAS Connectivity. The storage array field for Integrated NAS platforms is intentionally blank. Please consult the corresponding footnote for details about the storage array.

No.
1 2

Storage Array
EMC Celerra:VG2,VG8 EMC Celerra:VG2,VG8

Interoperable Storage Array


EMC Symmetrix VMAX 20K EMC Symmetrix VMAX 20K

Storage Operating Environment


EMC VNX File OE:7.0,7.1 EMC VNX File OE:7.0,7.1

Interoperable Storage Operating Environment


Enginuity 5876:.82.57,.85.59 Enginuity 5876:.82.57,.85.59

Comments
See1, 2 See1, 2, 3

1. VNX File OE 7.0.53 or later OR VNX File OE 7.1 is required for Enginuity 5876 on VMAXe/VMAX 10K, VMAX-1/VMAX 20K and VMAX 40K systems. VNX File OE 7.0.53 or later, and VNX File OE 7.1 remain compatible with Enginuity 5876 on these systems so long as new features are not used. Refer to the respective VNX OE release notes for details. VNX File OS 7.1.55 or later is required when attaching to VMAX systems running 5876.159 or later. 2. Symm VMAX-1 * All shipping drives are supported. Support for the 200GB Flash drive, part number 005049076 requires Celerra version 6.0.40.x or later. * RAID 1 (mirrored) * Parity RAID 3+1 or 7+1 * RAID 5 * RAID 6 * Restrictions: 1: Control Volumes must be created on mirrored devices, not on parity RAID, RAID 5 or RAID 6. 2: Control Volumes are not allowed on SATA devices. 3. Support for Symmetrix Business class systems requires Enginuity release 5875.198.148 or later and VNX File OE 7.0.35.x or later. There is no support in Celerra 5.6 or 6.0 for Symmetrix Business Class systems or Enginuity release 5875.198.148 or later on VMAX systems.

Return To Top

NAS Gateway Connectivity

17

Storage Array Properties

No.
1 2

Storage Array

Storage Operating Environment

Topology
FC-AL;FC-SW;FCoE;iSCSI FC-SW

FC-Fanout
5122 5122

FC-Luns/Storage Port
20483 20483

Comments

EMC Symmetrix VMAX 20K1, 3, 4, 5, EMC Enginuity 6 5876:.82.57,.85.59,.159.102


6

EMC Symmetrix VMAX 20K1, 3, 4, 5, EMC Enginuity 5876:.82.57,.85.59,.159.102

See1

1. For Symmetrix array: The no of FC-LUNs is equal to the no of the devices mapped to the storage port including the members of the meta devices mapped to the Symm port. 2. The posted limit for fan out is based on the architectural supported limit. If there are more than 128 active physical initiators involved, an RPQ is required. 3. The posted limit for max no supported LUNs per Symmetrix FC port does not imply the max supported LUNs per HBA. Please review the host base table for the information regarding to max supported LUNs per HBA. 4. The architectural support for iSCSI-Fanout per Symmetrix iSCSI port is 1024. If there are more than 128 active physical iSCSI initiator, an RPQ is required. 5. The posted limit for fan out is based on the architectural supported limit. If there are more than 128 active physical initiators involved to a single fan-out port, please submit RPQ. 6. The architectural support for FCoE-Fanout per Symmetrix FCoE port is 1024. If there are more than 32 active physical FC/FCoE initiator, an RPQ is required.

Return To Top

December 2012 ESM

Storage Array Properties

18

Вам также может понравиться