Вы находитесь на странице: 1из 68

DATA CENTER LAN MIGRATION GUIDE

Data Center LAN Migration Guide

Table of Contents
Chapter 1: Why Migrate to Juniper . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Introduction to the Migration Guide . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Audience . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Data Center Architecture and Guide Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Why Migrate? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 Scaling Is Too Complex with Current Data Center Architectures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .10 The Case for a High Performing, Simplified Architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 Why Juniper? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 Other Considerations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 Chapter 2: Pre-Migration Information Requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 Pre-Migration Information Requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .16 Technical Knowledge and Education . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .16 Chapter 3: Data Center Migration -Trigger Events and Deployment Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 How Migrations Begin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 Trigger Events for Change and Their Associated Insertion Points . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 Considerations for Introducing an Alternative Network Infrastructure Provider . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 Trigger Events, Insertion Points, and Design Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 IOS to Junos OS Conversion Tools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 Data Center Migration Insertion Points: Best Practices and Installation Tasks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 New Application/Technology Refresh/Server Virtualization Trigger Events . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 Design Options and Best Practices: New Application/Technology Refresh/Server Virtualization Trigger Events . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 Network Challenge and Solutions for Virtual Servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 Network Automation and Orchestration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 Data Center Consolidation Trigger Event. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 Best Practices: Designing the Upgraded Aggregation/Core Layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 Best Practices: Upgraded Security Services in the Core . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .41 Aggregation/Core Insertion Point Installation Tasks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .41 Consolidating and Virtualizing Security Services in the Data Center: Installation Tasks . . . . . . . . . . . . . . . . . . . . . . . . . 44 Business Continuity and Workload Mobility Trigger Events . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 Best Practices Design for Business Continuity and HADR Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 Best Practices Design to Support Workload Mobility Within and Between Data Centers . . . . . . . . . . . . . . . . . . . . . . . . 48 Best Practices for Incorporating MPLS/VPLS in the Data Center Network Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49 Six Process Steps for Migrating to MPLS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50 Completed Migration to a Simplified, High-Performance, Two-Tier Network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 Juniper Professional Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Chapter 4: Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 Troubleshooting Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 OSI Layer 1: Physical Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57 OSI Layer 2: Data Link Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58 Virtual Chassis Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58 OSI Layer 3: Network Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 OSPF . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 VPLS Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60 Multicast . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .61 Quality of Service/Class of Service (CoS) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .61 OSI Layer 4-7: Transport to Application Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 Tools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 Troubleshooting Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 Chapter 5: Summary and Additional Resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 Additional Resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 Data Center Design Resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 Training Resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65 Juniper Networks Professional Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65 About Juniper Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Table of Figures
Figure 1: Multitier legacy data center LAN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .7 Figure 2: Simpler two-tier data center LAN design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Figure 3: Data center traffic flows . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .10 Figure 4: Collapsed network design delivers increased density, performance, and reliability . . . . . . . . . . . . . . . . . . . . . . . . . . 11 Figure 5: Junos OS - The power of one . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 Figure 6: The modular Junos OS architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .18 Figure 7: Junos OS lowers operations costs across the data center . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .19 Figure 8: Troubleshooting with Service Now . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 Figure 9: Converting IOS to Junos OS using I2J . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 Figure 10: The I2J input page for converting IOS to Junos OS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 Figure 11: Inverted U design using two physical servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 Figure 12: Inverted U design with NIC teaming . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 Figure 13: EX4200 top-of-rack access layer deployment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 Figure 14: Aggregation/core layer insertion point . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 Figure 15: SRX Series platform for security consolidation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44 Figure 16: Workload mobility alternatives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48 Figure 17: Switching across data centers using VPLS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50 Figure 18: Transitioning to a Juniper two-tier high-performance network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

Copyright 2010, Juniper Networks, Inc.

Chapter 1: Why Migrate to Juniper

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Introduction to the Migration Guide


IT has become integral to business success in virtually all industries and markets. Todays data center is the centralized repository of computing resources enabling enterprises to meet their business objectives. Todays data center traffic flows and performance requirements have changed considerably from the past with the advent of cloud computing and service-oriented architecture (SOA)-based applications. In addition, increased mobility, unified communications, compliance requirements, virtualization, the sheer number of connecting devices, and changing network security boundaries present new challenges to todays data center managers. Architecting data centers based on old traffic patterns and outdated security models is inefficient and results in lower performance, unnecessary complexity, difficulty in scaling, and higher cost. A simplified, cloud-ready, two-tier data center design is needed to meet these new challengeswithout any compromise in performance. Migrating to such a data center network can theoretically take place at any time. Practically speaking, however, most enterprises will not disrupt a production data center except for a limited time window to perform scheduled maintenance and business continuity testing. Luckily and within this context, migration to a simpler two-tier design can begin at various insertion points and proceed in controlled ways in an existing legacy data center architecture. Junipers Data Center LAN Migration Guide identifies the most common trigger events at which migration to a simplified design can take place together with design considerations at each network layer for a successful migration. The guide is segmented into two parts. For the business decision maker, Chapter 1: Why Migrate to Juniper will be most relevant. The technical decision maker will find Chapters 2 and 3 most relevant, particularly Chapter 3, which covers the data center trigger events that can stimulate a transition and the corresponding insertion points, designs, and best practices associated with pre-install, install, and post-install tasks.

Audience
While much of the high-level information presented in this document will be useful to anyone making strategic decisions about a data center LAN, this guide is targeted primarily to: Data center network and security architects evaluating the feasibility of new approaches in network design Data center network planners, engineers, and operators designing and implementing new data center networks Data center managers, IT managers, network and security managers planning and evaluating data center infrastructure and security requirements

Data Center Architecture and Guide Overview


One of the primary ways to increase data center efficiency is to simplify the infrastructure. Most data center networks in place today are based on a three-tier architecture. A simplified twotier design, made possible by the enhanced performance and more efficient packaging of todays Ethernet switches, reduces cost and complexity, and increases efficiency without compromising performance. During the 1990s, Ethernet switches became the basic building block of enterprise campus network design. Networks were typically built in a three-tier hierarchical tree structure to compensate for switch performance limitations. Each tier performed a different function and exhibited different form factors, port densities, and throughputs to handle the workload. The same topology was deployed when Ethernet moved into the data center displacing Systems Network Architecture (SNA), DECnet, and token ring designs.

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

3-TIER LEGACY NETWORK


Data Center Interconnect WAN Edge

Ethernet

Core

Aggregation Layer

Access Layer

Servers

NAS

FC Storage

FC SAN

Figure 1: Multitier legacy data center LAN


This multitiered architecture, shown in Figure 1, worked well in a client/server world where the traffic was primarily north and south, and oversubscription ratios at tiers of the network closest to the endpoints (including servers and storage) could be high. However, traffic flows and performance requirements have changed considerably with the advent of applications based on SOA, increased mobility, Web 2.0, unified communications, compliance requirements, and the sheer number of devices connecting to the corporate infrastructure. Building networks today to accommodate 5 to 10 year old traffic patterns is not optimal, and results in lower performance, unnecessary complexity, and higher cost. A new data center network design is needed to maximize IT investment and easily scale to support the new applications and services a high-performance enterprise requires to stay competitive. According to Gartner, Established LAN design practices were created for an environment of limited switch performance. Todays highcapacity switches allow new design approaches, thus reducing cost and complexity in campus and data center LANs. The three-tier concept can be discarded, because all switch ports can typically deliver rich functionality without impacting performance. 1

Neil Rikard Minimize LAN Switch Tiers to Reduce Cost and Increase Efficiency, Gartner Research ID Number: G00172149 November 17, 2009

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

WAN Edge Data Center Interconnect MX Series

Aggregation/Core EX82XX SRX5800 Access Layer EX4200


Virtual Chassis

EX4500 EX82XX

Servers

NAS

FC Storage

FC SAN

Figure 2: Simpler two-tier data center LAN design


Juniper Networks offers a next-generation data center solution, shown in Figure 2, which delivers: Simplified design for high performance and ease of management Scalable services and infrastructure to meet the needs of a high-performance enterprise Virtualized resources to increase efficiency This two-tier data center LAN architecture provides a more elastic and more efficient network that can also easily scale. This guide covers the key considerations in migrating an existing three-tier data center network to a simplified, cloudready, two-tier design. From a practical perspective, most enterprises wont initiate a complete data center redesign for an existing, operational data center. However, there are several events, such as bringing a new application or service online or a data center consolidation, which require an addition to the existing data center infrastructure. We call these common events at which migration can begin trigger events. Trigger events generate changes in design at a given network layer, which we call an insertion point. In Chapter 3 of this guide, we cover the best practices and steps involved for migration at each of the insertion points presented by a specific trigger event. By following these steps and practices, it is possible to extend migration to other legacy network tiers and continue towards a simplified two-tier Juniper infrastructure over time. In summary, this Data Center LAN Migration Guide describes: Pre-migration information requirements Migration process overview and design considerations Logical migration steps and Juniper best practices for transitioning each network layer insertion point Troubleshooting steps Additional resources

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Why Migrate?
IT continues to become more tightly integrated with business across all industries and markets. Technology is the means by which enterprises can provide better access to information in near or real time to satisfy customer needs, while simultaneously driving new efficiencies. However, todays enterprise network infrastructures face growing scalability, agility, and security challenges. This is due to factors such as increased collaboration with business partners, additional workforce mobility, and the sheer proliferation of users with smart mobile devices requiring constant access to information and services. These infrastructure challenges are seriously compounded when growth factors are combined with the trend towards data center consolidation. What is needed is a new network infrastructure that is more elastic, more efficient, and can easily scale. Scalability is a high priority, as it is safe to predict that much of the change facing businesses today is going to come as a requirement for more storage, more processing power, and more flexibility. Recent studies by companies such as IDC suggest that global enterprises will be focusing their investments and resources in the next 5 to 10 years on lowering costs while continuing to look for new growth areas. Industry analysts have identified several key data center business initiatives that align with these directions: Data center consolidation: Enterprises combine data centers as a result of merger or acquisition to reduce cost as well as centralize and consolidate resources. Virtualization: Server virtualization is used to increase utilization of CPU resources, provide flexibility, and deliver on-demand services that easily scale (currently the most prevalent virtualization example). Cloud computing: Pooling resources within a cloud provides a cost-efficient way to reconfigure, reclaim, and reuse resources to deliver responsive services. I/O convergence or consolidation: Ethernet and Fibre Channel are consolidated over a single wire on the server side. Virtual Desktop Infrastructure (VDI): Applications are run on centralized servers to reduce operational costs and also provide greater flexibility. These key initiatives all revolve around creating greater data center efficiencies. While meeting these business requirements, it is vital that efficient solutions remain flexible and scalable systems are easy to manage to maximize all aspects of potential cost savings. In todays data center, applications are constantly being introduced, updated, and retired. Demand for services is unpredictable and ever changing. Remaining responsive, and at the same time cost efficient, is a significant resource management challenge, and adding resources needs to be a last resort since it increases the cost basis for service production and delivery. Having the ability to dynamically reconfigure, reclaim, and reuse resources positions the data center to effectively address todays responsiveness and efficiency challenges. Furthermore, existing three-tier architectures are built around a client/server model that is less relevant in todays application environment. Clearly, a new data center LAN design is needed to adapt to changing network dynamics, overcome the complexity of scaling with the current multitiered architecture, as well as capitalize on the benefits of high-performance platforms and a simplified design.

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Network topologies should mirror the nature of the tra c they transport

W UP TO 70% E

Figure 3: Data center traffic flows


Applications built on SOA architecture and those delivered in the software as a service (SaaS) model require an increasing number of interactions among servers in the data center. These technologies generate a significant amount of server-to-server traffic; in fact, up to 70% of data center LAN traffic is between servers. Additional server traffic may also be produced by the increased adoption of virtualization, where shared resources such as a server pool are used at greater capacity to improve efficiency. Todays network topologies need to mirror the nature of the traffic being transported. Existing three-tier architectures were not designed to handle server-to-server traffic without going up and back through the many layers of tiers. This is inherently inefficient, adding latency at each hop, which in turn impacts performance, particularly for real-time applications like unified communications, or in industries requiring high performance such as financial trading.

Scaling Is Too Complex with Current Data Center Architectures


Simply deploying ever more servers, storage, and devices in a three-tier architecture to meet demand significantly increases network complexity and cost. In many cases, it isnt possible to add more devices due to space, power, cooling, or throughput constraints. And even when it is possible, it is often difficult and time-consuming to manage due to the size and scope of the network. Or it is inherently inefficient, as its been estimated that as much as 50% of all ports in a typical data center are used for connecting switches to each other as opposed to doing the more important task of interconnecting storage to servers and applications to users. Additionally, large Layer 2 domains using Spanning Tree Protocol (STP) are prone to failure and poor performance. Similarly, commonly deployed data center technologies like multicast dont perform at scale across tiers and devices in a consistent fashion. Legacy security services may not easily scale and are often not efficiently deployed in a data center LAN due to the difficulty of incorporating security into a legacy,multitiered design. Security blades which are bolted into switches at the aggregation layer consume excessive power and space, impact performance, and dont protect virtualized resources. Another challenge of legacy security service appliances is the limited performance scalability, which may be far below the throughput requirements of most high-performance enterprises consolidating applications or data centers. The ability to cluster together firewalls as a single logical entity to increase scalability without added management complexity is another important consideration. Proprietary systems may also limit further expansion with vendor lock-in to low performance equipment. Different operating systems at each layer may add to the complexity to operate and scale the network. This complexity is costly, limits flexibility, increases the time it takes to provision new capacity or services, and restricts the dynamic allocation of resources for services such as virtualization.

10

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

The Case for a High Performing, Simplified Architecture


Enhanced, high-performance LAN switch technology can help meet these scaling challenges. According to Network World, Over the next few years, the old switching equipment needs to be replaced with faster and more flexible switches. This time, speed needs to be coupled with lower latency, abandoning spanning tree and support for the new storage protocols. Networking in the data center must evolve to a unified switching fabric.2 New switching technology such as that found in Juniper Networks EX Series of Ethernet Switches has caught up to meet or surpass the demands of even the most high-performance enterprise. Due to specially designed applicationspecific integrated circuits (ASICs) which perform in-device switching functions, enhanced switches now offer high throughput capacity of more than one terabit per second (Tbps) with numerous GbE and 10GbE ports, vastly improving performance and reducing the number of uplink connections. Some new switches also provide built-in virtualization that reduces the number of devices that must be managed, yet can rapidly scale with growth. Providing much greater performance, enhanced switches also enable the collapsing of unnecessary network tiersmoving towards a new, simplified network design. Similarly, scalable enhanced security devices can be added to complement such a design, providing security services throughout the data center LAN. A simplified, two-tier data center LAN design can lower costs without compromising performance. Built on highperformance platforms, a collapsed design requires fewer devices, thereby reducing capital outlay and the operational costs to manage the data center LAN. Having fewer network tiers also decreases latency and increases performance, enabling wider support of additional cost savings and high bandwidth applications such as unified communications. Despite having fewer devices, a simplified design still offers high availability (HA) with key devices being deployed in redundant pairs and dual homed to upstream devices. Additional HA is offered with features like redundant switching fabrics, dual power supplies, and the other resilient capabilities available in enhanced platforms.

MULTI-TIER LEGACY NETWORK

2-TIER DESIGN

Density Performance Reliability

Figure 4: Collapsed network design delivers increased density, performance, and reliability

Robin Layland/Layland Consulting 10G Ethernet shakes Net Design to the Core/Shift from three- to two-tier architectures accelerating, Network World September 14, 2009

Copyright 2010, Juniper Networks, Inc.

11

Data Center LAN Migration Guide

Two-Tier Design Facilitates Cloud Computing


By simplifying the design, by sharing resources, and by allowing for integrated security, a two-tier design also enables the enterprise to take advantage of the benefits of cloud computing. Cloud computing delivers on-demand services to any point on the network without requiring the acquisition or provisioning of location-specific hardware and software. These cloud services are delivered via a centrally managed and consolidated infrastructure that has been virtualized. Standard data center elements such as servers, appliances, storage, and other networking devices can be arranged in resource pools that are shared securely across multiple applications, users, departments, or any other way they should be logically shared. The resources are dynamically allocated to accommodate the changing capacity requirements of different applications and improve asset utilization levels. This type of on-demand service and infrastructure simplifies management, reduces operating and ownership costs, and allows services to be provisioned with unprecedented speed. Reduced application and service delivery times mean that the enterprise is able to capitalize on opportunities as they occur.

Achieving Power Savings and Operating Efficiencies


Fewer devices require less power, which in turn reduces cooling requirements, thus adding up to substantial power savings. For example, a simplified design can offer more than a 39% power savings over a three-tier legacy architecture. Ideally, a common operating system should be used on all data center LAN devices to reduce errors, decrease training costs, ensure consistent features, and thus lower the cost of operating the network.

Consolidating Data Centers


Due to expanding services, enterprises often have more than one data center. Virtualization technologies like server migration and application load balancing require multiple data centers to be virtually consolidated into a single, logical data center. Locations need to be transparently interconnected with technologies such as virtual private LAN service (VPLS) to interoperate and appear as one. All this is possible with a new, simplified data center LAN design from Juniper Networks. However, as stated earlier, Juniper recognizes that it is impractical to flash migrate from an existing, operational, three-tier production data center LAN design to a simpler two-tier design, regardless of the substantial benefits. However, migration can begin as a result of any of the following trigger events: Addition of a new application or service Refresh cycle Server virtualization migration Data center consolidation Business continuity and workload mobility initiatives Data center core network upgrade Higher performance and scalability for security services The design considerations and steps for initiating migration from any of these trigger events is covered in detail in Chapter 3: Data Center MigrationTrigger Events and Deployment Processes.

12

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Why Juniper?
Juniper delivers high-performance networks that are open to and embrace third-party partnerships to lower total cost of ownership (TCO) as well as to create flexibility and choice. Juniper is able to provide this based on its extensive investment in software, silicon, and systems. Software: Junipers investment in software starts with Juniper Networks Junos operating system. Junos OS offers the advantage of one operating system with one release train and one modular architecture across the enterprise portfolio. This results in feature consistency and simplified management throughout all platforms in the network. Silicon: Juniper is one of the few network vendors that invests in ASICs which are optimized for Junos OS to maximize performance and resiliency. Systems: The combination of the investment in ASICs and Junos OS produces high-performance systems that simultaneously scale connectivity, capacity, and the control capability needed to deliver new applications and business processes on a single infrastructure that also reduces application and service delivery time. Juniper Networks has been delivering a steady stream of network innovations for more than a decade. Juniper brings this innovation to a simplified data center LAN solution built on four core principles: simplify, share, secure, and automate. Creating a simplified infrastructure with shared resources and secure services delivers significant advantages over other designs. It helps lower costs, increase efficiency, and keep the data center agile enough to accommodate any future business changes or technology infrastructure requirements. Simplify the architecture: Consolidating legacy siloed systems and collapsing inefficient tiers results in fewer devices, a smaller operational footprint, and simplified management from a single pane of glass. Share the resources: Segmenting the network into simple, logical, and scalable partitions with privacy, flexibility, high performance, and quality of service (QoS) enables network agility to rapidly adapt to an increasing number of users, applications, and services. Secure the data flows: Integrating scalable, virtualized security services into the network core provides benefits to all users and applications. Comprehensive protection secures data flows into, within, and between data centers. It also provides centralized management and the distributed dynamic enforcement of application and identity-aware policies. Automate network operations at each stepAn open, extensible software platform reduces operational costs and complexity, enables rapid scaling, minimizes operator errors, and increases reliability through a single network operating system. A powerful network application platform with innovative applications enables network operators to leverage Juniper or third-party applications for simplifying operations and scaling application infrastructure to improve operational efficiency. Junipers data center LAN architecture embodies these principles and enables high-performance enterprises to build next-generation, cloud-ready data centers. For information on Building the Cloud-Ready Data Center, please refer to: www.juniper.net/us/en/solutions/enterprise/data-center.

Other Considerations
It is interesting to note that even as vendors introduce new product lines, the legacy three-tier architecture remains as the reference architecture for Data Centers. This legacy three-tier architecture retains the same limitations in terms of scalability and increased complexity.

Copyright 2010, Juniper Networks, Inc.

13

Data Center LAN Migration Guide

Additionally, migrating to a new product line, even with an incumbent vendor, may require adopting a new OS, modifying configurations, and replacing hardware. The potential operational impact of introducing new hardware is a key consideration for insertion into an existing data center infrastructure, regardless of the platform provider. Prior to specific implementation at any layer of the network, it is sound practice to test interoperability and feature consistency in terms of availability and implementation. When considering an incumbent vendor with a new platform, any Enterprise organization weighing migration to a new platform from their existing one, should also evaluate moving towards a simpler high performing Juniper-based solution, which can deliver substantial incremental benefits. (See Chapter 3: Data Center MigrationTrigger Events and Deployment Processes for more details about introducing a second switching infrastructure vendor into an existing single vendor network.) In summary, migrating to a simpler data center design enables an enterprise to improve the end user experience and scale without complexity, while also driving down operational costs.

14

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Chapter 2: Pre-Migration Information Requirements

Copyright 2010, Juniper Networks, Inc.

15

Data Center LAN Migration Guide

Pre-Migration Information Requirements


Migrating towards a simplified design is based on a certain level of familiarity with the following Juniper solutions: Juniper Networks Junos operating system Juniper Networks EX Series Ethernet Switches and MX Series 3D Universal Edge Routers Juniper Networks SRX Series Services Gateways Juniper Networks Network and Security Manager, STRM Series Security Threat Response Managers, and Junos Space network management solutions Juniper Networks Cloud-Ready Data Center Reference Architecture communicates Junipers conceptual framework and architectural philosophy in creating data center and cloud computing networks robust enough to serve the range of customer environments that exist today. It can be downloaded from: www.juniper.net/us/en/solutions/enterprise/ data-center/simplify/#literature.

Technical Knowledge and Education


This Migration Guide assumes some experience with Junos OS and its rich tool set, which will not only help simplify the data center LAN migration but also ongoing network operations. A brief overview of Junos OS is provided in the following section. Juniper also offers a comprehensive series of Junos OS workshops. Standardization of networking protocols should ease the introduction of Junos OS into the data center since the basic constructs are similar. Juniper Networks offers a rich curriculum of introductory and advanced courses on all of its products and solutions. Learn more about Junipers free and fee-based online and instructor-led hands-on training offerings at: www.juniper.net/us/en/training/technical_education. Additional education may be required for migrating security services such as firewall and intrusion prevention system (IPS). If needed, Juniper Networks Professional Services can provide access to industry-leading IP experts to help with all phases of the design, planning, testing, and migration process. These experts are also available as training resources, to help with project management, risk assessment, and more. The full suite of Juniper Networks Professional Services offerings can be found at: www.juniper.net/us/en/products-services/consulting-services.

Junos OS Overview
Enterprises deploying legacy-based solutions today are most likely familiar with the number of different operating systems (OS versions) running on switching, security, and routing platforms. This can result in feature inconsistencies, software instability, time-consuming fixes and upgrades. Its not uncommon for a legacy data center to be running many different versions of a switching OS, which may increase network downtime and require greater time, effort, and cost to manage the network. From its beginning, Juniper set out to create an operating system that addressed these common problems. The result is Junos OS, which offers one consistent operating system across all of Junipers routing, switching, and security devices.

16

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

T Series Junos Space

EX8216

EX8208 Junos Pulse SRX5000 Line

NSM

NSMXpress SRX3000 Line MX Series SRX650 SRX240 M Series SRX100 SRX210 J Series LN1000 EX4500 Line EX4200 Line EX3200 Line EX2200 Line

SECURITY

ROUTERS

SWITCHES

Branch

Core

Frequent Releases

Module X

One OS

One Release Track


Figure 5: Junos OS - The power of one

One Architecture

Junos OS serves as the foundation of a highly reliable network infrastructure and has been at the core of the worlds largest service provider networks for over 10 years. Junos OS offers identical carrier-class performance and reliability to any sized enterprise data center LAN. Also through open, standards-based protocols and an API, Junos OS can be customized to optimize any enterprise-specific requirement. What sets Junos OS apart from other network operating systems is the way it is built: one operating system (OS) delivered in one software release train, and with one modular architecture. Feature consistency across platforms and one predictable release of new features ensure compatibility throughout the data center LAN. This reduces network management complexity, increases network availability, and enables faster service deployment, lowering TCO and providing greater flexibility to capitalize on new business opportunities. Junos OS consistent user experience and automated tool sets make planning and training easier and day-to-day operations more efficient, allowing for faster changes. Further, integrating new software functionality protects not just hardware investments, but also an organizations investment in internal systems, practices, and knowledge.

Junos OS Architecture
The Junos OS architecture is a modular design conceived for flexible yet stable innovation across many networking functions and platforms. The architectures modularity and well-defined interfaces streamline new development and enable complete, holistic integration of services.

API

10.0

10.1

10.2

Copyright 2010, Juniper Networks, Inc.

17

Data Center LAN Migration Guide

OPEN MANAGEMENT INTERFACES

Scripts CLI

NSM/ Junos Space

J-Web

Toolkit

CONTROL PLANE

Management

Interfaces

Module n

Routing

Service App 1

Kernel

Service App 3

DATA PLANE

Packet Forwarding Service App n

Physical Interfaces

Figure 6: The modular Junos OS architecture


The advantages of modularity reach beyond the operating system softwares stable, evolutionary design. For example, the Junos OS architectures process modules run independently in their own protected memory space, so one module cannot disrupt another. The architecture also provides separation between control and forwarding functions to support predictable high performance with powerful scalability. This separation also hardens Junos OS against distributed denial-of-service (DDoS) attacks. Junos operating systems modularity is integral to the high reliability, performance, and scalability delivered by its software design. It enables unified in-service software upgrade (ISSU), graceful Routing Engine switchover (GRES), and nonstop routing.

Automated Scripting with Junoscript Automation


With Junoscript Automation, experienced engineers can create scripts that reflect their own organizations needs and procedures. The scripts can be used to flag potential errors in basic configuration elements such as interfaces and peering. The scripts can also automate network troubleshooting and quickly detect, diagnose, and fix problems as they occur. In this way, new personnel running the scripts benefit from their predecessors long-term knowledge and expertise. Networks using Junoscript Automation can increase productivity, reduce OpEx, and increase high availability (HA), since the most common reason for a network outage is operator error. For more detailed information on Junos Script Automation, please see: www.juniper.net/us/en/community/junos.

18

Copyright 2010, Juniper Networks, Inc.

SERVICES PLANE

Services Interfaces

Service App 2

Data Center LAN Migration Guide

A key benefit of using Junos OS is lower TCO as a result of reduced operational challenges and improved operational productivity at all levels in the network.

Critical Categories of Enterprise Network Operational Costs


Switch and Router Downtime Costs Switch and Router Maintenance and Support Costs Switch and Router Deployment Time Costs Unplanned Switch and Router Events Resolution Costs Overall Switch and Router Network Operations Costs

Baseline for all network operating systems

27%* Lower with Junos

54%* Lower with Junos

25%* Lower with Junos

40%* Lower with Junos

41%* Lower with Junos

(Based on reduction in frequency and duration of unplanned network events)

(A planned events category)

(The adding infrastructure task)

(The time needed to resolve unplanned network events)

(The combined total savings associated with planned, unplanned, planning and provisioning, and adding infrastructure tasks)

Multiple network operating systems diminish e ciency3

Figure 7: Junos OS lowers operations costs across the data center


An independent commissioned study conducted by Forrester Consulting3 (www.juniper.net/us/en/reports/junos_tei.pdf) found that the use of Junos OS and Juniper platforms produced a 41% reduction in overall operations costs for network operational tasks including planning and provisioning, deployment, and planned and unplanned network events.

Juniper Platform Overview


The ability to migrate from a three-tier network design to a simpler two-tier design with increased performance, scalability, and simplicity is predicated on the availability of hardware-based services found in networking platforms such as the EX Series Ethernet Switches, MX Series 3D Universal Edge Routers, and the SRX Series Services Gateways. A consistent and unified view of the data center, campus, and branch office networks is provided by Junipers single pane of glass management platforms, including the recently introduced Junos Space. The following section provides a brief overview of the capabilities of Junipers platforms. All of the Junos OS-based platforms highlighted provide feature consistency throughout the data center LAN and lower TCO.

EX4200 Switch with Virtual Chassis Technology


Typically deployed at the access layer in a data center, Juniper Networks EX4200 Ethernet Switch provides chassisclass, high availability features, and high-performance throughput in a pay as you grow 1 rack unit (1 U) switch. Depending on the size of the data center, the EX4200 may also be deployed at the aggregation layer. Offering flexible cabling options, the EX4200 can be located at the top of a rack or end of a row. There are several different port configurations available with each EX4200 switch, providing up to 48 wire-speed, non-blocking, 10/100/1000 ports with full or partial Power over Ethernet (PoE). Despite its small size, this high-performance switch also offers multiple GbE or 10Gbe uplinks to the core, eliminating the need for an aggregation layer. And because of its small size, it takes less space, requires less power and cooling, and it costs less to deploy and maintain sparing.

The Total Economic Impact of Junos Network Operating Systems, a commissioned study conducted by Forrester Consulting on behalf of Juniper Networks, February 2009

Copyright 2010, Juniper Networks, Inc.

19

Data Center LAN Migration Guide

Up to 10 EX4200 line switches can be connected, configured, and managed as one single logical device through built-in Virtual Chassis technology. The actual number deployed in a single Virtual Chassis instance depends upon the physical layout of your data center and the nature of your traffic. Connected via a 128 Gbps backplane, a Virtual Chassis can be comprised of EX4200 switches within a rack or row, or it can use a 10GbE connection anywhere within a data center or across data centers up to 40 km apart. Junipers Virtual Chassis technology enables virtualization at the access layer, offering three key benefits: 1. It reduces the number of managed devices by a factor of 10X. 2. The network topology now closely maps to the traffic flow. Rather than sending inter-server traffic up to an aggregation layer and then back down in order to send it across the rack, its sent directly east-to-west, reducing the latency for these transactions. This also more easily facilitates workload mobility when server virtualization is deployed. 3. Since the network topology now maps to the traffic flows directly, the number of uplinks required can be reduced. The Virtual Chassis also delivers best-in-class performance. According to testing done by Network World (see full report at www.networkworld.com/slideshows/2008/071408-juniper-ex4200.html), the EX4200 offers the lowest latency of any Ethernet switch they had tested, making the EX4200 an optimal solution for high-performance, low latency, real-time applications. There has also been EX4200 performance testing done in May 2010 by Network Test which demonstrates the low latency high performance and high availability capabilities of the EX 4200 series, viewable at http://networktest.com/jnprvc. When multiple EX4200 platforms are connected in a Virtual Chassis configuration, they offer the same software high availability as traditional chassis-based platforms. Each Virtual Chassis has a master and backup Routing Engine preelected with synchronized routing tables and routing protocol states for rapid failover should a master switch fail. The EX4200 line also offers fully redundant power and cooling. To further lower TCO, Juniper includes core routing features such as OSFP and RIPv2 in the base software license, providing a no incremental cost option for deploying Layer 3 at the access layer. In every deployment, the EX4200 reduces network configuration burdens and measurably improves performance for server-to-server communications in SOA, Web services, and other distributed application designs. For more information, refer to the EX4200 Ethernet Switch data sheet for a complete list of features, benefits, and specifications at: www.juniper.net/us/en/products-services/switching/ex-series.

EX4500 10GbE Switch


The Juniper Networks EX4500 Ethernet Switch delivers a scalable, compact, high-performance platform for supporting high-density 10 gigabit per second (10 Gbps) data center top-of-rack, as well as data center, campus, and service provider aggregation deployments .The Junos OS-based EX4500 is a 48 port wire-speed switch whose ports can be provisioned as either gigabit Ethernet (GbE) or 10GbE ports in a two rack unit (2 U) form factor. The 48 ports are allocated with 40 1000BaseT ports in the base unit and 8 optional uplink module ports. The EX4500 delivers 960 Gbps throughput (full duplex) for both Layer 2 and Layer 3 protocols. For enterprises introducing 10GbE into their racks, the EX4500 can be used to add 10GbE-attached servers, iSCSI, and network-attached storage (NAS) with minimal impact to the current switching infrastructure. The EX4500 is also in Junipers roadmap to support Virtual Chassis fabric technology. For smaller data centers, the EX4500 can be deployed as the core layer switch, aggregating 10GbE uplinks from EX4200 Virtual Chassis configurations in the access layer. Back-to-front and front-to-back cooling ensure consistency with server designs for hot and cold aisle deployments. Juniper plans to add support to the EX4500 for Converged Enhanced Ethernet (CEE) and Fibre Channel over Ethernet (FCoE) in upcoming product releases. Refer to the EX4500 Ethernet Switch data sheet for more information at: www.juniper.net/us/en/products-services/ switching/ex-series/ex4500/#literature.

20

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

EX2500 10GbE Switch


The Juniper Networks EX2500 Ethernet Switch is designed for high-density 10GbE data center top-of-rack applications where high performance and low latency are key requirements. The low latency offered by the EX2500approximately 700 nanosecondsmakes it ideal for delay sensitive applications such as high-performance server clusters and financial applications, where this degree of low latency is required. Note that the EX2500 is not Junos OS-based. Refer to the EX2500 Ethernet Switch data sheet for more information at: www.juniper.net/us/en/products-services/ switching/ex-series/ex2500/#literature.

EX8200 Line of Ethernet Switches


The Juniper Networks EX8200 line of Ethernet switches is a high-performance chassis platform designed for the high throughput that a collapsed core layer requires. This highly scalable platform supports up to 160,000 media access control (MAC) addresses, 64,000 access control lists (ACLs), and wire-rate multicast replication. The EX8200 line may also be deployed as an end-of-rack switch for those enterprises requiring a dedicated modular chassis platform. The advanced architecture and capabilities of the EX8200 line, similar to the EX4200, accelerate migration towards a simplified data center design. The EX8200-40XS line card brings 10GbE to the access layer for end-of-row configurations. This line card will deliver 25 percent greater density per chassis and consume half the power of competing platforms, reducing rack space and management costs. The EX8200 line is expected to add Virtual Chassis support later in 2010 with additional features being added in early 2011. With the new 40-port line card, the EX8200 line with Virtual Chassis technology will enable a common fabric of more than 1200 10GbE ports. The most fundamental challenge that data center managers face is the challenge of physical plant limitations. In this environment, taking every step possible to minimize power draw for the required functionality becomes a critical goal. For data center operators searching for the most capable equipment in terms of functionality for the minimum in rack space, power, and cooling, the EX8200 line delivers higher performance and scalability in less rack space with lower power consumption than competing platforms. Designed for carrier-class HA, each EX8200 line model also features fully redundant power and cooling, fully redundant Routing Engines, and N+1 redundant switch fabrics. For more information, refer to the EX8200 line data sheets for a complete list of features and specifications at: www.juniper.net/us/en/products-services/switching/ex-series.

MX Series 3D Universal Edge Routers


Its important to have a consistent set of powerful edge services routers to be able to interconnect the data center to other data centers and out to dispersed users. The MX Series with the new Trio chipset delivers cost-effective, powerful scaling that allows enterprises to support application-level replication for disaster recovery or virtual machine migration between data centers by extending VLANs across data centers using mature, proven technologies such as VPLS. It is interesting to note the following observation from the recent 2010 MPLS Ethernet World Conference from the Day 3 Data Center Interconnect session: VPLS is the most mature technology today to map DCI requirements. Delivering carrier-class HA, each MX Series model features fully redundant power and cooling, fully redundant Routing Engines, and N+1 redundant switch fabrics. For more information, refer to the MX Series data sheet for a complete list of features, benefits, and specifications at: www.juniper.net/us/en/products-services/routing/mx-series.

Copyright 2010, Juniper Networks, Inc.

21

Data Center LAN Migration Guide

Consolidated Security with SRX Series Services Gateways


The SRX Series Services Gateways replace numerous legacy security solutions by providing a suite of services in one platform, including a firewall, IPS, and VPN services. Supporting the concept of zones, the SRX Series can provide granular security throughout the data center LAN. The SRX Series can be virtualized and consolidated into a single pool of security services via clustering. The SRX Series can scale up to 10 million concurrent sessions allowing the SRX Series to massively and rapidly scale to handle any throughput without additional devices, multiple cumbersome device configurations, or operating systems. The highly scalable performance capabilities of the SRX Series platform, as with the EX Series switches, lays the groundwork for a simplified data center infrastructure and enable enterprises to easily scale to meet future growth requirements. This is in contrast to legacy integrated firewall modules and standalone appliances which have limited performance scalability. Even when multiple firewall modules are used, the aggregate performance may still be far below the throughput required for consolidating applications or data centers, where firewall aggregate throughput of greater than 100 gigabits may be required. The lack of clustering capabilities in some legacy firewalls not only limits performance scalability but also increases management and network complexity. The SRX Series provides HA features such as redundant power supplies and cooling fans, as well as redundant switch fabrics. This robust platform also delivers carrier-class throughput. The SRX5600 is the industrys fastest firewall and IPS by a large margin, according to Network World. For more information, refer to the SRX Series data sheet for a complete list of features, benefits, and specifications at: www.juniper.net/us/en/products-services/security/srx-series.

Altor Networks Virtual Firewall (VF 4.0)


To address the unique security challenges of virtualized networks and data centers, Juniper has integrated Altors virtual firewall and cloud protection software into its security portfolio to give network and application visibility and granular control over virtual machines (VM). Combining a powerful stateful virtual firewall with virtual intrusion detection (IDS), VM Introspection and automated compliance assessment, Altors comprehensive solution for protecting virtualized workloads slipstreams easily into Juniper environments featuring any of the following: SRX Series Services Gateways STRM Series Security Threat Response Managers IDP Series Intrusion Detection and Prevention Appliances Altors integrations focus on preserving customers investment into Juniper security, and extending it to the virtualized infrastructure with the similar feature, functionality, and enterprise-grade requirements like high-performance, redundancy, and central management. Juniper customers can deploy Altor on the virtualized server, and integrate security policies, logs, and related work flow into existing SRX Series, STRM Series, and IDP Series infrastructure. Customers benefit from layered, granular security without the management and OpEx overhead. Altor v4.0 will export firewall logs and inter-VM traffic flow information to STRM Series to deliver single-pane of glass for threat management. Customers who have deployed Juniper Networks IDP Series, and management processes around threat detection and mitigation can extend that to the virtualized server infrastructure with no additional CapEx investment. Altors upcoming enhancements with SRX Series and Junos Space continues on the vision to deliver gapless security with a common management platform. Altor-SRX Series integration will ensure trust zone integrity is guaranteed to the last mile - particularly relevant in cloud and shared-infrastructure deployments. Altors integration with Junos Space will bridge the gap between management of physical resources and virtual resources to provide a comprehensive view of the entire data center. Refer to the Securing Virtual Server Environments with Juniper Networks and Altor Networks solutions brief for more: www.juniper.net/us/en/local/pdf/solutionbriefs/3510354-en.pdf.

22

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

MPLS/VPLS for Data Center Interconnect


The consolidation of network services increases the need for Data Center Interconnect (DCI). Resources in one data center are often accessed by one or more data centers. Different business units, for example, may share information across multiple data centers via VPNs. Or compliance regulations may require that certain application traffic be kept on separate networks throughout data centers. Or businesses may need a real-time synchronized standby system to provide optimum HA in a service outage. MPLS is a suite of protocols developed to add transport and virtualization capabilities to large data center networks. MPLS enables enterprises to scale their topologies and services. An MPLS network is managed using familiar protocols such as OSPF or Integrated IS-IS and BGP. MPLS provides complementary capabilities to standard IP routing. Moving to an MPLS network provides business benefits like improved network availability, performance, and policy enforcement. MPLS networks can be employed for a variety of reasons: Inter Data Center Transport: To connect consolidated data centers to support mission critical applications. For example, real-time mainframe replication or disk, database, or transaction mirroring. Virtualizing the Network Core: For logically separating network services. For example, providing different levels of QoS for certain applications or separate application traffic due to compliance requirements. Extending L2VPNs for Data Center Interconnect: To extend L2 domains across data centers using VPLS. For example, to support application mobility with virtualization technologies like VMware VMotion, or to provide resilient business continuity for HA by copying transaction information in real time to another set of servers in another data center. The MX Series provides high capacity MPLS and VPLS technologies. MPLS networks can also facilitate migration towards a simpler, highly scalable and flexible data center infrastructure.

Junipers Unified Management Solution


Juniper provides three powerful management solutions for the data center LAN via its NSM and STRM Series platforms, as well as Junos Space. For more information on MPLS/VPLS, please refer to the Implementing VPLS for Data Center Interconnectivity Implementation Guide at: www.juniper.net/us/en/solutions/enterprise/data-center/simplify/#literature.

Network and Security Manager


NSM offers a single pane of glass to manage and maintain Juniper platforms as the network grows. It also helps maintain and configure consistent routing and security policies across the entire network. And NSM helps delegate roles and permissions as well. Delivered as a software application or a network appliance, NSM provides many benefits: Centralized activation of routers, switches, and security devices Granular role-based access and policies Global policies and objects Monitoring and investigative tools Scalable and deployable solutions Reliability and redundancy Lower TCO

Copyright 2010, Juniper Networks, Inc.

23

Data Center LAN Migration Guide

The comprehensive NSM solution provides full life cycle management for all platforms in the data center LAN. Deployment: Provides a number of options for adding device configurations into the database, such as importing a list of devices, or discovering and importing deployed network devices, or manually adding a device and configuration in NSM, or having the device contact NSM to add its configuration to the database. Configuration: Offers central configuration to view and edit all managed devices. Provides offline editing/modeling of device configuration. Facilitates the sharing of common configurations across devices via templates and policies. Provides configuration file management for backup, versioning, configuration comparisons, and more. Monitoring: Provides centralized event log management with predefined and user-customizable reports. Provides tools for auditing log trends and finding anomalies. Provides automatic network topology creation using standardsbased discovery of Juniper and non-Juniper devices based on configured subnets. Offers inventory management for device management interface (DMI)-enabled devices, and Job Manager to view device operations performed by other team members. Maintenance: Delivers centralized Software Manager to version track software images for network devices. Other tools also transform/validate between user inputs and device-specific data formats via DMI schemas. Using open standards like SNMP and system logging, NSM has support for third-party network management solutions from IBM, Computer Associates, InfoVista, HP, EMC, and others. Refer to the Network and Security Manager data sheet for a complete list of features, benefits, and specifications: www.juniper.net/us/en/products-services/security/nsmcm.

STRM Series Security Threat Response Managers


Complementing Junipers portfolio, the STRM Series offers a single pane of glass to manage security threats. It provides threat detection, event log management, compliance, and efficient IT access to the following: Log Management: Provides long-term collection, archival, search, and reporting of event logs, flow logs, and application data. Security Information and Event Management (SIEM): Centralizes heterogeneous event monitoring, correlation, and management. Unrivaled data management greatly improves ITs ability to meet security control objectives. Network Behavior Anomaly Detection (NBAD): Discovers aberrant network activities using network and application flow data to detect new threats that others miss. Refer to the STRM Series data sheet for a complete list of features, benefits, and specifications: www.juniper.net/us/ en/products-services/security/strm-series.

Junos Space
Another of ITs challenges has been adding new services and applications to meet the ever growing demand. Historically, this has not been easy, requiring months of planning and only making changes in strict maintenance windows. Junos Space is a new, open network application platform designed for building applications that simplify network operations, automate support, and scale services. Organizations can take control of their own networks through selfwritten programs or third-party applications from the developer community. Embodied in a number of appliances across Junipers routing, switching, and security portfolio, an enterprise can seamlessly add new applications, devices, and device updates as they become available from Juniper and the developer community, without ever restarting the system for full plug and play.

24

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Several applications will be available on Junos Space throughout 2010. Junos Space applications introduced as of the first half of 2010 include: Junos Space Virtual Control (expected availability in Q3 2010) allows users to monitor, manage, and control the virtual network environments that support virtualized servers deployed in the data center. Virtual Control provides a consolidated solution for network administrators to gain end-to-end visibility into, and control over, both virtual and physical networks from a single management screen. By enabling network-wide topology, configuration, and policy management, Virtual Control minimizes errors and dramatically simplifies data center network orchestration, while at the same time lowering total cost of ownership by providing operational consistency across the entire data center network. Virtual Control also greatly improves business agility by accelerating server virtualization deployment. Juniper has also formed a new collaboration with VMware that takes advantage of its open APIs to achieve seamless orchestration across both physical and virtual network elements by leveraging Virtual Control. The combination of Junos Space Virtual Control and VMware vSphere provides automated orchestration between the physical and virtual networks, wherein a change in the virtual network is seamlessly carried over the physical network and vice versa. Junos Space Ethernet Design (available now) is a Junos Space software application that enables end-to-end campus and data center network automation. Ethernet Design provides full automation including configuration, provisioning, monitoring, and administration of large switch and router networks. Designed to enable rapid endpoint connectivity and operationalization of the data center, Ethernet Design uses a best practice configuration and scalable workflows to scale data center operations with minimal operational overhead. It is a single pane of glass platform for end-to-end network automation that improves productivity via a simplified, create one, use extensively configuration and provisioning model. Junos Space Security Design (available now) enables fast, easy, and accurate enforcement of security state across the enterprise network. Security Design enables quick conversion of business intent to device-specific configuration, and it enables auto-configuration and provisioning through workflows and best practices to reduce the cost and complexity of security operations. Service Now and Junos Space Service Insight (available now) consists of Junos Space applications that enable fast and proactive detection, diagnosis, and resolution of network issues. (See Automated Support with Service Now for more details.) Junos Space Network Activate (expected availability Q4 2010) facilitates fast and easy setup of VPLS services, and allows for full lifecycle management of MPLS services In addition, the Junos Space Software Development Kit (SDK) will be released to enable development of a wide range of third-party applications covering all aspects of network management. Junos Space is designed to be open and provides northbound, standards-based APIs for integration to third-party data center and service provider solutions. Junos Space also includes DMI based on NetConf, an IETF standard, which can enable management of DMI-compliant third-party devices. Refer to the following URL for more information on Junos Space applications: www.juniper.net/us/en/productsservices/software/junos-platform/junos-space/applications.

Copyright 2010, Juniper Networks, Inc.

25

Data Center LAN Migration Guide

Automated Support with Service Now


Built on the Junos Space platform, Service Now delivers on Junipers promise of network efficiency, agility, and simplicity by delivering service automation that leverages Junos OS embedded technology. For devices running Junos OS 9.x and later releases, Service Now aids in troubleshooting for Junipers J-Care Technical Services. Junos OS contains the scripts which provide device and incident information that is relayed to the Service Now application where it is logged, stored, and with the customers permission, forwarded to Juniper Networks Technical Services for immediate action by the Juniper Networks Technical Assistance Center (JTAC). Not only does Service Now provide automated incident management, it offers automated inventory management for all Junos OS devices running release 9.x and later. These two elements provide substantial time savings in the form of more network uptime and less time spent on administrative tasks like inventory data collection. This results in a reduction of operational expenses and streamlined operations, allowing key personnel to focus on the goals of the network rather than its maintenanceall of which enhance Junipers ability to simplify the data center.

AI Scripts Installed

JMB Hardware So ware Resources Calibration CUSTOMER NETWORK

Service Now

Juniper Support System Service Insight

CUSTOMER OR PARTNER NOC

INTERNET

Gateway

JUNIPER

Figure 8: Troubleshooting with Service Now


The Service Insight application, available in Fall 2010 on the Junos Space platform, takes service automation to the next level by delivering proactive, customized support for networks running Juniper devices. While Service Now enables automation for reactive support components such as incident and inventory management for efficient network management and maintenance, Service Insight brings a level of proactive, actionable network insight that helps manage risk, lower TCO, and improve application reliability. The first release of Service Insight will consist of the following features: Targeted product bug notification: Proactive notification to the end user of any new bug notification that could impact network performance and availability with analysis of which devices could be vulnerable to the defect. This capability can avoid network incidents due to known product issues, as well as save numerous hours of manual impact analysis for system-wide impact of a packet-switched network (PSN). EOL/EOS reports: On-demand view of the end of life (EOL), end of service (EOS), and end of engineering (EOE) status of devices and field-replaceable units (FRUs) in the network. This capability brings efficiency to network management operations and mitigates the risk of running obsolete network devices and/or software/firmware. With this capability, the task of taking network inventory and assessing the impact of EOL/EOS announcements is reduced to the touch of a button instead of a time-consuming analysis of equipment and software revision levels and compatibility matrices.

26

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Chapter 3: Data Center Migration -Trigger Events and Deployment Processes

Copyright 2010, Juniper Networks, Inc.

27

Data Center LAN Migration Guide

How Migrations Begin


Many enterprises have taken on server, application, and data center consolidations to reduce costs and to increase the return on their IT investments. To continue their streamlining efforts, many organizations are also considering the use of cloud computing in their pooled, consolidated infrastructures. While migrating to a next-generation cloud-ready data center design can theoretically take place at any time, most organizations will not disrupt a production facility except for a limited time-window to perform scheduled maintenance and continuity testing, or for a suitably compelling reason whose return is worth the investment and the work. In Chapter 3 of this guide, we identify a series of such reasonstypically stimulated by trigger eventsand the way these events turn into transitions at various insertion points in the data center network. We also cover the best practices and steps involved in migration at each of the insertion points presented by a specific trigger event. By following these steps and practices, it is possible to extend migration to legacy network tiers and move safely towards a simplified data center infrastructure.

Trigger Events for Change and Their Associated Insertion Points


Change in the data center network is typically determined by the type of event triggering the organization to make that change. What follows is a short description of trigger events which can stimulate an organization to make the investments related to these events: Provisioning a new area of infrastructure area or Point of Delivery (POD) in an existing data center due to additional capacity required for new applications and services. The new applications may also have higher performance requirements that cannot be delivered by the existing infrastructure. Technology refresh due to either EOL on a given product line or an upgrade to the latest switching and/or server technology. A refresh can also be driven by the end of an equipment depreciation cycle, company policy regarding available headroom capacity, or for adding capacity to meet planned future expansion. Infrastructure redesign due to increased use of server virtualization. Data center consolidation due to merger or acquisition, cost saving initiatives, or moving from an existing colocation facility. Due to the increased scalability, performance, and high availability requirements, data center consolidation may also require a technology refresh. Business continuity and workload mobility initiatives. Delivering HA and VM/application mobility typically involves VLAN stretching within or between data centers. Upgrade to the core data center network for higher bandwidth and capacity to support new capabilities such as server virtualization/workload mobility or higher application performance. This may also be due to a technology refresh as a result of the retirement of legacy equipment which is at end of life (EOL). Need for higher performance and scale in security. Existing security gateways, whether integrated in a chassis or running as standalone appliances, may not be able to deliver the higher performance required to support the increased traffic from data center consolidation, growth in connected devices, increased extranet collaboration, and internal/external compliance and auditing requirements. Server, desktop, and application virtualization may also drive changes in the security model, to increase the strength of security in the new environments and ease complexity in management. Enhancements can be made to the core, edge, or virtual server areas of the data center network to deal with these requirements. Addressing any or all of these trigger events results in deployment of new technology into the access, aggregation, core, or services tiers of an existing data center network.

28

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Considerations for Introducing an Alternative Network Infrastructure Provider


In some installations, a key consideration when evolving an existing infrastructure is the impact of introducing another vendor. Organizations can minimize any impact by using the same best practices they employ in a single vendor network. For example, it is sound practice to test interoperability and feature consistency before an implementation at any network layer. Many enterprises do this today, since there are often multiple inconsistent versions of an operating system within a single vendors portfolio, or even completely different operating systems within that portfolio. For example, the firewall or intrusion detection and prevention (IDP) platforms may have a different OS and interface from the switching products. Even within a switching portfolio, there may be different operating systems, each supporting different feature implementations. It is also sound practice to limit fault domains and contain risks when introducing an additional vendor. This can be accomplished with a building block design for the target insertion point, when deploying into an existing LAN. This approach allows for definition of the new insertion as a functional module, testing of the module in proof-of-concept (PoC) environments before deployment, and clean insertion of the new module into production after testing. As mentioned earlier, PoC testing is often done as a best practice in a single vendor network as well. Other steps that can ensure successful insertion of Juniper Networks technology into an existing data center LAN include: Training Multivendor automation and management tools

Training
The simplicity of Junipers implementations typically minimizes the need for extensive training to accompany deployment; however, Juniper also offers a variety of training resources to accelerate deployments. To start with, standardization of protocols within the network typically eases introduction, since basic constructs are similar and interoperability has usually been tested and proven ahead of time by Juniper. Beyond the protocols, differences in command-line interface (CLI) are usually easier to navigate than people initially think. Time after time, people familiar with other CLIs find themselves able to make the transition quickly due to the consistent, intuitive nature of Junos operating systems implementation (it is easy to learn and use). Junos OS also has a tremendous amount of flexibility and user support built into it. For example, to ease migration from Ciscos IOS, there is a Junos OS command to display a configuration file in a format similar to IOS. Additionally, hands-on training is available in the form of a two-day boot camp. Customized training can also be mapped to address any enterprises specific environment. Training not only gives an opportunity to raise the project teams skill level, but also to get experience with any potential configuration complexities prior to entering the implementation phase of a project. Junos OS also provides embedded automation capabilities. A library of scripts that automate common operations tasks is readily available online for viewing and downloading. Categorized by function, the script with the best fit can easily be found. Refer to the Junos OS Script Library for a complete list: www.juniper.net/us/en/community/junos/ script-automation/library.

Multivendor Automation and Management Tools


In a multivendor environment, it is often critical to establish a foundation of multivendor management tools that work with existing suppliers as well as with Juniper. There are well established multivendor tools available in the fault and performance analysis areas. These tools work with equipment from all of the major vendors in the market including Juniper and other vendors. In the provisioning, configuration, inventory management, and capacity planning areas, existing third-party tools typically dont scale or leverage the capabilities and best practices of each vendors equipment. In these situations, it works well to leverage an integrated platform like Junos Space to support the Juniper infrastructure consistently, and, where possible, to incorporate support for other vendors platforms and applications if the APIs and SDKs of Junos Space can be used to complete that integration. Junos Space provides APIs and SDKs to customize existing applications, and also enables partners to integrate their applications into this homogenous application platform.

Copyright 2010, Juniper Networks, Inc.

29

Data Center LAN Migration Guide

Trigger Events, Insertion Points, and Design Considerations


To summarize the preceding discussions, the table below highlights the mapping between a trigger event, the corresponding network insertion point, and the network design considerations that are pertinent in each of the data center tiers.

Table 1: Trigger Event Design Consideration Summary


TRIGGER EVENT NETWORK INSERTION LAYER(S) DESIGN CONSIDERATIONS

New application or service Technology refresh Server virtualization

New switching infrastructure for: Access Aggregation Core

Topof-rack or end-of-row deployment Cabling changes Traffic patterns among servers VLAN definitions for applications/services segmentation L2 domain for VM workload mobility Server connectivity speed: GbE/10GbE Application latency requirements Network interface card (NIC) teaming Uplinks Uplink oversubscription ratios Number and placement of uplinks GbE/10GbE uplinks Use of L2 or L3 for uplinks IEEE Spanning Tree Protocol (STP) Redundant trunk groups as STP alternative - Link aggregation sizing/protocol QoS Classification and prioritization Policing High availability (HA) requirements Multicast scalability/performance requirements Interoperability testing definitions IOS configuration for conversion to Junos OS

Data center consolidation

Aggregation/core layer Access layer Services layer

Sufficiency of existing physical capacity Interior gateway protocol (IGP) and exterior gateway protocol (EGP) design Multicast requirements L2/L3 domain boundaries IEEE STP Default gateway/root bridge mapping Virtualized services support requirements (VPN/VRF) Uplinks Density Speed Link aggregation Scaling requirements for MAC addresses and ACLs Latency requirements HA features QoS Security Policies Existing policy migration ACL migration New policy definitions Interoperability testing definitions IOS configuration for conversion to Junos OS

30

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

TRIGGER EVENT

NETWORK INSERTION LAYER(S)

DESIGN CONSIDERATIONS

Business continuity Workload mobility Data center core network upgrade

Core layer

Sufficiency of existing physical capacity VLAN definitions for applications/services segmentation VLAN/VPN routing and forwarding (VRF) mapping Latency/performance requirements Traffic engineering/QoS requirements Connecting with legacy/proprietary IGP protocols Data Center Interconnect (DCI) method Stretching VLANs Physical layer extension via dense wavelength-division multiplexing (DWDM) VPLS/MPLS Interoperability testing definitions IOS configuration for conversion to Junos OS

Higher performance security services

Aggregation/core layer Insertion

Capacity and throughput requirements Compliance and audit requirements Existing security policy rules migration Zone definitions Virtual machine security requirements for virtual firewall Interoperability testing definitions

IOS to Junos OS Conversion Tools


For organizations with existing Cisco IOS configurations, Juniper provides an IOS to Junos OS (I2J) migration tool. I2J is a web-based software translation tool that converts Cisco IOS configurations for Cisco Catalyst switches and Cisco routers into their Junos OS equivalent. I2J lowers the time and cost of migrating to Juniper router and switching platforms. The I2J tool supports configurations for physical and logical interfaces, routing protocols, routing policies, packet filtering functions, and system commands. I2J also automatically flags any conversion errors or incompatible configurations. It can be used for IOS configuration conversions at all network insertion points. Figure 9 shows a sample I2J screen. Since all network operating systems are not created equal, it should be noted that a straight IOS to Junos OS conversion may mask opportunities to improve the overall network configuration or implement features or functions not possible with IOS. Conversely, I2J also helps identify IOS-specific features which may not be implementable under Junos OS in a direct 1:1 mapping but need to be implemented in an alternate manner.

Figure 9: Converting IOS to Junos OS using I2J


I2J is only available to Juniper customers or partners with a support contract via a secure Advanced Encryption Standard (AES) 256-bit encrypted website found at: https://i2j.juniper.net.

Copyright 2010, Juniper Networks, Inc.

31

Data Center LAN Migration Guide

IOS Input Page


The main interface for I2J supports upload or cut and paste of IOS configuration files. You can adjust a variety of translation options, such as outputting verbose IOS comments or consolidating firewall terms. Then use the Translate button to convert the IOS file into Junos OS. The output is displayed with statistics, the Junos OS configuration output, and the IOS source with messages.

Figure 10: The I2J input page for converting IOS to Junos OS Data Center Migration Insertion Points: Best Practices and Installation Tasks
The legacy three-tier design found in many of todays data centers was depicted in Figure 1. This is the baseline for layer insertion points addressed in this document. For a specific insertion point such as the access layer, for example, the recommended Juniper best practices pertaining to that layer are provided first. This is then followed by the recommended preinstallation, installation, and post installation tasks. Recommended best practices and installation-related tasks focus primarily on currently shipping products and capabilities. The next revision of the Data Center LAN Migration Guide will include more implementation detail on the switching platforms and software introduced by Juniper as part of the New Network for the Data Center announcements, May 2010: www.juniper.net/us/en/solutions/enterprise/data-center. A dedicated Troubleshooting chapter detailing Juniper recommended guidelines for the most commonly encountered migration and installation issues is also included in this guide.

New Application/Technology Refresh/Server Virtualization Trigger Events


These events are often driven by a lack of capacity in the existing infrastructure to support a new application or service. They may also occur when an organization is trying to maximize its processor capacity through use of virtual servers. Redesigns based on these triggers can involve upgrades to either data center access or aggregation tiers (or both) as described later in this section. They may also involve a redesign across data centers, addressed later in the Business Continuity and Workload Mobility Trigger Events section. In general, server virtualization poses an interesting set of design challenges, detailed at the end of this section. The insertion point for each of these triggers often involves provisioning one or more new Points of Delivery (PODs) or designated sections of a data center layout, including new switches for the networks access layer. The new POD(s) may also include an upgrade of the related aggregation layer which, depending on the requirements, could potentially later serve as the core in a simplified two-tier design. The process may also involve a core switch/router upgrade or replacement to increase functionality and bandwidth for each new POD requirement. For 10GbE server connectivity in a top-of-rack deployment, Junipers recommended design would be based on EX4500 or EX2500 switches, with latency requirements largely determining the platform decision. EX2500 switches support applications with sub microsecond switching requirements, such as high-performance compute cluster and financial services transaction processing. Use of Virtual Chassis technology in the EX4500 is in Junipers roadmap for delivery in 2011.

32

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

EX4500 and EX4200 switches will be able to work in the same Virtual Chassis, allowing companies to mix GbE and 10GbE attached devices within the same fabric. When the requirement is for GbE server connectivity, the EX4200 would be the platform of choice, where up to 480 GbE server connections may be accommodated within a single Virtual Chassis.

Design Options and Best Practices: New Application/Technology Refresh/Server Virtualization Trigger Events
When deploying a new access layer as part of this trigger event, there are issues related to uplink oversubscription, STP, and Virtual Chassis. Understanding design options and best practices for each of these topics is important to a successful deployment.

Tunable Oversubscription in Uplink Link Aggregation


In legacy networks, oversubscription is an ongoing issue, leading to unpredictable performance and the inability to deploy or provision a new application with confidence. Oversubscription typically occurs between the access and the core/aggregation switches (on the access network uplinks). Juniper provides a tunable oversubscription ratio of from 1 to 12 between the access and the core with the EX4200 Virtual Chassis. Up to 10 EX4200 Virtual Chassis systems can be configured together to perform and be managed as one device. Each Virtual Chassis supports up to 48 GbE connections and up to two 10GbE uplinks. A full configuration of 10 EX4200s has up to 480 GbE ports + 2 x (m) x 10GbE uplinks, where m is 1 to 10. The oversubscription ratio is tuned by adjusting the number of units in the Virtual Chassis, the number of GbE ports, and the 10GbE uplinks. An oversubscription ratio of 1, delivering full wire rate, is achieved when there is no greater than 2 units in a Virtual Chassis configuration and 40 gigabit user ports provisioned. An oversubscription ratio of 12:1 is achieved if there are 480 GbE ports and 4x10GbE uplinks. The target oversubscription ratio is always going to be based on the applications and services that the Virtual Chassis is expected to support. It can be easily adjusted by adding or removing 10GbE uplinks, or by increasing or reducing the number of member switches in the Virtual Chassis. For example, a 5 to 6 member Virtual Chassis using two to four 10GbE uplinks delivers oversubscription ratios between 7:1 and 12:1. Or, a 7 to 8 member Virtual Chassis using 4 10GbE uplinks, two in the middle and two at the ends, delivers oversubscription ratios between 8:1 and 9:1. While these oversubscription levels are common, the most important point is that they can be adjusted as needed.

Spanning Tree Alternatives for Access Layer Insertion Point


STP offers benefits to the data center, but it is also plagued with a number of well-known and hard to overcome issues. These include inefficient link utilization, configuration errors that can potentially bring an entire L2 domain down, and other problems. To avoid these issues, enterprises may consider alternative architectures to STP when inserting a new POD into the network. These can include an inverted U design, Redundant Trunk Groups (RTGs), and L3 uplinks.

Inverted U Design
An enterprise can create an STP-free data center topology when using an L2 domain to facilitate virtual machine mobility through the way it connects servers to an access layer switch. An inverted U design can be used such that no L2 loops are created. While technologies like STP or RTG arent required to prevent the loop, best practices still recommend provisioning STP to prevent accidental looping due to incorrect configuration.

Copyright 2010, Juniper Networks, Inc.

33

Data Center LAN Migration Guide

There are two basic options to provision this design: two physically separate servers connected to two separate access layer switches, or two separate Virtual Chassis units in a Juniper deployment as depicted in Figure 11.

802.3ad LAG 802.1q Trunking

AGGREGATION / CORE

EX82XX
802.3ad LAG 802.1q Trunking 802.3ad LAG 802.1q Trunking

EX82XX

EX4200

Virtual Chassis

EX4200

Virtual Chassis

Access

Figure 11: Inverted U design using two physical servers


The bold lines show that there are no loops in this upside down or inverted U design. Load balancers are typically used in this design for inbound server traffic as well as Global Load Balancing Protocol on outbound server traffic. The next option is to use NIC teaming within a VMware infrastructure as depicted in Figure 12.

802.3ad LAG 802.1q Trunking

AGGREGATION / CORE

EX82XX
802.3ad LAG 802.1q Trunking 802.3ad LAG 802.1q Trunking

EX82XX

EX4200

EX4200

Access
pNIC1 pNIC2

vNIC1 10.0.1.4/24 vNIC2 10.0.2.4/24

Figure 12: Inverted U design with NIC teaming


This option, using network interface teaming in a VMware environment, is the more prevalent form of an inverted U design. NIC teaming is a feature of VMware Infrastructure 3 that allows you to connect a single virtual switch to multiple physical Ethernet adapters. A team can share traffic loads between physical and virtual networks and provide passive failover in case of an outage. NIC teaming policies are set at the port group level.

34

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

For more detailed information on NIC teaming using VMware Infrastructure 3, refer to: www.vmware.com/technology/virtual-networking/virtual-networks.html www.vmware.com/files/pdf/virtual_networking_concepts.pdf The main advantages to the inverted U design are that all ports on the aggregation switches have useable bandwidth 100% of the time, traffic flows between access and aggregation are always deterministic, and there is a deterministic latency to all Virtual Chassis connected to a single aggregation or core switch.

Redundant Trunk Groups


LAN designs using the EX4200 Ethernet Switch with Virtual Chassis technology also benefit from RTG protocol as a built-in, optimized replacement to STP for sub-second convergence and automatic load balancing. RTG is an HA link feature of the EX Series Ethernet Switches that eliminates the need for STP. In fact, STP cant be provisioned on an RTG link. Ideally implemented on a switch with a dual-home connection, RTG configures one link as active and forwarding traffic, and the other link as blocking and backup to the active link. RTG provides extremely fast convergence in the event of a link failure. It is similar in practice to Rapid Spanning Tree Protocol (RSTP) root and alternate port, but doesnt require an RSTP configuration.

Layer Three Uplinks


Another alternative is to use L3 uplinks from the access layer to the aggregation/core switches. This limits L2 domain and VLANs to a single Virtual Chassis. Equal-cost multipath (ECMP), which is included in the base license for L3 OSPF on EX Series data center switches, is used for uplinks. An advanced license is needed if BGP, IS-IS, or Ipv6 are required. Up to 480 servers can be provisioned with one EX4200 Virtual Chassis. This allows applications to maintain L2 adjacency within a single switch, however servers are typically collocated within a single row which is part of the Virtual Chassis. The VLAN boundary is the access layer for low latency application data transfer. Virtual Chassis extension would allow distant servers to be accommodated as well. These servers would be part of a single Virtual Chassis domain.

Virtual Chassis Best Practices


These best practices should be followed when deploying and operating the EX4200 at the access layer: When designing a Virtual Chassis configuration, consider a deployment in which ports are distributed across as many switches as possible to provide the highest resiliency and the smallest failure domain. When possible, place uplink modules in the Virtual Chassis configuration line card switches, and place uplinks in devices which are separated at equal distances by member hop. Use the virtual management Ethernet (VME) interface as the management interface to configure Virtual Chassis technology options. Evenly space master and backup switches by member hop when possible. When installing a Virtual Chassis configuration, explicitly configure the mastership priority of the members that you want to function as the master and backup switches. When removing a switch from a Virtual Chassis configuration, immediately recycle its member ID so that the ID becomes the next lowest available unused ID. In this way, the replacement switch automatically is assigned that member ID and inherits the configuration of the original switch. Specify the same mastership priority value for the master and backup switches in a Virtual Chassis configuration. Configure the highest possible mastership priority value (255) for the master and backup switches. Place the master and backup switches in separate locations when deploying an extended Virtual Chassis configuration. For maximum resiliency, interconnect Virtual Chassis member devices in a ring topology. When changing configuration settings on the master switch, propagate changes to all other switches in the Virtual Chassis configuration via a commit sync command. Refer to Junipers Virtual Chassis Technology Best Practices Implementation Guide for more information: www.juniper.net/us/en/products-services/switching/ex-series/ex4200/#literature.

Copyright 2010, Juniper Networks, Inc.

35

Data Center LAN Migration Guide

Figure 13 below depicts an EX4200 Virtual Chassis top-of-rack access layer deployment.

TOR DEPLOYMENT
Legacy Aggregation Layer
L2/L3 Switch L2/L3 Switch

Access Layer

EX4200

Virtual Chassis

EX4200

Virtual Chassis

Figure 13: EX4200 top-of-rack access layer deployment


Access Layer Preinstallation Tasks
One of the first tasks is to determine space requirements for the new equipment. If the new access layer switches are to be housed in new racks, make sure that there is adequate space for the racks in the POD or data center. Or, if it is a switch refresh, ensure that there is sufficient space in the existing racks to accommodate the new switches. If the existing racks have the capacity, the eventual switchover becomes a matter of simply switching server cables from the old switches to the new ones. New racks are usually involved when a server refresh is combined with a switch refresh. It is assumed that data center facilities already have the power, cooling, airflow, and cabling required for any new equipment being provisioned. In a top-of-rack configuration, the EX4200 with Virtual Chassis technology can be logically viewed as a single chassis horizontally deployed across racks. It is important to understand the traffic profiles, since this determines the number of required uplinks. If the traffic flows are predominantly between servers, often referred to as east-west traffic, fewer uplinks to the core network layer are required since inter-server traffic primarily traverses the Virtual Chassis 128 Gbps backplane. The number of uplinks required is also a function of acceptable oversubscription ratios, which can be easily tuned as per the Virtual Chassis Best Practices section. The EX4200 with Virtual Chassis technology may also be used in an end-of-row deployment taking these same considerations into account. When connecting to an existing non-Juniper aggregation layer switch, its important to use open standard protocols such as 802.1Q for trunking VLANs and Multiple VLAN Registration Protocol (MVRP) if VLAN propagation is desired. To ensure interoperability, one of the standards-based STPs (IEEE STP, Rapid Spanning Tree Protocol, or Multiple Spanning Tree Protocol) should also be used. If they exist, company standard IOS-based access layer configurations should be collected. They can be quickly and easily converted into Junos OS using Junipers I2J translation tool as previously described. To simplify the deployment process and ensure consistent configurations when installing multiple access layer switches, Junos OS automation tools such as AutoInstall may be used. Refer to the following for more information: www.juniper.net/techpubs/software/junos-security/junos-security96/junos-security-admin-guide/configautoinstall-chapter.html. To further save time and ensure consistency, configuration files may be generated in advance, and then uploaded to a Trivial File Transfer Protocol (TFTP)/HTTP or FTP server for later downloading. The operations support systems (OSS) in place for switch provisioning and configuration management determine the type of server to use. To test feature consistency and feature implementations for the new Juniper access layer switches, a proof-ofconcept (PoC) lab could be set up, as previously noted. Feature and compatibility testing is greatly reduced with a single OS across all platforms such as Junos OS, which maintains a strict, serial release cycle. Feature testing in a

36

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

PoC lab could include: - Interface connectivity - Trunking and VLAN mapping - Spanning-tree interoperability with aggregation layer - QoS policyclassification marking, rate limiting - Multicast Provision the new access layer switches into any existing terminal servers for out of band CLI access via SSH. Network Automation: The approach to use for automating management in these scenarios depends on which other vendors equipment will be collocated in the data center, what other multivendor tools will be employed (Tivoli, OpenView, etc.), and how responsibilities will be partitioned across teams for different parts of the infrastructure. If the insertion is relatively small and contained, and other tools supporting each vendors platforms are already deployed, testing for compatible integration of the Juniper equipment is the best PoC step. However if this insertion is the start of a longer term, more expanded, strategic deployment of Juniper infrastructure into multiple PODs and sites (because of the importance of new applications, for example), then it may make sense to also test for integration of Junos Space into the environment because of the likely value of Space to the installation over time. Ethernet Design, Network Activate, or other Junos Space data center management applications could be considered. Space could be deployed as either a physical or a virtual appliance, depending on the allocation of responsibilities for management and the allocation of resources within the design. VM requirements for running Junos Space are 8 GB RAM and 40 GB hard disk in a production environment. If Junos Space VM is installed for lab/trial purposes, 2 GB RAM and 8 GB hard disk space is sufficient. For more information on deploying the Junos Space platform and managing nodes in the Junos Space fabric, please refer to the technical documentation: www.juniper.net/techpubs/en_US/junos-space1.0/information-products/ index-junos-space.html.

Installation
After comprehensive PoC testing, physical installation of the new POD should be relatively straightforward. As previously mentioned, if the installation only involves adding new access layer switches to existing racks, switchover becomes a matter of changing server cables from the old switches to the new ones. If a server refresh is combined with a switch refresh, those components should be physically installed in the new racks. Uplink interfaces are then connected to the aggregation switch(es).

Post Installation
Procedures which are similar to installing a new POD within a single vendor environment should be employed after physical installation is complete. As a best practice, verify: Access to the new POD via CLI or network management tools. Monitoring and alerting through OSS tools is functional. Security monitoring is in place and working. To limit false positives, security tuning should be conducted if a Security Incident and Event Manager (SIEM) system such as the STRM Series is in place. Interface status for server connections and uplinks via CLI or network management tools. L2/STP state between access and aggregation tiers. QoS consistency between access and aggregation tiers. Multicast state (if applicable). Traffic is being passed via ping tests and traffic monitoring tools. Applications are running as anticipated via end user test, performance management, and other verifications.

Copyright 2010, Juniper Networks, Inc.

37

Data Center LAN Migration Guide

Network Challenge and Solutions for Virtual Servers


Server virtualization has introduced a new layer of software to the data center called the hypervisor. This layer allows multiple dissimilar operating systems, and the applications running within them, to share the servers physical resources such as CPU, memory, and I/O using entities called virtual machines (VMs). A Virtual Ethernet Bridge (VEB) within the physical server switches traffic between the VMs and applications running within the same server, as well as between VMs and external network resources. Instead of interfacing directly to the physical NICs, the applications now talk to virtual ports on the internal VEB (or vSwitch). This creates virtual network endpoints, each with its own virtual IP and MAC addresses. Unique to server virtualization, a virtual switch (vSwitch) is a Layer 2 switch with limited function and security. Physical switches (or pSwitches) like the EX Series Ethernet Switches are actual pieces of network equipment. The network design for server virtualization requires an organization to consider the connections between the vSwitch and the pSwitch, including network functionality such as VLAN tagging and link aggregation (LAG) between network nodes. As VM to physical server ratios increase, server-based networking becomes more complex and may require multiple virtual switches, different VLANs, QoS tags, security zones, and more. Adding this amount of network processing to physical server infrastructures adds a great deal of overhead to their loads. It also requires more networking functionality to be added into hypervisors. To reduce this overhead and simplify operations, an alternative approach, strategically, is to use a Virtual Ethernet Port Aggregator (VEPA), where servers remain focused on application processing, and the hypervisor has visibility into the network, delegating switching tasks to collaborating physical switches. Networking for a large number of physical and logical servers can be simplified by attaching many of them to the wide configuration of a single Juniper Virtual Chassis switch. As previously described, EX Series switches, initially, the EX4200, can be logically combined into a single switch spanning over 400 physical server access ports and supporting several thousand virtual machines using Virtual Chassis technology. It is important to note that vSwitches and pSwitches are not competing technologies. The IEEE standards body is working on solving vSwitch issues within its VEPA standards working group. VEPA proposes to offload all switching activities from hypervisor-based vSwitches to the actual physical switches. In addition, VEPA proposes to ease management issues. The IEEE 802.1Qbg (Edge Virtual Bridging) and 802.1Qbh (Bridge Port Extension) VEPA standards specify how to move networking from virtual servers to dedicated physical Ethernet switches. This will help by concentrating networking functions into equipment that is purpose-built for tasks, enhancing performance, security, and management within the data centers. It will also help reduce computing overhead on virtual servers, allowing the CPU cycles to be spent on application processing as opposed to networking tasks. Offloading switching activities from hypervisor-based vSwitches will also result in an increase in the supported number of VMs. Most importantly, a ratified VEPA standard will be universal as opposed to a vendor-specific, proprietary solution. For more information on VEPA, please see: www.ieee802.org/1/files/public/docs2008/new-congdon-vepa-1108-v01.pdf. In addition to Virtual Chassis and VEPA-based forwarding, a management application such as Junos Space Virtual Control (expected availability in Q3 2010) allows users to monitor, manage, and control the virtual network environments that support virtualized servers deployed in the data center. Virtual Control provides a consolidated solution for network administrators to gain end-to-end visibility into, and control over, both virtual and physical networks from a single management screen.

Network Automation and Orchestration


Automation is a crucial tool for running a scalable and resilient data center at all levels. When developing an automation plan, it often works well to partition functions to be covered by the automation platforms and detail how these functions map into operations teams and data center domains. Within these boundaries, an open architecture should be selected that addresses the functions across the respective equipment suppliers platforms and maps them to the appropriate applications. In a multivendor infrastructure, a difference typically exists in the level at which vendor integration occurssometimes integration is possible in a multivendor application, and in other cases the integration is less complete and some number of each vendors specific tools needs to be kept in service to accomplish the necessary tasks.

38

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Significant levels of integration can be done in fault, performance, and network configuration and change management. As an example, IBMs Tivoli NetCool is a well established leading multivendor solution for fault management. Junipers Junos Space platform licenses and integrates NetCool into its Junos Space offering. There are also well established multivendor solutions for performance management that include IBM NetCool Proviso, CA eHealth, InfoVista, and HP OpenView that an enterprise can consider for managing network and application performance for both current vendor and Juniper network infrastructures. There are currently over a dozen Network Configuration and Change Management (NCCM) vendors with multivendor tools. These tools bring more structure to the change management process and also enable automated configuration management. NCCM vendors include IBM Tivoli, AlterPoint, BMC (EmprisaNetworks), EMC (Voyence), HP (Opsware), and others. Prior to introducing Juniper into an existing single vendor infrastructure, Juniper recommends that you replace manual network configuration management processes and vendor-specific tools with automated multivendor NCCM tools. It is also good practice to establish standard network device configuration policies which would apply to all vendors in the network infrastructure. Automated network configuration management is more efficient and also reduces operational complexity. An IT management solution should be built around the standards outlined by the Fault, Configuration, Accounting, Performance, Security (FCAPS) Model. Refer to the following URL for more information: http://en.wikipedia.org/wiki/FCAPS.

Integrating Management and Orchestration with Junos Space


Some amount of integration for management and orchestration can be accomplished using Junos Space. Junos Space is a network application platform for developing and deploying applications that simplify data center operations. Junos Space abstracts network intelligence such as traffic flows, routing topology, security events, and user statistics (to name a few), and makes this available as services used both by Juniper-developed applications and applications from third-party vendors leveraging the Junos Space SDKs/APIs. Junos Space applications are designed to be collaborative. They share common services such as inventory and configuration management, job scheduling, HA/clustering, etc., and they use a common security framework. This enables users to optimize scale, security, and resources across their application environment. The Junos Space application portfolio currently includes Ethernet Design (rapid endpoint and switch port configuration in campus/data center environments), Security Design (fast, accurate deployment of security devices and services), Network Activate (point/click setup and management of VPLS services), Route Insight (visibility, troubleshooting, and change modeling for L3 MPLS/IP networks), and Service Now (automated case and incident management).

Data Center Consolidation Trigger Event


Numerous large enterprises are consolidating their geographically distributed data centers into mega data centers to take advantage of cost benefits, economies of scale, and increased reliability, and to fully exploit the latest virtualization technologies. According to industry research, more than 50% of the companies surveyed had consolidated data centers within the last year and even more planned to consolidate in the upcoming year. Data center consolidation can involve multiple insertion points into access and core aggregation layers as well as consolidation of security services.

Best Practices in Designing for the Access Layer Insertion Point


We have already discussed the recommended best practices for the access layer insertion point. In this section, we will highlight key best practice/design considerations for the other insertion points related to the consolidation trigger event.

Copyright 2010, Juniper Networks, Inc.

39

Data Center LAN Migration Guide

Best Practices: Designing the Upgraded Aggregation/Core Layer


The insertion point for an enterprise seeking to initially retain its existing three-tier design may be at the current designs aggregation layer. In this case, the recommended Juniper aggregation switch, typically an EX8200 line or MX Series platform, should be provisioned in anticipation of it eventually becoming the collapsed core/aggregation layer switch. If the upgrade is focused on the aggregation layer in a three-tier design (to approach transformation of the data center network architecture in an incremental way, as just described), the most typical scenario is for the aggregation switches to be installed as part of the L2 topology of the data center network, extending the size of the L2 domains within the data center and interfacing them to the organizations L3 routed infrastructure that typically begins at the core tier, one tier upin the design from the aggregation tier. In this case, the key design considerations for the upgraded aggregation tier include: Ensuring sufficient link capacity for the necessary uplinks from the access tier, resiliency between nodes in the aggregation tier, and any required uplinks between the aggregation and core tiers in the network Supporting the appropriate VLAN, LAG, and STP configurations within the L2 domains Incorporating the correct configurations for access to the L3 routed infrastructure at the core tier, especially for knowledge of the default gateway in a VRRP (Virtual Router Redundancy Protocol) environment Ensuring continuity in QoS and policy filter configurations appropriate to the applications and user groups supported At a later point of evolution (or perhaps at the initial installation, depending on your requirements), it may be that these nodes perform an integrated core and aggregation function in a two-tier network design. This would be the case if it suited the organizations economic and operational needs, and could be accomplished in a well managed insertion/upgrade. In such a case, the new consolidated core of the network would most typically perform both L2 and L3 functions. The L2 portions would, at a minimum, include the functions described above. They could also extend or stretch the L2 domains in certain cases to accommodate functions like application mirroring or live migration of workloads in a virtual server operation between parts of the installation in other areas of the data center or even in other data centers. We describe design considerations for this later in this guide. In addition to L2 functions, this consolidated core will provide L3 routing capabilities in most cases. As a baseline, the L3 routing capabilities to be included are: Delivering a resilient interface of the routed infrastructure to the L2 access portion of the data center network. This is likely to include VRRP default gateway capabilities. It is also likely to include one form or another of an integrated routing/bridging interface in the nodes such as routed VLAN interfaces (RVIs), or integrated routing and bridging interfaces (IRBs), to provide transition points between the L2 and L3 forwarding domains within the nodes. Resilient, HA interfaces to adjacent routing nodes, typically at the edge of the data center network. Such high availability functions can include nonstop active routing (NSR), GRES, Bidirectional Forwarding Detection (BFD), and even MPLS fast reroute depending on the functionality and configuration of the routing services in the site. For definitions of these terms, please refer to the section on Node-Link Resiliency. MPLS fast reroute is a local restoration network resiliency mechanism where each path in MPLS is protected by a backup path which originates at the node immediately upstream. Incorporation of the appropriate policy filters at the core tier for enforcement of QoS, routing area optimization, and security objectives for the organization. On the QoS level, this may involve the use of matching Differentiated Services code points (DSCPs) and MPLS traffic engineering designs with the rest of the routed infrastructure to which the core is adjacent at the edge, as well as matching priorities with the 802.1p settings being used in the L2 infrastructure in the access tier. On the security side, it may include stateless filters that forward selected traffic to security devices such as firewall/IDP platforms at the core of the data center to enforce appropriate protections for the applications and user groups supported by the data center (see the next section of the core tier best practices for a complementary discussion of the firewall/IDP part of the design). In some cases, the core design may include use of VPN technologymost likely VPLS and MPLSto provide differentiated handling of traffic belonging to different applications and user communities, as well as to provide special networking functions between various data center areas, and between data centers and other parts of the organizations network.

40

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

The most common case will be use of VPLS to provide stretched VLANs between areas of a large data center network, or between multiple distant data centers using VPLS (over MPLS) to create a transparent extension of the LAN to support nonstop application services (transparent failovers), transaction mirroring, data base backups, and dynamic management of virtual server workloads across multiple data center sites. In these cases, the core nodes will include VPLS instances matching the L2 topology and VLAN configurations required by the applications, as well as the appropriate implementation of MPLS between the core nodes and the rest of the organizations routed IP/MPLS network. This design will include ensuring high availability and resilient access of the L2 access tier into the elastic L2 infrastructure enabled by VPLS in the core; use of appropriate traffic engineering, and HA features of MPLS to enable the proper QoS and degree of availability for the traffic being supported in the transparent VLAN network. Details on these design points are included in the section of the Migration Guide on incorporating multiple sites into the data center network design using MPLS in the Six Process Steps for Ensuring MPLS Migration section.

Best Practices: Upgraded Security Services in the Core


Frequently a data center consolidation requires consolidating previously separate and siloed security appliances into a more efficient security tier integrated into the L2 and L3 infrastructures at the core network layer. Here we describe design considerations for accomplishing that integration of security services in the core. All security appliances should be consolidated and virtualized into a single pool of security services with a platform such as the SRX Series Services Gateways. To connect to and protect all core data center network domains, the virtual appliance tier should optimally participate in the interior gateway routing protocols within the data center network. Security zones should be defined to apply granular and logically precise protection for network partitions and virtualized resources within the network wherever they reside, above and beyond the granularity of traditional perimeter defenses. The security tier should support the performance required by the data centers applications and be able to inspect information up to L7 at line rate. A powerful application decoder is necessary on top of the forwarding, firewall filtering, and IDP signature detection also applied to the designated traffic streams. Including this range of logic modularly in a high-performance security architecture for the core helps reduce the number of devices in the network and increase overall efficiency. Scalable, strong access controls for remote access devices and universal access control should be employed to ensure that only those with an organizational need can access resources at the appropriate level. Integration of secure access with unified policies and automation using coordinated threat control not only improves security strength but also increases efficiency and productivity of applications within the data center. Finally, incorporation of virtual appliances such as virtual firewalls and endpoint verification servers into the data centers security design in a way that integrates protection for the virtual servers, desktops, and related network transports provides an extension of the common security fabric into all of the resources the IT team needs to protect.

Aggregation/Core Insertion Point Installation Tasks


Preinstallation Tasks
The tasks described in this section pertain to a consolidation within an existing data center that has the required space and power to support consolidation. Alternatively, a consolidation may take place in a new facility, sometimes referred to as a greenfield installation. That scenario would follow the best practices outlined in Junipers Cloud-Ready Data Center Reference Architecture: www.juniper.net/us/en/solutions/enterprise/data-center/simplify/#literature. The steps outlined here also apply to a case in which the organization wants to stay with its existing three-tier design, at least for the initial steps in the process. In such a case, deployment and provisioning should be done leaving flexibility to move to a two-tier design at some future date.

Copyright 2010, Juniper Networks, Inc.

41

Data Center LAN Migration Guide

The scenario for transitioning the core data center network in a design upgrade triggered by a consolidation project should encompass the following kinds of tasks, suited to each organizations case: Ensure that the power, cooling, airflow, physical rack space, and cabling required for any new equipment has been designed, ordered, and installed (however your organization breaks down these responsibilities between departments, suppliers, and integrators). Size the new/upgraded switching platforms using your organizations policy for additional performance and capacity headroom for future growth. Design should include a pair of switches that can eventually serve as the new collapsed core/aggregation layer. Initial design can be tuned to the exact role the switches will perform. For example, if they will be focused on pure core functions in the initial phase, they can be focused on capacity/functionality required for that role (e.g., IGP policies and area design appropriate to a core data center switch). If they will be focused on pure aggregation functions initially, design can focus on L2 and L3 interface behaviors appropriate to that role. Or, if they will be performing a blended core/aggregation role (as would be appropriate in many two-tier data center networks), a mix of L2 and L3 functions can be designed to fit the networks requirements. If the new switches are aggregating existing access layer switches, industry standard protocols (such as 802.1D, 802.1Q, 802.1s, 802.3ad, etc.) should be used to ensure interoperability. Configuration checklist should include: - IGP and EGP requirements such as area border roles, routing metrics and policies, and possible route redistributions - L2/L3 domain demarcations VLAN to VRF mapping Virtualized service support requirements (VPN/VRF) Default gateway/root bridge mapping Hot Standby Routing Protocol (HSRP) to Virtual Router Redundancy Protocol (VRRP) mappings - Uplink specifications Density Speeds (number of GbE and 10GbE links) Link aggregations Oversubscription ratios - Scaling requirements for MAC addresses - Firewall filters (from IOS ACL mapping) - QoS policies - Multicast topology and performance - Audit/compliance requirements should be clarified and any logging or statistics collection functions designed in. - IOS configurations should be mapped to Junos OS as outlined in the section on IOS to Junos OS translation tools. As noted earlier, a PoC lab could be set up to test feature consistency and implementation in the new Juniper infrastructure. Testing could include: - Interface connections - Trunking and VLAN mapping - STP interoperability with any existing switches - QoS policy (classification marking; rate limiting) - Multicast

42

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Installation Tasks
Refer to Figure 14 when considering the tasks described in this section:
EX8200 EX8216 West
Legacy Switch West

EX8200 EX8216 East


Legacy Switch East

Legacy Aggregation Layer Switches and Security Applicances

Legacy Access Layer

Figure 14: Aggregation/core layer insertion point


If the new aggregation layer switches are part of a new POD that includes new access layer switches, the installation should be straightforward. If the new aggregation layer switches are part of a replacement for existing legacy switches, it is important to create a fallback position in the event of any unforeseen issues. Typically, EX8200 line switches would be provisioned and deployed as a pair, replacing the legacy switches. Once the EX8200 line switches are installed, they should be connected to the existing core layer. Again, this scenario is appropriate in cases where the organization initially will maintain its existing three-tier architecture. Appropriate IGP and EGP configurations as identified in the preinstallation configuration checklist would have been provisioned and interoperability verified by checking neighbor state and forwarding tables. An initial access layer switchs uplinks would then be connected to the EX8200 line switches and connectivity verified as outlined in the post installation checklist. Once this baseline is established, additional access layer switches would then be migrated to the EX8200 line.

Post Installation
As previously noted, procedures which are similar to those employed when installing a new aggregation switch in a single vendor environment can be used, after physical installation is complete, to verify successful operations. As a best practice, verify: Access to new configuration via CLI or network management tools. Interface status for server connections and uplinks via CLI or network management tools. L2/STP status between network tiers. QoS consistency between network tiers. Multicast state (if applicable). Traffic passing via ping tests. Application connectivity and flows with statistics or end user tests (or both).

Copyright 2010, Juniper Networks, Inc.

43

Data Center LAN Migration Guide

Consolidating and Virtualizing Security Services in the Data Center: Installation Tasks
In addition to cyber theft and increasing malware levels, organizations must guard against new vulnerabilities introduced by data center technologies themselves. To date, security in the data center has been applied primarily at the perimeter and server levels. However, this approach isnt comprehensive enough to protect information and resources in new system architectures. In traditional data center models, applications, compute resources, and networks have been tightly coupled, with all communications gated by security devices at key choke points. However, technologies such as server virtualization and Web services eliminate this coupling and create a mesh of interactions between systems that create subtle and significant new security risks within the interior of the data center. For a complete discussion of security challenges in building cloud-ready, next-generation data centers, refer to the white paper, Security Considerations for Cloud-Ready Data Center: www.juniper.net/us/en/local/pdf/implementationguides/8010046-en.pdf. A key requirement for this insertion point is for security services platforms to provide the performance, scalability, and traffic visibility needed to meet the increased demands of a consolidated data center. Enterprises deploying platforms which do not offer the performance and scalability of Juniper Networks SRX Series Services Gateways and their associated management applications are faced with a complex appliance sprawl and management challenge, where numerous appliances and tools are needed to meet requirements. This is a more costly, less efficient, and less scalable approach.

Preinstallation Tasks for Security Consolidation and Virtualization

SRX5800

EX82XX

Legacy Security Appliances

Figure 15: SRX Series platform for security consolidation

Ensure that the appropriate power, cooling, airflow, physical rack space, and cabling required to support the new equipment have been ordered and installed. Ensure that the security tier is sized to meet the organizations requirements for capacity headroom for future growth. Define and provision routing/switching Infrastructure first (see prior section). This sets the L3/L2 foundation domains upon which the security zones the SRX Series enforces will be built. The SRX Series supports a pool of virtualized security services that can be applied to any application flow traversing the data center network. Setting up the network with this foundation of subnets and VLANs feeding the dynamic security enforcement point segments the data center resources properly and identifies what is being protected and what level of protection is needed. With SOA, for example, there are numerous data flows between servers within the data center and perimeter but security is often insufficient for securing these flows. Policies based on role/function, applications, business goals, or regulatory requirements can be achieved using a mix of VLAN, routing, and security zone policies enabling the SRX Series to enforce the appropriate security posture for each flow in the network. The performance and scalability requirements should be scoped. SRX Series devices can be paired together in a cluster to scale to 120 Gbps of firewall throughput, as well as providing HA. Virtual machine security requirements should also be defined. Junipers partnership with Altor Networks is hypervisor neutral, eliminating VM security blind spots. For more information on Altors virtual firewall solution, refer to Chapter 2 (Altor Networks Virtual Firewall).

44

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

In the preinstallation phase, security policies must be developed. This typically takes time and can be complex to coordinate. Juniper Professional Services can be used as a resource to help analyze and optimize security policies at all enforcement points. The full suite of Juniper Networks Professional Services offerings can be found at: www.juniper.net/us/en/products-services/consulting-services. Establish a migration plan, identifying a time line and key migration points, if all appliances cannot be migrated in a flash cut. As with the other insertion points, PoC testing can be done and could include: - Establishing the size of the target rule base to be used post conversion - Checking the efficacy of the zone definitions - Determining the effectiveness of the IPS controls - Determining the suitability and implementation of the access controls to be used

Installation Tasks
As with the aggregation/core insertion point, it is important to have a fallback position to existing appliances in the event of any operational issues. The current firewall appliances should be kept on hot standby. The key to a successful migration is to have applications identified for validation and to have a clear test plan for success criteria. There are three typical options for migration, with option 1 being the one most commonly used. Migration Test Plan (Option 1) Test failover by failing the master firewall (legacy vendor) to the backup (this confirms that HA works and the other devices involved in the path are working as expected). Replace the primary master, which was just manually failed, with the Juniper firewall. The traffic should still be flowing through the secondary, which is the legacy vendor firewall. Turn off the virtual IP (VIP) address or bring the interface down on the backup (legacy vendor firewall) and force everything through the new Juniper firewall. A longer troubleshooting window helps to ensure that the switchover has happened successfully. Also, turn off synchronization checks (syn-checks) initially, to process already established sessions, since the TCP handshake has already occurred on the legacy firewall. This will ensure that the newly established Juniper firewall will not drop all active sessions as it starts up. Migration Test Plan (Option 2) This is essentially a flash cut option where an alternate IP address for the new firewall is configured along with the routers and the hosts then point to the new firewall. If there is an issue, gateways and hosts can then be provisioned to fall back to the legacy firewalls. With this option, organizations will sometimes choose to leave IPsec VPNs or other termination on their old legacy firewalls and gradually migrate them over a period of time. Migration Test Plan (Option 3) This option is typically used by financial organizations due to the sensitive nature of their applications. A Switched Port Analyzer (SPAN) session will be set up on the relevant switches with traffic sent to the Juniper firewalls, where traffic is analyzed and session tables are built. This provides a clear understanding of traffic patterns and provides more insight into the applications being run. This option also determines whether there is any interference due to filter policies or IPS, and it creates a more robust test and cutover planning scenario. This option typically takes more time than the other options, so organizations typically prefer to go with option 1. Again, this is a more common option for companies in the financial sector.

Copyright 2010, Juniper Networks, Inc.

45

Data Center LAN Migration Guide

Post Installation
As previously noted, procedures similar to those used when installing new security appliances in a single vendor environment could be used after physical installation is complete. As a best practice: Verify access via CLI or network management tools. Verify interface status via CLI or network management tools. Verify that traffic is passing through the platform. Verify that rules are operational and behaving as they should. Confirm that Application Layer Gateway (ALG) policies/IPS are stopping anomalous or illegal traffic in the application layer, while passing permitted traffic. Confirm that security platforms are reporting appropriately to a centralized logging or SIEM platform.

Business Continuity and Workload Mobility Trigger Events


Sometimes an improvement in availability of systems to external and internal users drives a critical initiative to enhance the availability of data center infrastructures, either within an individual data center or between sets of data centers such as primary, backup, and distributed data center sites. The goal is almost always to preserve a business value to its stakeholders, and it often requires upgrades or extensions to critical infrastructure areas to achieve this goal. Business continuity or disaster recovery sites can be set up as active/active, warm-standby, or cold-standby configurations. A cold-standby site could involve an agreement with a provider such as SunGuard in which backup tapes are trucked to a SunGuard backup data center facility. A warm-standby site could interconnect primary and standby data centers for resumption of processing after a certain amount of backup/recovery system startup has occurred. A hot-standby, active/active configuration involves continuously available services running in each site that allow transparent switching between primary and secondary as needed, driven by planned or unplanned outages. A large organization may have instances of each. Business continuity and workload mobility are tightly coupled. Business continuity or high availability disaster recovery (HADR) often involves provisioning between two or more data centers. The design could involve replicating an entire data center (essentially a Greenfield installation), or the design could involve adding additional capacity to one or more existing data centers. The specific insertion points could be at any of the tiers of an existing three-tier design. We have already outlined best practices and specific installation tasks for several of these network insertion points in this chapter. Once provisioning for the disaster recovery data center has been done, users should be able to connect into any of the data centers transparently. Since we have already described the installation tasks for access and aggregation/core switching and services tiers of the new data center network, we wont repeat those here. The same procedures can be used to enhance the data center infrastructures that will take part in the HADR system. To the extent that MPLS and VPLS are involved in the configuration between centers, we will address the steps associated with that part of the network in the section on workload mobility further on in this guide.

Best Practices Design for Business Continuity and HADR Systems


Business continuity is enabled using a mix of device-level, link-level, and network-level resiliency within and between an organizations data center sites. In most cases, it also involves application and host system resiliency capabilities that need to interwork seamlessly with the network to achieve continuity across multiple sites. In this section, we first concentrate on the network-level design within the data center sites. In the following section (on workload mobility), we also describe capabilities that extend continuity to the network supporting multiple data center sites and to certain considerations around host and application resiliency interworking with the network.

46

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Link-level redundancy in data center networks can be implemented with the following network technologies: - Link Aggregation Group (LAG) - Redundant Trunk Groups (RTGs) - Spanning Tree Protocol (STP) and its variations - Bidirectional Forwarding Detection (BFD) - MPLS We have already discussed LAG, RTG, and STP earlier in this guide. BFD is rapidly gaining popularity in data center deployments, because it is a simple protocol aiding rapid network convergence (30 to 300 ms resulting in a subsecond convergence time). BFD is a simple low layer protocol involving a hello mechanism between two devices. The communication can be across directly connected links or across a virtualized communications path like MPLS. Node level resiliency can be achieved using the following technologies: - Graceful Routing Engine switchover (GRES) - Graceful restart - Nonstop active routing (NSR) and nonstop bridging (NSB) GRES is a feature used to handle planned and unplanned platform restarts gracefully, without any disruptions, by deploying a redundant Routing Engine in a chassis. The Routing Engines synchronize and share their forwarding state and configuration. Once synchronized, if the primary Routing Engine fails due to a hardware or software problem, the secondary Routing Engine comes online immediately, resulting in minimum traffic forwarding interruption. Rather than being a feature, graceful restart is a standards-based protocol that relies on routing neighbors to orchestrate and help a restarting router to continue forwarding. There is no disruption in control or in the forwarding path when the graceful restart node and its neighbors are participating fully and are employing the standard procedures. NSR builds on GRES and implements a higher level of synchronization between the Routing Engines. In addition to the synchronizations and checkpoints between the Routing Engines that GRES achieves, NSR employs additional protective steps and results in no disruption to the control or data planes, hiding the failure from the rest of the network. And, it does not require any help from its neighbors to achieve these results. Note that graceful restart and NSR are mutually exclusivethey are two different means to achieve the same high availability goal. NSB is similar to GRES and preserves interface and Layer 2 protocol information. In the event of a planned or unplanned disruption in the primary Routing Engine, forwarding and bridging are continued during the switchover resulting in minimal packet loss. Node-level HA includes many aspects, starting with the architecture of the node itself and ending with the protocols that the node uses to network with other components. Paralleling the Open Systems Interconnection (OSI) Reference Model, you can view the network infrastructure components starting from the physical layer, which is the nodes internal architecture and protocols, and ending with the upper tiers, which would include components such as OSPF, IS-IS, BGP, MPLS, GRES, NSR, etc. No single component provides HA. It is all of the components working together which creates one architecture and results in high availability. For more detailed information on HA features on Juniper platforms, refer to: www.juniper.net/techpubs/en_US/ junos10.1/information-products/topic-collections/swconfig-high-availability/frameset.html. To complement the node-level availability mechanisms highlighted above, devices and systems deployed at critical points in the data center design should include redundancy of important common equipment such as power supplies, fans, and Routing Engines at the node levels, so that the procedures mentioned above can have a stable hardware environment to build upon. In addition, the software/firmware in these devices should be based on a modular architecture to prevent software failures or upgrade events from impacting the entire device. There should also be a clean separation between control plane and data processes to ensure system availability. Junos OS is an example of a multitasking OS that operates in this manner, ensuring that a failure in one process doesnt impact any others.

Copyright 2010, Juniper Networks, Inc.

47

Data Center LAN Migration Guide

Best Practices Design to Support Workload Mobility Within and Between Data Centers
In current and future data centers, the application workloads will increasingly be handled on an infrastructure of extensively virtualized servers, storage, and supporting network infrastructure. In these environments, operations teams will frequently need to move workloads from test into production, and distribute the loads among production resources based on performance, time of day, and other considerations. This kind of workload relocation among servers may occur between servers within the same rack, within the same data center, and increasingly, between multiple data centers depending on the organizations size, support for cloud computing services for bursting overloads, and other similar considerations. In general, we refer to this process of relocating virtual machines and their supporting resources as workload mobility. Workload mobility plays an important part in maintaining business continuity and availability of services, and, as such, is discussed within this section of the guide. Workload mobility can be deployed in two different ways, as shown in Figure 16.

VPLS

Virtual Chassis

Virtual Chassis

Virtual Chassis

Cloud Center

Cloud Center

Rack A

Layer 2 domain across racks and across data center

Rack B

Layer 2 domain across virtual private LAN

RACK TO RACK

CLOUD TO CLOUD

Figure 16: Workload mobility alternatives


With Junipers Virtual Chassis technology, a Layer 2 domain can be easily extended across racks or rows, allowing server administrators to move virtual machines quickly and easily within a data center. When moving workloads between data center sites, consideration should be given to the latency requirements of applications spanning the data center sites to be sure that the workload moves can meet application needs. If it is important to support the movement of workloads between data centers, the L2 domain supporting the VM infrastructure can be extended, or stretched, across data centers in two different ways: If the data centers are under 100 km apart, and the sites are directly connected physically, a Virtual Chassis can be extended across the sites, using the 10 Gigabit chassis extension ports supporting a configuration up to the maximum of 10 switches in a single Virtual Chassis. Note that directly connected means that ports are directly connected to one another, i.e., a single fiber link connecting one device to another device or L1 signal repeater is used. If the connection between data centers is at a distance requiring services of a WAN, MPLS can be used to create transparent virtual private LANs at very high scale using a Juniper infrastructure.

48

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

At the level of the data center network core and the transparent virtual WAN that an organization can use to support its business continuity and workload mobility goals, MPLS provides a number of key advantages over any other alternative: MPLS virtualization enables the physical network to be run as many separate virtual networks. The benefits include cost savings, improved privacy through traffic segmentation, improved end user experience with traffic engineering and QoS, and improved resiliency with functionality such as MPLS fast reroute and BFD. This can be done in a completely private network context (e.g., the enterprise owns the entire infrastructure), or it can be achieved through the interworking of the organizations private data center and WAN infrastructures with an appropriately deployed carrier service. VPLS provides Ethernet-based point-to-point, point-to-multipoint, and multipoint-to-multipoint (full mesh) transparent LAN services over an IP/MPLS infrastructure. It allows geographically dispersed LANs to connect across an MPLS backbone, allowing connected nodes (such as servers) to interpret that they are on the same Ethernet LAN. VPLS thus provides an efficient and cost-effective method for communicating at L2 across two or more data center sites. This can be useful for transaction mirroring in active/active or other backup configurations. And it is necessary for supporting workload mobility and migration of virtual machines between locations over a WAN. MPLS can provide private L3VPN networks between data center sites that share the same L3 infrastructure. A composite, virtualized L2 and L3 infrastructure can thus be realized. Very useful security properties can be achieved in such a design as well. For example, by mapping L3VPNs to virtual security zones in an advanced firewall such as the SRX Series, many security policies can be selectively layered on the traffic. Also in support of business continuity, MPLS traffic engineering (TE) and fast reroute capabilities combine sophisticated QoS and resiliency features into a multiservice packet core for superior performance and economics. TE could be used to support real-time data replication and transaction mirroring, along with service-level agreement (SLA) protection for real-time communications such as video conferencing and collaboration services. Fast reroute delivers rapid path protection in the packet-based network without requiring redundant investments in SONET or SDH level services (e.g., superior performance for lower cost). For workload mobility that involves extending a Layer 2 domain across data centers to support relevant applications like VMware VMotion, archiving, backup, and mirroring, L2VPNs using VPLS could be used between data center(s). VPLS allows the connected data centers to be in the same L2 domain, while maintaining the bandwidth required for backup purposes. This feature ensures that other production applications are not overburdened.

Best Practices for Incorporating MPLS/VPLS in the Data Center Network Design
Current L2/L3 switching technologies designed for the LAN do not scale well with the appropriate levels of rerouting, availability, security, QoS, and multicast capabilities to achieve the required performance and availability. As a result, when redesigning or upgrading the data center, an upgrade to MPLS is frequently appropriate and justified to meet business operational demands and cost constraints. MPLS often simplifies the network for the data center, removing costly network equipment and potential failure points while providing complete network redundancy and fast rerouting. When fine grained QoS is required with traffic engineering for the data center, RSVP should be used to establish bandwidth reservations based upon priorities, available bandwidth, and server performance capacities. MPLS-based TE is a tool made available to the data center network administrators which is not presently available in common IP networks. Furthermore, MPLS virtualization capabilities can be leveraged to segment and secure server access, becoming a very important part of maintaining a secure data center environment. For this section of the Data Center LAN Migration Guide, the prior construct of best practice, preinstall, install, and post install is going to be combined into six process steps for migrating to MPLS, keeping in mind that VPLS runs over an IP/ MPLS network.

Copyright 2010, Juniper Networks, Inc.

49

Data Center LAN Migration Guide

Switching across data centers using VPLS is depicted below in Figure 17.

CORE

M Series

M Series

MX Series

MX Series

EX4200
Virtual Chassis

EX4500

EX4200
Virtual Chassis

EX4500

Apps

Apps

Apps

Apps

VLAN 10

VLAN 20

Mirroring VLAN 10

Mirroring VLAN 20

DATA CENTER 1

DATA CENTER 2

Figure 17: Switching across data centers using VPLS Six Process Steps for Migrating to MPLS
The following approach using a phased series of steps is one we have found useful in many enterprises. However, there may be specific circumstances which could dictate a different approach in a given case. Step 1: Upgrade the IP network to MPLS-capable platforms, yet continue to run it as an IP network. In step one, upgrade the routers connecting the data centers to routers capable of running MPLS, yet configure the network as an IP network without MPLS. Use this time to verify a stable and properly performing inter data center connection. This will provide the opportunity to have the MPLS network in place and to be sure routers are configured and working correctly to support IP connectivity. If youre presently running Extended IGRP (EIGRP), use this opportunity to migrate to OSPF or one of the other L3 protocols that will perform better with MPLS. Depending upon how many data centers will be inter connected, once youve migrated to OSPF and/or IS-IS, it is a good time to enable BGP as well. BGP can be used for automatic MPLS label distribution. Juniper has multiple sources of design guidelines and practical techniques for accomplishing these tasks, which can be delivered in either document-based or engineering professional services modes. Please refer to the Additional Resources sections in Chapter 5 for specific URLs.

50

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Step 2: Build the MPLS layer. Once you have migrated to an MPLS capable network and tested and verified its connectivity and performance, activate the MPLS overlay and build label-switched paths (LSPs) to reach other data centers. Label distribution is automated with the use of LDP and RSVP with extensions, to support creation and maintenance of LSPs and to create bandwidth reservations on LSPs (RFC 3209). BGP can also be used to support label distribution at the customers choice. The choice of protocol for label distribution on the network depends on the needs of the organization and applications supported by the network. If traffic engineering or fast reroute are required on the network, you must use RSVP with extensions for MPLS label distribution. It is the decision of whether or not to traffic engineer the network or require fast reroute that frequently makes the decision between the use of LDP or RSVP for MPLS label distribution. Step 3: Configure MPLS VPNs. MPLS VPNs can segregate traffic based on departments, groups, or users, as well as by applications or any combination of user group and application. Lets take a step back and look at why we call MPLS virtualized networks VPNs. First, they are networks because they provide connectivity between separately defined locations. They are private because they have the same properties and guarantees as a private network in terms of network operations and in terms of traffic forwarding. And lastly, they are virtual because they may use the same transport links and routers to provide these separated transport services. Since each network to be converged onto the newly built network has its own set of QoS, security, and policy requirements, you will want to define MPLS-based VPNs that map to the legacy networks already built. MPLS VPNs can be defined by: Department, business unit, or other function: Where there is a logical separation of traffic that goes to a more granular level than the network, perhaps down to the department or business unit, application, or specific security requirement level, you will want to define VPNs on the MPLS network, each to support the logical separation required for a unique QoS, security, and application support combination. Service requirements: In the context of this Data Center LAN Migration Guide, VPLS makes it appear as though there is a large LAN extended across the data center(s). VPLS can be used for IP services or it can be used to interconnect with a VPLS service provider to seamlessly network data center(s) across the cloud. IP QoS in the LAN can be carried over the VPLS service with proper forwarding equivalence class (FEC) mapping and VPN configuration. If connecting to a service providers VPLS service, you will either need to collocate with the service provider or leverage a metro Ethernet service, as VPLS requires an Ethernet hand-off from the enterprise to the service provider. QoS needs: Many existing applications within your enterprise may run on separate networks today. To be properly supported, these applications and their users make specific and unique security and quality demands on the network. This is why, as a best practice, it is suggested that you start by creating VPNs that support your existing networks. This is the minimum number of VPNs you will need. Security requirements: Security requirements may be defined by user groups such as those working on sensitive and confidential projects, by compliance requirements to protect confidential information, and by application to protect special applications. Each special security zone can be sectioned off with enhanced security via MPLS VPNs. Performance requirements: Typically, your applications and available bandwidth will determine traffic engineering and fast reroute requirements, however users and business needs may impact these considerations as well. Additional network virtualization: Once MPLS-based VPNs are provisioned to support your existing networks, user groups, QoS, and security requirements, consideration should be given to new VPNs that may be needed. For example, evolving compliance processes supporting requirements such as Sarbanes-Oxley or Health Insurance Portability and Accountability Act (HIPAA) may require new and secure VPNs. Furthermore, a future acquisition of a business unit may require network integration and this can easily be performed on the network with the addition of a VPN to accommodate the acquisition. Step 4: Transfer networks onto the MPLS VPNs. All of the required VPNs do not have to be defined before initiating the process of migrating the existing network(s) to MPLS. In fact, you may wish to build the first VPN and then migrate the related network, then build the second VPN and migrate the next network and so on. As the existing networks converge to the MPLS network, monitor network performance and traffic loads to verify that expected transport demands are being met. If for some reason performance or traffic loads vary from expected results, investigate further as MPLS can provide deterministic traffic characteristics, and resulting performance should not vary greatly from the expected results. Based upon findings, there may be opportunities to further optimize the network for cost and performance gains.

Copyright 2010, Juniper Networks, Inc.

51

Data Center LAN Migration Guide

Step 5: Traffic engineer the network. Step 5 does not require steps 3 or 4 above to be completed before initiating this step. Traffic engineering may begin as soon as the MPLS network plane is established. However, as a best practice it is recommended to first migrate some of the existing traffic to the MPLS plane before configuring TE. This will allow you to experience firsthand the benefits and granular level of control you have over the network through the traffic engineering of an MPLS network. Start by assessing the existing traffic demand of applications across data center(s). Group traffic demand into priority categories, for instance, voice and video may be gathered into a real time priority category, while private data is grouped into a second and Internet traffic is grouped into a third category. Step 6: Monitor and manage. As with any network, you must continue to monitor and manage the network once it is deployed and running while supporting new service loads and demands. An advantage MPLS provides above and beyond IP is its capability to traffic engineer based upon utilization and application demands as the business evolves. For more information on MPLS, refer to: www.juniper.net/techpubs/software/junos/junos53/swconfig53-mplsapps/html/mpls-overview.html. For more information on VPLS, refer to: www.juniper.net/techpubs/en_US/junos10.2/information-products/ pathway-pages/config-guide-vpns/config-guide-vpns-vpls.html.

High-Performance Security Services Trigger Event


Please refer to the best practices and related installation information outlined in the Upgrading Security Services in the Core section. as that trigger event covers the security services network insertion point.

Completed Migration to a Simplified, High-Performance, Two-Tier Network


As discussed throughout this document, enterprises with an existing legacy three-tier data center architecture can begin their migration to a next-generation, cloud-ready, Juniper-based two-tier design from any of the common trigger events outlined in this chapter. Weve identified the best practices and the key prescriptive installation steps needed to ensure successful insertion into the existing architecture. You can transition to a simplified architecture from an existing legacy multitier architecture, as shown in Figure 18, by provisioning each of these network insertion points.

3-TIER LEGACY NETWORK


Core

SIMPLER, 2-TIER DESIGN

Ethernet

Aggregation Layer

Collapsed Aggregation/Core Layer

SRX5800

EX82XX

Access Layer

EX4200
Virtual Chassis

EX4500 EX82XX

Access Layer

Servers

NAS

FC Storage

Servers

NAS

FC Storage

FC SAN

FC SAN

Figure 18: Transitioning to a Juniper two-tier high-performance network

52

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Juniper Professional Services


Juniper Networks offers a suite of professional services that span the entire migration processfrom the design and planning to the implementation and operation phases of a new environment. These services provide an expeditious and thorough approach to getting the network up and running, while at the same time minimizing the risk of business disruptions and optimizing the networks performance. High- and low-level design services, as well as network consulting, help determine the high-level technical requirements to support business needs, including network assessment and recommendations. Implementation services offer a thorough review of the system configurations to be migrated, complete the actual migration/implementation tasks, and provide the necessary testing and troubleshooting activities to ensure a successful network implementation. Custom and fixed scope installation, quick start and transition services are available to ease the installation and enable a complete knowledge transfer of Juniper technology. These services also help accelerate the migration and minimize the risk and cost in moving from legacy products. Subsequent to the migration activities, Juniper consultant expertise can be obtained when and where it provides the most benefit during operation, to efficiently fill potential skill gaps and to help assess current network performance, providing recommendations to meet specific operational needs and maintain the network in the most optimal way. The full suite of Juniper Networks Professional Services offerings can be found at: www.juniper.net/us/en/productsservices/consulting-services.

Copyright 2010, Juniper Networks, Inc.

53

Data Center LAN Migration Guide

54

Copyright 2010, Juniper Networks, Inc.

Chapter 4: Troubleshooting

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Troubleshooting
Introduction
The scope of this section is to provide an overview of common issues that might be encountered at different insertion points when inserting Juniper platforms as a result of a trigger event (adding a new application or service to the organization). This section wont provide exhaustive troubleshooting details, however, we do describe the principal recommended approaches to troubleshooting the most common issues and provide guidelines for identification, isolation, and resolution.

Troubleshooting Overview
When investigating the root cause of a problem, it is important to determine the problems nature and analyze its symptoms. When troubleshooting a problem, it is generally advisable to start at the most general level and work progressively into the details, as needed. Using the OSI model as a reference, troubleshooting typically begins at the lower layers (physical and data link) and works progressively up toward the application layer until the problem is found. This approach tends to quickly identify what is working properly so that it can be eliminated from consideration, and narrows the problem domain for quick problem identification and resolution. The following list of questions provides a methodology on how to use clues and visible effects of a problem to reduce the diagnostic time. Has the issue appeared just after a migration, a deployment of new network equipment, a new link connection, or a configuration change? This is the context being presented in this Data Center LAN Migration Guide. The Method of Procedure (MOP) detailing the steps of the operation in question should include the tasks to be performed to return to the original state before the network event, should any abnormal conditions be identified. If any issue arises during or after the operation that cannot be resolved in a timely manner, it may be necessary to roll back and disconnect newly deployed equipment while the problem is researched and resolved. The decision to back out should be made well in advance, prior to the expiration of the maintenance window. This type of problem is likely due to an equipment misconfiguration or planning error. Does the problem have a local or a global impact on the network? The possible causes of a local problem may likely be found at L1 or L2, or it could be related to an Ethernet switching issue at the access layer. An IP routing problem may potentially have a global impact on networks, and the operator should focus its investigation on the aggregation and core layer of the network. Is it an intermittent problem? When troubleshooting an intermittent problem, system logging and traceoptions provide the primary debugging tools on Juniper Networks platforms, and can be focused on various protocol mechanisms at various levels of detail. Events occurring in the network will cause the logging of state transitions related to physical, logical, or protocols to local or remote files for analysis. Is it a total or partial loss of connectivity or is it a performance problem? All Juniper Networks platforms have a common architecture in that there are separate control and forwarding planes. For connectivity issues, Juniper recommends that you first focus on the control plane to verify routing and signaling states and then concentrate on the forwarding or data plane, which is implemented in the forwarding hardware (Packet Forwarding Engine or PFE). If network performance is adversely affected by packet loss, delays, and jitter impacting one or multiple traffic types, the root cause is most likely related to network congestion, high link utilization, and packet queuing along the traversed path.

56

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Hardware
The first action to take when troubleshooting a problem and also before making any change in the network is to ensure proper functionality and integrity of the network equipment and systems. A series of validation checks and inspection tests should be completed to verify that the hardware and the software operate properly and there are not any fault conditions. The following presents a list of show commands from the Junos OS CLI relative to this, as well as a brief description of expected outcomes. show system boot-messages Review the output and verify that no abnormal conditions or errors occurred during the booting process. POST (poweron self-test) results are captured in the bootup message log and stored on the hard drive. show chassis hardware detail Verify that all hardware appears in the output (i.e., routing engines, control boards, switch fabric boards, power supplies, line cards, and physical ports). Verify that no hardware indicates a failure condition. show chassis alarms Verify that there are no active alarms. show log messages Search log for errors and failures and review the log for any abnormal conditions. The search can be narrowed to specific keywords using the grep function. show system core-dumps Verify any transient software failures. Junos OS under fatal fault condition will create a core file of the kernel and processes in question for diagnostic analysis. For more details on platform specifics, please refer to the Juniper technical documentation that can be found at: www.juniper.net/techpubs.

OSI Layer 1: Physical Troubleshooting


An OSI Layer 1 problem or physical link failure can occur in any part of the network. Each media type has different physical and logical properties and provides different diagnostic capabilities. Focus here will be on Ethernet, as it is universally deployed in data centers at all tiers and in multiple flavors: GbE, 10GbE, copper, fiber, etc. show interface extensive command produces the most detailed and complete information about all interfaces. It displays input and output errors for the interface displayed in multiple categories such as carrier transition, cyclic redundancy check (CRC) errors, L3 incomplete errors, policed discard, L2 channel errors, static RAM (SRAM) errors, packet drops, etc. It also contains interface status and setup information at both physical and logical layers. Ethernet networks can present many symptoms, but troubleshooting can be helped by applying common principles: verify media type, speed, fiber mode and length, interface and protocol maximum transmission unit (MTU), flow control and link mode. The physical interface may have a link status of up because the physical link is operational with no active alarm, but the logical interface has a link status of down because the data link layer cannot be established end to end. If this occurs, refer to the next command. monitor interface provides real-time packets and byte counters as well as displaying error and alarm conditions. After an equipment migration or a new link activation, the network operator should ping a locally connected host to verify that the link and interface are operating correctly and monitor if there are any incrementing error counters. The do-not-fragment flag in a ping test is a good tool to detect MTU problems which can adversely affect end-to-end communication. For 802.3ad aggregated Ethernet interfaces, we recommend enabling Link Aggregation Control Protocol (LACP) as a dynamic bundling protocol to form one logical interface with multiple physical interfaces. LACP is designed to provide link monitoring capabilities and fast failure detection over an Ethernet bundle connection.

Copyright 2010, Juniper Networks, Inc.

57

Data Center LAN Migration Guide

OSI Layer 2: Data Link Troubleshooting


Below are some common steps to assist in troubleshooting issues at Layer 2 in the access and aggregation tiers: Are the devices utilizing DHCP to obtain an IP addresses? Is the Dynamic Host Configuration Protocol (DHCP) server functioning properly so that host devices receive an IP address assignment from the DHCP server? If routed, is the DHCP request being correctly forwarded? monitor traffic interface ge-0/0/0 command provides a tool for monitoring local traffic. Expect to see all packets that are sent out and received to and from ge-0/0/0. This is particularly useful to verify the Address Resolution Protocol (ARP) process over the connected LAN or VLAN. Use the show arp command to display ARP entries. Is the VLAN in question active on the switch? Is a trunk active on the switch that could interfere with the ability to communicate? Is the routed VLAN interface (RVI) configured with the correct prefix and attached to the corresponding VLAN? Is VRRP functioning properly and showing one unique routing node as master for the virtual IP (VIP) address? Virtual Chassis, Layer 3 uplinks, inverted U designs, and VPLS offer different alternatives to prevent L2 data forwarding loops in a switching infrastructure without the need to implement Spanning Tree Protocols (STPs). Nevertheless, it is common best practice to enable STP as a protection mechanism to prevent broadcast storms in the event of a switch misconfiguration or a connection being established by accident between two access switches.

Virtual Chassis Troubleshooting


Configuring a Virtual Chassis is essentially plug and play. However, if there are connectivity issues, the following section provides the relevant commands to perform operational analysis and troubleshooting. To troubleshoot the configuration of a Virtual Chassis, perform the following steps. Check and confirm Virtual Chassis configuration and status with the following commands: show configuration virtual-chassis show virtual-chassis member-config all-members show virtual-chassis status Check and confirm Virtual Chassis interfaces: show interfaces terse show interfaces terse vcp* show interfaces terse *me* Verify that the mastership priority is assigned appropriately: show virtual-chassis status show virtual-chassis vc-port all-members Verify the Virtual Chassis active topology and neighbors: show virtual-chassis active-topology show virtual-chassis protocol adjacency show virtual-chassis protocol database extensive show virtual-chassis protocol route show virtual-chassis protocol statistics In addition to the verifications above, also check the following: Check the cable to make sure that it is properly and securely connected to the ports. If the Virtual Chassis port (VCP) is an uplink port, make sure that the uplink module is model EX-UM-2XFP. If the VCP is an uplink port, make sure that the uplink port has been explicitly set as a VCP. If the VCP is an uplink port, make sure that you have specified the options (pic-slot, port-number, member-id) correctly.

58

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

OSI Layer 3: Network Troubleshooting


While L1 and L2 problems have limited effect and are local, L3 or routing issues may affect other networks by propagation and may have a global impact. In the data center, the aggregation/core tiers may be affected. The following focuses on the operation of OSPF and BGP as they are commonly implemented in data centers to exchange internal and external routing information. In a next -generation or newly deployed network, OSPFs primary responsibility typically is to discover endpoints for internal BGP. Unlike OSPF, BGP may play multiple roles that include providing connectivity to an external network, information exchange between VRFs for L3 MPLS VPN or VPLS, eventually carrying data centers internal routes to access routers.

OSPF
A common problem in OSPF is troubleshooting adjacency issues which can occur for multiple reasons: mismatched IP subnet/mask, area number, area type, authentication, hello/dead interval, network type, or mismatched IP MTU. The following are useful commands for troubleshooting an OSPF problem: show ospf neighbor displays information about OSPF neighbors and the state of the adjacencies which must be shown as full. show ospf interface displays information about the status of OSPF interfaces. show ospf log logs shortest-path-first (SPF) calculation. show ospf statistics displays number and type of OSPF packets sent and received. show ospf databases displays entries in the OSPF link-state database (LSDB). OSPF traceoptions provide the primary debugging tool, and the OSPF operation can be flagged to log error packets and state transitions along with the events causing them.

BGP
show bgp summary is a primary command used to verify the state of BGP peer sessions, and it should display that the peering is established to be fully operational. BGP has multiprotocol capabilities made possible through simple extensions that add new address families. This command also helps to verify which address families are carried over the BGP session, for example, inet-vpn if L3 MPLS VPN service is required or L2VPN for VPLS. BGP is a policy-driven routing protocol. It offers flexibility and granularity when implementing routing policy for path determination and for prefix filtering. A network operator must be familiar with the rich set of attributes that can be modified and also with the BGP route selection process. Routing policy controls and filters can modify routing information entering or leaving the router in order to alter forwarding and routing decisions based on the following criteria: What should be learned about the network from all protocols? What routes should be shared with other routing protocols? What should be advertised to other routers? What routing information should be modified, if any? Consistent policies must be applied across the entire network to filter/advertise routes and modify BGP route attributes. The following commands assist in the troubleshooting of routing policies: show route receive-protocol bgp <neighbor> displays received attributes. show route advertising-protocol bgp <neighbor> displays route and attributes sent by BGP to a specific peer. show route hidden extensive displays routes not usable due to BGP next-hop problems and routes filtered by an inbound route filter. Logging of peer state transitions and flagging BGP operations provides a good source of information when investigating BGP problems.

Copyright 2010, Juniper Networks, Inc.

59

Data Center LAN Migration Guide

VPLS Troubleshooting
This section provides a logical approach to take when determining the root cause of a problem in a VPLS network. A good place to start is to verify the configuration setup. Following is a brief configuration snippet with corresponding descriptor.

routing-instance { vpls_vpn1 { instance-type vpls; vlan-tags outer 4094 inner 4093; interface ge-1/0/0.3001; route-distinguisher 65000:1001; vrf-target target:65000:1001; protocols { vpls { mac-table-size { 100; } interface-mac-limit {

#arbitrary name #VPLS type #VLAN Normalization must match if configured # int.unit # RD carried in MPGP # VPN RT must match on all PEs in this VPLS

# max mac table size # max mac that may be learned from all CE facing interfaces # lsi interfaces for tunneling # arbitrary name # unique site ID # list of int.unit in this VPN

50; } no-tunnel-services; site site-1 { site-identifier 1001; interface ge-1/0/0.3001;

The next step is to verify the control plane with the following operation commands: show route receive-protocol bgp <neighbor> table <vpls-vpn> detail displays BGP routes received from an MPiBGP peer for a VPLS instance. Use the detail/extensive option to see other BGP attributes such as route-target RT, label base, and site-ID. The BGP next hop must have a route in the routing table for the mapping to a transport MPLS LSP. show vpls connections is an excellent command to verify the VPLS connection status and to aid in troubleshooting. After the control plane has been validated as fully functional, the forwarding plane should be checked next by issuing the following commands. Note that the naming of devices maps to a private MPLS network, as opposed to using a service provider MPLS network. On local switch: show arp show interfaces ge-0/0/0 On MPLS edge router show vpls mac-table show route forwarding-table On MPLS core router show route table mpls

60

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

The commands presented in this section should highlight the proper VPLS operation as follows: Sending to unknown MAC address, VPLS edge router floods to all members of the VPLS. Sending to a known MAC address, VPLS edge router maps to an outer and inner label. Receiving a MAC address, VPLS edge router identifies the sender and maps the MAC address to a label stack in the MAC address cache. VPLS provider edge (PE) router periodically ages out unused entries from the MAC address cache.

Multicast
Looked at simplistically, multicast routing is upside down unicast routing. Multicast routing functionality is focused on where the packet came from and directs traffic away from its source. When troubleshooting multicast, the following methodology is recommended: Gather information In one-to-many and many-to-many communications, it is important to have a good understanding of the expected traffic flow to clearly identify all sources and receivers for a particular multicast group. Verify receiver interest by issuing the following commands: show igmp group <mc_group> displays information about Internet Group Management Protocol (IGMP) group membership received from the multicast receivers on the LAN interface. show pim interfaces is used to verify the designated router for that interface or VLAN. Verify knowledge of the active source by issuing the following commands: show multicast route group <mc_group> source-prefix <ip_address> extensive displays the forwarding state (pruned or forwarding) and the rate for this multicast route. show pim rps extensive determines if the source designated router has the right rendezvous point (RP) and displays tunnel interface-related information for register message for encapsulation/de-encapsulation. Trace the forwarding state backwards, working your way back towards the source IP and looking for Physical Interface Module (PIM) problems along the way with the following commands: show pim neighbors displays information about PIM neighbors. Show pim join extensive <mc_group> validates outgoing interface list and upstream neighbor and displays source tree and shared tree (Real-Time Transport Protocol or RTP) state, with join/prune status. Show multicast route group <mc_group> source-prefix <ip_address> produces extensive checks if traffic is flowing and has a positive traffic rate. show multicast rpf <source_address>Multicast routing uses reverse path forwarding check. A router forwards only multicast packets if received on the upstream interface to the source. Otherwise, the RPF check fails, and the packet is discarded.

Quality of Service/Class of Service (CoS)


Link congestion in the network can be the root cause of packet drops. The show interfaces queue command provides CoS queue statistics for all physical interfaces to assist in determining the number of packets dropped due to tail drop, and the number of packets dropped due to random early detection (RED).

Copyright 2010, Juniper Networks, Inc.

61

Data Center LAN Migration Guide

OSI Layer 4-7: Transport to Application Troubleshooting


This type of problem is most likely to occur on firewalls or on routers secured with firewall filters. Below are some important things to remember when troubleshooting Layer 4-7 issues: Standard troubleshooting tools such as ping and traceroute may not work. Generally, ping and traceroute are not enabled through a firewall except in specific circumstances. Firewalls are routers too, In addition to enforcing stateful policies on traffic, firewalls also have the responsibility of routing packets to their next hop. To do this, firewalls must have a working and complete routing table statically or dynamically defined. If the table is incomplete or incorrect, the firewall will not be able to forward traffic correctly. Firewalls are stateful and build state for every session that has passed through the firewall. If a non-SYN packet comes to the firewall and the firewall does not have a session open for that packet, it is considered an out of state packet. This can be the sign of an attack or an application that is dormant beyond the firewall session timeout duration attempting to send traffic. By definition, stateful firewalls enforce traffic though their policy based on the network and transport layers of the OSI model. In addition, firewalls may also do protocol anomaly checks and signature matches on the application layer for selected protocols. This function is implemented by ALGs. ALGs recognize application-specific sequences, change the application layer to make protocols compatible with Port Address Translation (PAT) attempting to send traffic and Network Address Translation (NAT), and deliver higher layer content to deep inspection (DI), antivirus, URL filter, and spam filter features, if enabled. If you experience a problem that involves the passing or blocking of traffic, the very first place to look is the firewall logs. Often the log messages will give strong hints about the problem.

Tools
Junos OS has embedded script tools to simplify and automate some tasks for network engineers. Commit scripts, operation (op) scripts, and event scripts provide self monitoring, self diagnosing, and self healing capabilities to the network. The apply-macro command feeds a commit script to extend and customize the router configuration based on user-defined data and templates. Together, these tools offer an almost infinite number of applications to reduce downtime, minimize human error, accelerate service deployment, and reduce overall operational costs. For more information, refer to: www.juniper.net/us/en/community/junos/script-automation.

Troubleshooting Summary
Presenting an exhaustive and complete troubleshooting guide falls outside the scope of this Data Center LAN Migration Guide. Presented in this section is a methodology to understand the factors contributing to a problem and a logical approach to the diagnostics needed to investigate root causes. This method relies on the fact that IP networks are modeled around multiple layered architectures. Each layer depends on the services of the underlying layers. From the physical network topology comprised of access, aggregation, and core tiers to the model of IP communication founded on the 7 OSI layers, matching symptoms to the root cause layer is a critical step in the troubleshooting methodology. Juniper platforms have also implemented a layered architecture by integrating separate control and forwarding planes. Once the root cause layer is correctly identified, the next steps are to isolate the problem and to take the needed corrective action at that specific layer. For more details on platform specifics, please refer to the Juniper technical documentation that can be found at: www.juniper.net/techpubs.

62

Copyright 2010, Juniper Networks, Inc.

Chapter 5: Summary and Additional Resources

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Summary
Todays data center is a vital component to business success in virtually all industries and markets. New trends and technologies such as cloud computing and SOA-based applications have significantly altered data center traffic flows and performance requirements. Data center designs based on the performance characteristics of older technology and traffic flows have resulted in complex architectures which do not easily scale and also may not provide the performance needed to efficiently meet todays business objectives. A simplified, next-generation, cloud-ready two-tier data center design is needed to meet these new challenges without compromising performance or security. Since most enterprises wont disrupt a production data center except for scheduled maintenance and business continuity testing, a gradual migration to the new network is often most practical. This guide has identified and described how migration towards a simpler two-tier design can begin at any time and at various insertion points within an existing legacy data center architecture. These insertion points are determined by related trigger events such as adding a new application or service, or by larger events such as data center consolidation. This guide has outlined the design considerations for migration at each network layer, providing organizations with a path towards a simplified high-performance network which can not only lower TCO, but also provide the agility and efficiency to enable organizations to gain a competitive advantage by leveraging their data center network. Juniper Networks has been delivering a steady stream of network innovations for more than a decade. Juniper brings this innovation to a simplified data center LAN solution built on three core principles: simplify, share, and secure. Creating a simplified infrastructure with shared resources and secure services delivers significant advantages over other designs. It helps lower costs, increase efficiency, and keep the data center agile enough to accommodate any future business changes or technology infrastructure requirements.

Additional Resources
Data Center Design Resources
Juniper has developed a variety of materials that complement this Migration Guide and are helpful to the network design process in multiple types of customer environments and data center infrastructures. On our public Internet website, we keep many of these materials at the following locations: 1. Data Center Solution Reference Materials Site: www.juniper.net/us/en/solutions/enterprise/data-center/simplify/#literature At this location you will find information helpful to the design process at a number of levels, organized by selectable tabs according to the type of information you are seekinganalyst reports, solution brochures, case studies, reference architecture, design guides, implementation guides, and industry reports. The difference between a reference architecture, design guide, and implementation guide is the level of detail the document addresses. The reference architecture is the highest level organization of our data center network approach; the design guide provides guidance at intermediate levels of detail appropriate to, say, insertion point considerations (in the terms of this Migration Guide); and implementation guides provide specific guidance for important types of network deployments at different tiers of the data center network. Implementation guides give customers and other readers enough information to start specific product implementation tasks appropriate for the most common deployment scenarios, and are quite usable in combination with an individual products installation, configuration, or operations manual. 2. Information on Junipers individual lines relevant to the insertion scenarios described in this guide can be found at: Ethernet switching: www.juniper.net/us/en/products-services/switching IP routing: www.juniper.net/us/en/products-services/routing Network security: www.juniper.net/us/en/products-services/security Network management: http://www.juniper.net/us/en/products-services/software/junos-platform/junos-space

64

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

3. Information on the way Junipers offerings fit with various alliance partners in the data center environment can be found at: www.juniper.net/us/en/company/partners/enterprise-alliances. 4. Information on Junipers Professional Services to support planning and design of migration projects can be found at: www.juniper.net/us/en/products-services/consulting-services/#services. For more detailed discussions of individual projects and requirements, please contact your Juniper or authorized Juniper partner representative directly.

Training Resources
Juniper Networks offers a rich curriculum of introductory and advanced courses on all of its products and solutions. Learn more about Junipers free and fee-based online and instructor-led hands-on training offerings: www.juniper.net/us/en/training/technical_education.

Juniper Networks Professional Services


Juniper Networks Professional Services organization is uniquely qualified to help enterprises or service providers design, implement, and optimize their networks for confident operation and rapid returns on infrastructure investments. The Professional Services team understands todays Internet demands and those that are just around the corner for bandwidth efficiency, best-in-class security, solid reliability, and cost-effective scaling. These highly trained, experienced professionals augment your team to keep your established network protected, up-to-date, performing at its best, and aligned with the goals of your organization. Planning and design consultingEnsures that the new design is aligned with your stated business and technical goals. Incorporates critical business information in the network re-architecture. Eliminates costly redesigns and lost time. Implementation and configuration servicesHelps you achieve the deployment of complex products and technologies efficiently with minimal business disruption. These services shorten the interval between network element installation and revenue generation. This may includes services such as onsite and remote support for installation, configuration, and testing. It may also include production of site requirements, site survey documentation, hardware installation documentation, and network test documentation. Knowledge transferJuniper Networks IP experts can train internal teams in best-in-class migration practices or any specific enterprise issues. Project managementA dedicated project manager can be assigned to assist with administration and management throughout the entire migration project. In general, this service provides an emphasis on tasks such as project planning, reports that track project tasks against scheduled due dates, and project documentation. Resident engineerA Juniper Networks Resident Engineer (RE) can be placed onsite at any desired enterprise location to engage with the engineering or operations staff on a daily basis to support a data center network. Functioning as part of the enterprise team, REs are available for a 12-month engagement to the specific networking environment, and provide technical assistance such as network implementation and migration, troubleshooting and operations support, network and configuration analysis, assistance in testing Juniper product features and functionality to help optimize the value of high-performance networking to meet an evolving business environment. Learn more about Juniper Networks Professional Services at: www.juniper.net/us/en/products-services/consultingservices. Juniper Networks Technical Assistance Center (JTAC) provides a single point of contact for all support needs with skilled engineers and automatic escalation alerts to senior management. The Customer Support Center (CSC) provides instant, secure access to critical information, including the Juniper Networks Knowledge Base, frequently asked questions, proactive technical bulletins, problem reports, technical notes, release notes, and product documentation.

Copyright 2010, Juniper Networks, Inc.

65

Data Center LAN Migration Guide

About Juniper Networks


Juniper Networks, Inc. is the leader in high-performance networking. Juniper offers a high-performance network infrastructure that creates a responsive and trusted environment for accelerating the deployment of services and applications over a single network. This fuels high-performance business. Additional information can be found at www.juniper.net.

66

Copyright 2010, Juniper Networks, Inc.

Data Center LAN Migration Guide

Copyright 2010, Juniper Networks, Inc.

67

Data Center LAN Migration Guide

Corporate and Sales Headquarters Juniper Networks, Inc. 1194 North Mathilda Avenue Sunnyvale, CA 94089 USA Phone: 888.JUNIPER (888.586.4737) or 408.745.2000 Fax: 408.745.2100 www.juniper.net

APAC Headquarters Juniper Networks (Hong Kong) 26/F, Cityplaza One 1111 Kings Road Taikoo Shing, Hong Kong Phone: 852.2332.3636 Fax: 852.2574.7803

EMEA Headquarters Juniper Networks Ireland Airside Business Park Swords, County Dublin, Ireland Phone: 35.31.8903.600 EMEA Sales: 00800.4586.4737 Fax: 35.31.8903.601

To purchase Juniper Networks solutions, please contact your Juniper Networks representative at 1-866-298-6428 or authorized reseller.

Copyright 2010 Juniper Networks, Inc. All rights reserved. Juniper Networks, the Juniper Networks logo, Junos, NetScreen, and ScreenOS are registered trademarks of Juniper Networks, Inc. in the United States and other countries. All other trademarks, service marks, registered marks, or registered service marks are the property of their respective owners. Juniper Networks assumes no responsibility for any inaccuracies in this document. Juniper Networks reserves the right to change, modify, transfer, or otherwise revise this publication without notice.

7100128-001-EN

Aug 2010

Printed on recycled paper

68

Copyright 2009, Juniper Networks, Inc.

Вам также может понравиться