IBM Network Design Guide
IBM Network Design Guide
IBM Data Center Networking: Planning for virtualization and cloud computing
Drivers for change in the data center
IBM systems networking capabilities The new data center design landscape
ibm.com/redbooks
7928edno.fm
International Technical Support Organization IBM Data Center Networking: Planning for virtualization and cloud computing January 2011
SG24-7928
7928edno.fm
Note: Before using this information and the product it supports, read the information in Notices on page vii.
First Edition (January 2011) This document created or updated on February 10, 2011.
Copyright International Business Machines Corporation 2011. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix The team who wrote this book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .x Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xii Chapter 1. Drivers for a dynamic infrastructure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.1 Key operational challenges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.1.1 Costs and service delivery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.1.2 Energy efficiency. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 1.1.3 Business resiliency and security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 1.1.4 Changing applications and business models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 1.1.5 Harnessing new technologies to support the business . . . . . . . . . . . . . . . . . . . . . . 6 1.1.6 Evolving business models. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 1.2 Cloud computing can change how IT supports business . . . . . . . . . . . . . . . . . . . . . . . . 7 1.2.1 The spectrum of cloud solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 1.3 Benefits and challenges of cloud computing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 1.4 Perceived barriers to cloud computing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 1.5 Implications for todays CIO . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 1.6 Dynamic infrastructure business goals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 1.6.1 Reduce cost . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 1.6.2 Improve service. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 1.6.3 Manage risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 1.7 Enabling the dynamic infrastructure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 1.8 The shift in the data center network architectural thinking . . . . . . . . . . . . . . . . . . . . . . 20 1.8.1 Networking non-functional requirements. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 Chapter 2. Servers, storage and software components . . . . . . . . . . . . . . . . . . . . . . . . 2.1 Virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.1 Server virtualization techniques . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.2 Storage virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 The System z platform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.1 Introduction to System z architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2 Mainframe virtualization technologies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.3 Linux on System z. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.4 z/VM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.5 System z network connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.6 z/OS Communications Server for IP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3 Power Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.1 PowerVM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.2 PowerVM Editions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.3 POWER Hypervisor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.4 Live Partition Mobility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.5 Virtual I/O Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.6 Virtual Ethernet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 33 33 37 38 39 40 45 47 51 55 59 60 60 61 64 64 65
2.3.8 External connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 2.3.9 IPv4 routing features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70 2.4 System x and BladeCenter virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 2.4.1 Benefits of virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72 2.4.2 Hardware support for full virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73 2.4.3 IBM BladeCenter. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74 2.5 Other x86 virtualization software offerings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 2.5.1 Xen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78 2.5.2 KVM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80 2.5.3 VMware vSphere 4.1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83 2.5.4 Hyper-V R2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 2.6 Storage virtualization. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104 2.6.1 Storage Area Networks (SAN) and SAN Volume Controller (SVC) . . . . . . . . . . 105 2.6.2 Virtualization Engine TS7520: virtualization for open systems . . . . . . . . . . . . . . 108 2.6.3 Virtualization Engine TS7700: mainframe virtual tape . . . . . . . . . . . . . . . . . . . . 109 2.6.4 XIV Enterprise Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 2.6.5 IBM Disk Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110 2.6.6 Storwize V7000 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110 2.6.7 Network Attached Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111 2.6.8 N-Series . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112 2.7 Virtualized infrastructure management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112 2.7.1 IT Service Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114 2.7.2 Event Management - Netcool/OMNIbus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118 2.7.3 IBM Tivoli Provisioning Manager. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121 2.7.4 Systems and Network management - IBM Systems Director . . . . . . . . . . . . . . . 124 2.7.5 Network management - IBM Tivoli Network Manager. . . . . . . . . . . . . . . . . . . . . 128 2.7.6 Network configuration change management - Tivoli Netcool Configuration Manager 132 2.7.7 System and network management product integration scenarios . . . . . . . . . . . 136 2.8 IBM integrated data center solutions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139 2.8.1 iDataplex . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140 2.8.2 CloudBurst . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146 Chapter 3. Data center network functional components. . . . . . . . . . . . . . . . . . . . . . . 3.1 Network virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Impact of server and storage virtualization trends on the data center network . . . . . . 3.3 Impact of Data Center Consolidation on the Data Center Network. . . . . . . . . . . . . . . 3.4 Virtualization technologies for the Data Center Network. . . . . . . . . . . . . . . . . . . . . . . 3.4.1 Data Center Network Access Switching Techniques . . . . . . . . . . . . . . . . . . . . . 3.4.2 Traffic patterns at the access layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4.3 Network Node virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4.4 Building a Single Distributed Data Center. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4.5 Network Services Deployment Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4.6 Virtual Network Security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4.7 Virtualized network resources in servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 4. The new data center design landscape . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1 The changing IT landscape and data center networking . . . . . . . . . . . . . . . . . . . . . . 4.1.1 Increasing importance of the data center network . . . . . . . . . . . . . . . . . . . . . . . 4.1.2 Multi-vendor networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.3 Networks - essential to the success of cloud computing initiatives. . . . . . . . . . . 4.1.4 IPv6 and the data center . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149 150 152 157 162 162 163 165 168 175 177 184 193 195 196 197 197 199
4.2 Assuring service delivery : The data center network point of view . . . . . . . . . . . . . . . 4.2.1 Todays data center network challenges. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.2 Burgeoning solutions for the data center network. . . . . . . . . . . . . . . . . . . . . . . . 4.2.3 Additional points of concern . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3 Setting the evolutionary imperatives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3.1 Developing a strategic approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3.2 The value of standardization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3.3 Service Delivery Oriented DCN Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.4 IBM network strategy, assessment, optimization and integration services . . . . . . . . . 4.4.1 Services lifecycle approach. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.4.2 Networking services methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.4.3 Data center network architecture and design . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.4.4 Why IBM? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Related publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Online resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . How to get Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Help from IBM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
201 202 208 210 212 213 215 217 220 222 224 225 226 229 229 230 231 231
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233
Notices
This information was developed for products and services offered in the U.S.A. IBM may not offer the products, services, or features discussed in this document in other countries. Consult your local IBM representative for information on the products and services currently available in your area. Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to evaluate and verify the operation of any non-IBM product, program, or service. IBM may have patents or pending patent applications covering subject matter described in this document. The furnishing of this document does not give you any license to these patents. You can send license inquiries, in writing, to: IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A. The following paragraph does not apply to the United Kingdom or any other country where such provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you. This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice. Any references in this information to non-IBM Web sites are provided for convenience only and do not in any manner serve as an endorsement of those Web sites. The materials at those Web sites are not part of the materials for this IBM product and use of those Web sites is at your own risk. IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you. Information concerning non-IBM products was obtained from the suppliers of those products, their published announcements or other publicly available sources. IBM has not tested those products and cannot confirm the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. This information contains examples of data and reports used in daily business operations. To illustrate them as completely as possible, the examples include the names of individuals, companies, brands, and products. All of these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is entirely coincidental. COPYRIGHT LICENSE: This information contains sample application programs in source language, which illustrate programming techniques on various operating platforms. You may copy, modify, and distribute these sample programs in any form without payment to IBM, for the purposes of developing, using, marketing or distributing application programs conforming to the application programming interface for the operating platform for which the sample programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or imply reliability, serviceability, or function of these programs.
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines Corporation in the United States, other countries, or both. These and other IBM trademarked terms are marked on their first occurrence in this information with the appropriate symbol ( or ), indicating US registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law trademarks in other countries. A current list of IBM trademarks is available on the Web at http://www.ibm.com/legal/copytrade.shtml The following terms are trademarks of the International Business Machines Corporation in the United States, other countries, or both:
AIX BladeCenter CloudBurst DB2 DS6000 DS8000 Dynamic Infrastructure ESCON FICON GDPS Geographically Dispersed Parallel Sysplex Global Business Services HiperSockets IBM iDataPlex Informix Maximo Micro-Partitioning MVS Netcool Parallel Sysplex POWER Hypervisor Power Systems POWER5 POWER6 POWER7 PowerPC PowerVM POWER PR/SM Processor Resource/Systems Manager Proventia pSeries RACF Redbooks Redguides Redbooks (logo) Service Request Manager Solid System i System p System Storage System x System z Systems Director VMControl Tivoli TotalStorage Virtual Patch X-Architecture X-Force XIV z/Architecture z/OS z/VM z/VSE z10
The following terms are trademarks of other companies: ITIL is a registered trademark, and a registered community trademark of the Office of Government Commerce, and is registered in the U.S. Patent and Trademark Office. IT Infrastructure Library is a registered trademark of the Central Computer and Telecommunications Agency which is now part of the Office of Government Commerce. Data ONTAP, and the NetApp logo are trademarks or registered trademarks of NetApp, Inc. in the U.S. and other countries. Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. Intel, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries. UNIX is a registered trademark of The Open Group in the United States and other countries. Linux is a trademark of Linus Torvalds in the United States, other countries, or both. Other company, product, or service names may be trademarks or service marks of others.
7928pref.fm
Preface
The enterprise data center has evolved dramatically in recent years. It has moved from a model that placed multiple data centers closer to end users to a more centralized dynamic model. The factors influencing this evolution are varied but can mostly be attributed to regulatory, service level improvement, cost savings, and manageability. Multiple legal issues regarding the security of data housed in the data center have placed security requirements at the forefront of data center architecture. As the cost to operate data centers has increased, architectures have moved towards consolidation of servers and applications in order to better utilize assets and reduce server sprawl. The more diverse and distributed the data center environment becomes, the more manageability increasingly becomes an issue. These factors have led to a trend of data center consolidation and resources on demand using technologies such as virtualization, higher WAN bandwidth technologies, and newer management technologies. The network has been widely viewed as the plumbing of the system. It has been architected without much if any consideration for the type of device at the end point. The usual consideration for the end-point requirements consisted of speed, duplex, and possibly some traffic engineering technology. There have been well-documented designs that allowed redundancy and availability of access ports that were considered interchangeable at best and indistinguishable at worst. With the rise of highly virtualized environments and the drive towards dynamic infrastructures and cloud computing, the network can no longer remain just plumbing. It must be designed to become dynamic itself. It will have to provide the ability to interconnect both the physical and virtual infrastructure of the new enterprise data center and provide for cutting edge features such as workload mobility that will drive enterprise IT architecture for years to come. In this book we discuss: The current state of the data center network The business drivers making the case for change The unique capabilities and network requirements of system platforms The impact of server and storage consolidation on the data center network The functional overview of the main data center network virtualization and consolidation technologies The new data center network design landscape
ix
7928pref.fm
7928pref.fm
Jeffery Sanden Global Technology Services, Executive IT Architect, Global Integrated Communications Services CTO team Alan Fishman, Stephen Sauer, Iain Neville, Joe Welsh, and Dave Johnson and Wes Toman IBM Special thanks to the Data Center Networking Center of Excellence: Miguel Cachim Gino DePinto Roberta J Flinn Patrick Frejborg Hua He Mikio Itaba Raymund Schuenke
Comments welcome
Your comments are important to us! We want our books to be as helpful as possible. Send us your comments about this book or other IBM Redbooks in one of the following ways: Use the online Contact us review Redbooks form found at: ibm.com/redbooks
Preface
xi
7928pref.fm
Send your comments in an e-mail to: redbooks@us.ibm.com Mail your comments to: IBM Corporation, International Technical Support Organization Dept. HYTD Mail Station P099 2455 South Road Poughkeepsie, NY 12601-5400
xii
7928ch01.fm
Chapter 1.
7928ch01.fm
A more integrated IT approach is needed. IT service delivery needs to be changed to help move beyond todays operational challenges to a data center model that is more efficient, service-oriented, and responsive to business needs, with improved levels of economies, rapid service delivery, and one that can provide tighter alignment with business goals. IBM uses an evolutionary approach for efficient IT delivery that helps to drive business innovation. This approach allows organizations to be better positioned to adopt integrated new technologies, such as virtualization and cloud computing, to help deliver dynamic and seamless access to IT services and resources. As a result, IT departments will spend less time fixing IT problems and more time solving real business challenges. With this approach to efficient IT service delivery, IBM captures an end-to-end view of the IT data center and its key components. As such, the strategy for planning and building the dynamic infrastructure integrates the following elements: Highly virtualized resources Efficient, green, and optimized infrastructures and facilities Business-driven service management Business resiliency and security Information infrastructure In this chapter, we examine the operational challenges leading to an approach to implement a dynamic infrastructure.
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
The ability to deliver quality service is critical to businesses of all sizes. Service management enables visibility, control, and automation to deliver quality service at any scale. Maintaining stakeholder satisfaction through cost efficiency and rapid return on investment depends upon the ability to see the business (visibility), manage the business (control), and leverage automation (automate) to drive efficiency and operational agility. High responsiveness and business goal-driven infrastructure A highly efficient, shared infrastructure can help businesses respond quickly to evolving demands. It creates opportunities to make sound business decisions based on information obtained in real time. Alignment with a service-oriented approach to IT delivery provides the framework to free up resources from more traditional operational demands and to focus them on real-time integration of transactions, information, and business analytics. IT professionals have continued to express concerns about the magnitude of the operational issues they face. These operational issues are described in the following sections.
Virtualization 2.0: The Next Phase in Customer Adoption. Doc. 204904 IDC, Dec. 2006 - from the IBM's Vision for the New Enterprise Data Center white paper http://www-05.ibm.com/innovation/nl/shapeyourfuture/pdf/New_Enterprise_Data_Center.pdf IBM Information Infrastructure Newsletter, March 2009 http://www-05.ibm.com/il/systems/storage/pdf/information_infastructure_custnews1Q.pdf
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
Virtualization 2.0: The Next Phase in Customer Adoption. Doc. 204904 IDC, Dec. 2006 - from the IBM's Vision for the New Enterprise Data Center white paper http://www-05.ibm.com/innovation/nl/shapeyourfuture/pdf/New_Enterprise_Data_Center.pdf CIO Magazine, Survey 2007 as cited in ftp://public.dhe.ibm.com/software/uk/itsolutions/optimiseit/energy-efficiency-solutions /1-new-enterprise-data-centre-ibm-s-vision-nedc.pdf http://www.internetworldstats.com/stats.htm
7928ch01.fm
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
Furthermore, mobile devices now give us the ability to transport information, or access it online, nearly anywhere. Today, the people at the heart of this technology acceleration, Generation Y, cannot imagine a world without the Internet. They are the ones entering the workforce in droves, and they are a highly attractive and sought-after consumer segment. Because of this, business models are no longer limited to business-to-business or business-to-consumer. Instead, these new generations of technologyand peoplehave created a bi-directional business model that spreads influential content from consumer-to-consumer to communities-to-business. Today, the power of information, and the sharing of that information, rests firmly in the hands of the end user while real-time data tracking and integration will become the norm.
Location-independent resource pooling Rapid elasticity and provisioning Pay-per-use Cloud has evolved from on demand and grid computing, while building on significant advances in virtualization, networking, provisioning and multi-tenant architectures. As with any new technology, the exciting impact comes from enabling new service consumption and delivery models that support business model innovation.
10
Youseff, Lamia: Toward a Unified Ontology of Cloud Computing, November 2008, Available from: http://www.cs.ucsb.edu/~lyouseff/CCOntology/CloudOntology.pdf
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
The distinction between the five categories of cloud offering is not necessarily clear-cut. In particular, the transition from Infrastructure as a Service to Platform as a Service is a very gradual one, as shown in Figure 1-2.
Vendor manages the (virtual) hardware (Like having your own data center but with infinitely flexible capacity) Vendor provides pre-configured OS (Customers then have to manage it themselves.) Vendor manages the OS (Customers can install any stack they want, but have to manage it and pay for it themselves.) Vendor manages the framework and/or database (Applications can be dropped into a fully-managed environment) Vendor hides the framework and/or database (Applications have to be (re)written in the vendors environment) IaaS GridLayer FlexiScale Joyent Amazon EC2 GoGrid Mosso IBM Private Clouds Google AppEngine Force.com
PaaS
7928ch01.fm
10
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
Challenges Tight binding of the applications with the platform makes portability across vendors extremely difficult. PaaS in general is still maturing, and the full benefits of componentization of, and collaboration between, services is still to be demonstrated. PaaS offerings lack the functionality needed for converting legacy applications into fully-fledged cloud services.
Benefits PaaS PaaS saves costs by: Reducing up-front software licensing and infrastructure costs Reducing ongoing operational costs for development, test and hosting environments. PaaS significantly improves development productivity by removing the challenges of integration with services such as database, middleware, Web frameworks, security, and virtualization. Software development and delivery times are shortened since software development and testing are performed on a single PaaS platform. There is no need to maintain separate development and test environments. PaaS fosters collaboration among developers and also simplifies software project management. This is especially beneficial to enterprises that have outsourced their software development. IaaS saves costs by: Eliminating the need to over-provision computing resources to be able to handle peaks in demand. Resources dynamically scale up and down as required. Reducing capital expenditure on infrastructure, and ongoing operational costs for support, maintenance and administration. Organizations can massively increase their data center resources without significantly increasing the number of people needed to support it. The time required to provision new infrastructure resources is reduced from typically months to just minutes - the time required to add the requirements to an online shopping cart, submit it and have it approved. IaaS platforms are generally open platforms, supporting a wide range of operating systems and frameworks. This minimizes the risk of vendor lock-in.
IaaS
Infrastructure resources are leased on a pay-as-you-go basis, according to the hours of usage. Applications that need to run 24x7 may not be cost-effective. To benefit from the dynamic scaling capabilities, applications have to be designed to scale and execute on the vendor's infrastructure. There can be integration challenges with third-party software packages. This should improve over time, however, as and when ISVs adopt cloud licensing models and offer standardized APIs to their products.
11
7928ch01.fm
12
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
delivery model. Overall performance of a cloud service is dependent on the performance of components outside the direct control of both the client and the cloud service provider, such as the network connection. 4. Integration and interoperability Identifying and migrating appropriate applications to the cloud is made complicated by the interdependencies typically associated with business applications. Integration and interoperability issues include: A lack of standard interfaces or APIs for integrating legacy applications with cloud services. This is worse if services from multiple vendors are involved. Software dependencies which must also reside in the cloud for performance reasons, but which may not be ready for licensing on the cloud. Interoperability issues between cloud providers. There are worries about how disparate applications on multiple platforms, deployed in geographically dispersed locations, can interact flawlessly and can provide the expected levels of service.
13
7928ch01.fm
- The client pays for resources as used allowing for capacity fluctuations over time. - Simple Web interface for self-service provisioning of infrastructure capacity.
- The client pays for resources as used, allowing for capacity fluctuations over time. - Self-service provisioning of
infrastructure capacity only possible up to a point. Standard capacity planning and purchasing processes required for major increases.
Economies of scale
- Potentially significant cost savings are possible from providers economies of scale. - Operating costs for the cloud are absorbed in the usage-based pricing.
SLAs
- SLAs and contractual terms and conditions are negotiable between client and the cloud vendor to meet specific requirements.
- All data and secure information remains behind the corporate firewall.
14
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
Private cloud
Public cloud
Platform choice
- Private clouds can be designed for specific operating systems, applications and use cases unique to the client.
There is no clear right answer, and the choice of cloud model will depend on the application requirements. For example, a public cloud could be ideally suited for development and testing environments, where the ability to provision and decommission capacity at short notice is the primary consideration, while the requirements on SLAs are not particularly strict. Conversely, a private cloud could be more suitable for a production application where the capacity fluctuations are well-understood, but security concerns are high. Many of IBMs largest clients are looking at the introduction of Web 2.0 applications and cloud computing style data centers. Numerous newer Web 2.0-enabled data centers are starting to expand and while they have some key strengths, such as scalability and service delivery, they are also facing operational challenges similar to traditional data centers. The infrastructures that are leveraging some of the early cloud IT models for either acquiring or delivering services, or both, are able to be very nimble and scale very quickly. As a style of computing in which IT-enabled capabilities are delivered as a service to external clients using Internet technologies, popularized by such companies as Google and SalesForce.com, cloud computing can enable large, traditional-style enterprises to start delivering IT as a service to any end user, at any time, in a highly responsive way. Yet, data centers using these new models are facing similar issues of resiliency and security as the need increases for consistent levels of service.
15
7928ch01.fm
The solution to these challenges will not come from todays distributed computing models, but from more integrated and cost-efficient approaches to managing technology and aligning it with the priorities of organizations and end users. To that end, we see an evolutionary computing model emerging, one that takes into account and supports the interconnected natures of the following: The maturing role of the mobile Web The rise of social networking Globalization and the availability of global resources The onset of real time data streaming and access to information
16
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
Cloud computing brings a new paradigm to the cost equation, as well. This approach of applying engineering discipline and mainframe-style security and control to an Internet-inspired architecture can bring in improved levels of economics through virtualization and standardization. Virtualization provides the ability to pool the IT resources to reduce capital expense of hardware, software and facilities. Standardization, with common software stacks and operational policies, helps to reduce operating expense, like labor and downtime - which is far and away the fastest-growing piece of the IT spend. While these technologies have improved the way we do business, they can put a tremendous strain on data centers and IT operations. IT professionals must balance the challenges associated with managing data centers as they increase in cost and complexity against the need to be highly responsive to ongoing demands from the business.
17
7928ch01.fm
individual assets towards complex processes and services. From the data center to the plant floor to the delivery fleet, smart organizations are finding ways to leverage their dynamic infrastructures in support of accelerated, differentiated, and cost-effective service delivery.
18
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
One implementation of a dynamic infrastructure is cloud computing. A dynamic infrastructure conditions the IT environment for a cloud execution model. Cloud resources can be accessed through a variety of methods, and can be rapidly scaled up and scaled down in a secure way to deliver a high quality of service. As discussed in Public clouds versus private clouds on page 13, clouds can be created for internal (private) or external (public) use but always with scalability, elasticity and share-ability of the application deployment and management environment in mind. Users can gain access to their applications from anywhere through their connected devices. With regards to infrastructures, the good news is that there is more than one right answer. The correct solution is to pull the best from existing infrastructure designs and harness new technologies, solutions and deployment options that will free up IT to help run the businessnot just support it. Combining the old best practices (learned from years of managing high-volume, information-intense data centers) with the new best practices enabled by technologies such as virtualization and cloud computing, or pairing ones own IT resources with strategic outsourcing such as managed services from IBM, can assist clients in creating a more dynamic infrastructure. The key initiatives for implementing a dynamic infrastructure are: Virtualization Breaking out of the barriers of physical devices in a data center, such as servers, storage, networks, data, and applications giving clients improved TCO, resiliency and flexibility for a Dynamic Infrastructure. Optimize the energy efficiency of the IT infrastructure to reduce costs, resolve space, power, and cooling constraints, and achieve green strategy objectives
Energy efficiency
Service managementAchieving integrated visibility, control, and automation across all of the business and IT infrastructure components that support differentiated service delivery and accelerated business growth. Service Management includes all aspects of managing complex services environment by controlling both physical and logical resources, targeting Service Level Agreements agreed with the users of any specific service. Asset management Improving asset reliability, availability, and uptime that underpin quality delivery of service according to the priorities of the business, while also maximizing the return on lifetime asset investment along with inventory optimization, labor efficiency, and mitigating the risk of equipment failures that jeopardize the environment and the health and safety of people.
19
7928ch01.fm Security
A new approach to managing risk, providing a full range of security capabilities to organizations, processes, and information as the IT and business infrastructure become more interconnected.
Business resiliency Provides the ability to rapidly adapt and respond to risks, as well as opportunities, in order to maintain continuous business operations, reduce operational costs, enable growth and be a more trusted partner. Information infrastructure A comprehensive approach to a resilient infrastructure for securely storing and managing information and mitigating business risks.
20
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
Figure 1-3 illustrates a traditional, multitier, physical server oriented, data center infrastructure.
Access/Distribution
Switch Pairs
Core Switches
Si
Si
Si
Si
Storage Services
Switches
21
7928ch01.fm
Figure 1-4 illustrates a logical architectural overview of a virtualized data center network that supports server and storage consolidation and virtualization.
Consolidation began as a trend towards centralizing all the scattered IT assets of an enterprise for better cost control, operational optimization, and efficiency. Virtualization introduced an abstraction layer between hardware and software, allowing enterprises to consolidate even furthergetting the most out of each physical server platform in the data center by running multiple virtual servers on it. The leading edge of virtualization is beginning to enable dramatic shifts in the way the data center infrastructures are designed, managed, sourced, and delivered. The emerging trend toward dynamic infrastructures, such as cloud, call for elastic scaling and automated provisioning; attributes that drive new and challenging requirements for the data center network. By comparing the two diagrams we can highlight some important trends that impact the data center network architects and managers and that will be described in detail throughout this document:
22
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
The virtual machine is the new IT building block inside the data center. The physical server platform is no longer the basic component but instead is made up of several logical resources that are aggregated in virtual resource pools Network architects can no longer stop their design at the NIC level, but need to take into account the server platforms network-specific features and requirements, such as vSwitches. These will be presented in detail in Chapter 2, Servers, storage and software components on page 31 The virtualization technologies that are available today (for servers, storage and networks) decouple the logical function layer with the physical implementation layer so physical connectivity becomes less meanigful than in the past. A deeper understanding of the rest of the infrastructure becomes paramount for Network Architects and Managers to understand for a proper data center network design. The network itself can be virtualized and these technologies will also be described in Chapter 3, Data center network functional components on page 149. The architecture view is moving from physical to logical, focused on functions and how they can be deployed rather than on appliances and where they should be placed Infrastructure management integration becomes more important in this environment because the inter-relations between appliances and functions are more difficult to control and manage. Without integrated tools that simplify the DC operations, managing the infrastructure box-by-box becomes cumbersome and more difficult. IBM Software technologies that can be leveraged in order to achieve this goal will be presented in the next chapter. The shift to logical applies also when interconnecting distributed data centers. In order to achieve high availability and maximize resource utilization, it is key to have all the available data centers in an active state. In order to do this, from a network standpoint, the network core must be able to offer layer 1 virtualization services (lamda on fiber connections), layer 2 virtualization services (e.g. VPLS connectivity to enable VM mobility) together with the usual layer 3 services leveraging proven technologies like MPLS.
23
7928ch01.fm
Availability
Availability means that data or information is accessible and usable upon demand by an authorized person. Network availability is affected by failure of a component such as a link. Two factors that allow better availability are redundancy of components and convergence in case of a failure. Redundancy - Redundant data centers involve complex solution sets depending on a clients requirements for backup and recovery, resilience, and disaster recovery. Convergence - Convergence is the time required for a redundant network to recover from a failure and resume traffic forwarding. Data center environments typically include strict uptime requirements and therefore need fast convergence.
Configuration management
Configuration management covers the identification, recording, and reporting of IT/network components, including their versions, constituent components, states and relationships to other network components. Configuration Items (CIs) that should be under the control of configuration management include network hardware, network software, services, and associated documentation.
24
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
It also includes activities associated with change management policies and procedures, administration, and implementation of new installations and equipment moves, additions, or changes. The goal of configuration management is to ensure that a current backup of each network devices configuration exists in a format that can be quickly re stored in case of failure. Also a trail of the changes that have been made on the different components is needed for auditing purposes.
Disaster recovery
The requirement for disaster recovery is the ability to resume functional operations following a catastrophic failure. The definition of functional operations depends upon the depth and scope of an enterprises operational requirements and its business continuity and recovery plans. Multi-data center environments that provide a hot-standby solution is one example of a disaster recovery plan.
Environment
There are environmental factors such as availability of power or air conditioning and maximum floor loading that influence the average data center today. The network architecture must take these factors into consideration.
Failure management
All failures must be documented and tracked. The root cause of failures must be systematically determined and proactive measures taken to prevent a repeat of the failure. Failure management processes and procedures will be fully documented in a well-architected data center environment.
Performance
Network performance is usually defined by the following terms: Capacity Capacity refers to the amount of data that can be carried on the network at any time. A network architecture should take into account anticipated minimum, average, and peak utilization of traffic patterns.
25
7928ch01.fm Throughput
Throughput is related to capacity, but focuses on the speed of data transfer between session pairs versus the utilization of links. Delay Delay, also known as lag or latency, is defined as a measurement of end-to-end propagation times. This requirement is primarily related to isochronous traffic, such as voice and video services. Jitter Jitter is the variation in the time between packets arriving, caused by network congestion, timing drift, or route changes. It is most typically associated with telephony and video-based traffic. Quality of Service Quality of Service (QoS) requirements include the separation of traffic into predefined priorities. QoS helps to arbitrate temporary resource contention. It also provides an adequate service level for business-critical administrative functions, as well as for delay-sensitive applications such as voice, video, and high-volume research applications.
Reliability
Reliabilty is the ability of a system to perform a given function, under given conditions, for a given time interval. There is a difference between availability and reliability. A system can be unavailable but not considered as unreliable because this system is not in use at the given time. An example is a maintenance window where the system is shut down for an upgrade and restarted. During this window, we know the system is not available, but the reliability is not affected. Since todays data center houses critical applications and services for the enterprise, outages are becoming less and less tolerable. Reliability is expressed in terms of the percentage of time a network is available, as shown in Table 1-3 on page 26.
Table 1-3 Reliability Total downtime (HH:MM:SS) Availability 99.9999% 99.999% 99.99% 99.9% per day 00:00:00.08 00:00:00.4 00:00:08 00:01:26 per month 00:00:02.7 00:00.26 00:04:22 00:43:49 per year 00:00:32 00:05:15 00:52:35 08:45:56
26
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
Very few enterprises can afford a 99.9999% level of reliability, because it is usually too costly. Instead, many moderate-sized businesses opt for a 99.99% or 99.9% level of reliability.
Scalability
In networking terms, scalability is the ability of the network to grow incrementally in a controlled manner. For enterprises that are constantly adding new servers and sites, architects may want to specify something more flexible, such as a modular-based system. Constraints that may affect scalability, such as defining spanning trees across multiple switching domains or additional IP addressing segments to accommodate the delineation between various server functions, must be considered.
Security
Security in a network is the definition of permission to access devices, services, or data within the network. The following components of a security system will be considered: Security policy Security policies define how, where, and when a network can be accessed. An enterprise will normally develop security policies related to networking as a requirement. The policies will also include the management of logging, monitoring, and audit events and records. Network segmentation Network segmentation divides a network into multiple zones. Common zones include various degrees of trusted and semi-trusted regions of the network. Firewalls and inter-zone connectivity Security zones are typically connected with some form of security boundary, often firewalls or access control lists. This may take the form of either physical or logical segmentation. Access controls Access controls are used to secure network access. All access to network devices will be via user-specific login credentials; there must be no anonymous or generic logins. Security monitoring
27
7928ch01.fm
To secure a data center network, a variety of mechanisms are available including IDS, IPS, content scanners, and so on. The depth and breadth of monitoring will depend upon both the clients requirements as well as legal and regulatory compliance mandates. External regulations External regulations often play a role in network architecture and design due to compliance policies such as Sarbanes-Oxley (SOX), Payment Card Industry Standards (PCI), the Health Insurance Portability and Accountability Act (HIPAA); and a variety of other industry and non-industry-specific regulatory compliance requirements.
Serviceability
Serviceability refers to the ability to service the equipment. Several factors can influence serviceability, such as modular or fixed configurations or requirements of regular maintenance.
Standards
Network standards are key to the ongoing viability of any network infrastructure. They define such things as: Infrastructure naming standards Port assignment standards Server attachment standards Protocol Standards, ratified for example by IEEE and IETF IP addressing standards
System management
Network management must facilitate key management processes, including: Network Discovery and Topology Visualization
28
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch01.fm
This includes the discovery of network devices, network topology, and the presentation of graphical data in an easily understood format. Availability management This provides for the monitoring of network device connectivity. Event management This provides for the receipt, analysis, and correlation of network events. Asset management This facilitates the discovery, reporting, and maintenance of the network hardware infrastructure. Performance management This provides for monitoring and reporting network traffic levels and device utilization. Incident management The goal of incident management is to recover standard service operation as quickly as possible. The incident management process is used by many functional groups to manage an individual incident. The process includes minimizing the impact of incidents affecting the availability and/or performance, which is accomplished through analysis, tracking, and solving of incidents that have impact on managed IT resources. Problem management This includes identifying problems through analysis of incidents that have the same symptoms, then finding the root cause and fixing it in order to prevent malfunction reoccurrence. User and accounting management This is responsible for ensuring that only those authorized can access the needed resources. Security management This provides secure connections to managed devices and management of security provisions in device configurations.
29
7928ch01.fm
30
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Chapter 2.
31
7928ch02.fm
Section 2.5 will describe the main x86 virtualization technologies available on the market today, together with the main network virtualization features. Section 2.6 will briefly describe the IBM Storage platforms portfolio and the main consolidation and virtualization technologies Section 2.7 will present the main IBM Software products for the management and provisioning of the virtualized data center, from a networking point of view. Section 2.8 will describe the main IBM Integrated Data Center Solutions : iDataPlex and Cloudburst, highlighting the networking options that are available today.
32
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
2.1 Virtualization
Virtualization refers to the abstraction of logical resources away from their underlying physical resources to improve agility and flexibility, reduce costs, and thus enhance business value. Virtualization allows a set of underutilized physical infrastructure components to be consolidated into a smaller number of better utilized devices, contributing to significant cost savings. Server virtualization A physical server is abstracted to provide, or host, multiple virtual servers or multiple operating systems on a single platform. Storage virtualization The storage devices used by servers are presented in the form of abstracted devices, partitioned and dedicated to servers as needed, independent of the actual structure of the physical devices. Network virtualization (will be descibed in more detail in chapter 3) Network devices such as switches, routers, links and network interface cards (NICs) are abstracted to provide many virtualized network resources on few physical resources, or combine many physical resources into few virtual resources.
33
7928ch02.fm
hardware made available to it by the traditional operating system on which it is executed. These are also known as hosted hypervisors. Containers - Virtualization code that runs as an application within a traditional operating system that creates encapsulated, isolated virtual instances that are pointers to the underlying host operating system on which it is executed. This is also known as operating system virtualization. Figure 2-1 depicts the three hypervisor types.
App OS
...
App OS
App OS
...
App OS App
...
App
Type 1 or bare metal hypervisors are the most common in the market today, and they can be further classified into three main sub-types: Stand-alone (for example, VMware ESX and vSphere) Hybrid (for example, Hyper-V or XenServer) Mixed (Kernel Virtual Machine)
Type 1 Stand-alone
In a stand-alone hypervisor, all hardware virtualization and virtual machine monitor (VMM) functions are provided by a single, tightly integrated set of code. This architecture is synonymous with the construct of VMware vSphere and previous generations of the ESX hypervisor. Figure 2-2 on page 35 shows a sample diagram of the architectural overview of VMware vSphere 4.0 (also referred to as ESX 4). Contrary to common belief, VMware is not a Linux-based Hypervisor. Rather, ESX is comprised of a strictly proprietary, highly sophisticated operating system called VMkernel, providing all virtual machine monitor and hardware virtualization functions. The full version of ESX does provide a Linux-based service console (shown to the left of the diagram). As described in the VMware
34
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
section of this document, ESXi (the embedded version of the hypervisor) does not contain this service console instance.
Type 1 Hybrid
The hybrid type 1 architecture includes a split software model where a thin hypervisor provides hardware virtualization in conjunction with a parent partition (privileged virtual machine) which provides virtual machine monitor (VMM) functionality. This model is associated primarily with Microsoft Hyper-V and Xen-based hypervisors.The parent partition, also called Domain 0 (Dom0), is typically a virtual machine that runs a full version of the native operating system with root authority. For example, Dom0 for Xen-enabled and executed within Novell SUSE Linux Enterprise Server (SLES) would execute as a full instance of SLES, providing the management layer of VM creation, modification, deletion, and other similar configuration tasks. At system boot, the Xen-enabled kernel loads initially, followed by the parent partition which runs with VMM privileges, serves as the interface for VM management, and manages the I/O stack. Similar to VMware, all hybrid products available today provide paravirtualized drivers for guests, enabling improved performance to network and I/O resources. Guests not implementing paravirtualized drivers must traverse the I/O stack in the parent partition, degrading guest performance. Operating system (OS) paravirtualization is becoming increasingly common to achieve optimal guest performance and improved interoperability across hypervisors. For example, Microsoft Hyper-V/Windows Server 2008 R2 provides full OS paravirtualization
35
7928ch02.fm
support (also known as Enlightened guest support) for Windows Server 2008 and SUSE Enterprise Linux guests; see Figure 2-3 on page 36.
Traditional OS
Virtual Machine
Traditional OS
Vir tual Machine
Traditional OS
Virtual Machine
Traditional OS
Virtual Machine
Key:
Type 1 Mixed
The Linux-based Kernel Virtual Machine (KVM) hypervisor model provides a unique approach to Type 1 architecture. Rather than executing a proprietary hypervisor on bare-metal, the KVM approach leverages open-source Linux (including RHEL, SUSE, Ubuntu, and so on) as the base operating system and provides a kernel-integrated module (named KVM) that provides hardware virtualization. The KVM module is executed in user-mode (unlike stand-alone and hybrid hypervisors, which run in kernel/root mode), but it enables virtual machines to execute with kernel-level authority using a new instrucion execution context called Guest Mode; see Figure 2-4.
36
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Traditional OS
Virtual Machine
Memory & CPU Paravirtualized I/O
QEMU
Virtualization is a critical part of the data center. It offers the capability to balance loads, provide better availability, reduce power and cooling requirements, and allow resources to be managed fluidly. For more detailed information, see IBM Systems Virtualization: Servers, Storage, and Software, REDP-4396.
37
7928ch02.fm
LUN 0
LUN 1
LUN 2
This SAN is connected to the server through a Fibre Channel connection switch. The advantages of SAN virtualization are: Centralized stored data Ease of backup of stored data Ability to remap a LUN to a different server in case of server failure Storage appears local to servers and to users Improves utilization and reduces storage growth Reduces power and cooling requirements See 2.6 Storage virtualization on page 104 for more details.
38
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
service downtime. The System z platform is commonly referred to as a mainframe. The newest zEnterprise System consists of the IBM zEnterprise 196 central processor complex, IBM zEnterprise Unified Resource Manager, and IBM zEnterprise BladeCenter Extension. The z196 is designed with improved scalability, performance, security, resiliency, availability, and virtualization. The z196 Model M80 provides up to 1.6 times the total system capacity of the z10 EC Model E64, and all z196 models provide up to twice the available memory of the z10 EC. The zBX infrastructure works with the z196 to enhance System z virtualization and management through an integrated hardware platform that spans mainframe and POWER7 technologies. Through the Unified Resource Manager, the zEnterprise System is managed as a single pool of resources, integrating system and workload management across the environment. The following sections contain network-related information about the System z architectural in general. The zEnterpise, however, introduces some new network-specific functions. In fact, the most significant change is that other kinds of servers can now be plugged into the mainframe to create an Ensemble Network where security exposures are minimized and the data center can be managed as if it were a single pool of virtualized resources. For more information on zEnterprise Networking refer to this site: http://www.ibm.com/systems/z/faqs
39
7928ch02.fm
40
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
There are two types of partitions, dedicated and shared, as explained here: A dedicated partition runs on the same dedicated physical processors at all times, which means the processors are not available for use by other partitions even if the operating system running on that partition has no work to do. This eliminates the need for PR/SM to get involved with swapping out one guest and dispatching another.
41
7928ch02.fm
For a better understanding of the above IBM System z functionalities, we will briefly describe the most important network-related topics in more detail in the following sections.
System z
Channel Subsystem
Channel Subsystem: Controls queuing, de-queuing, and priority management, and I/O identification of all I/O operations performed by any logical partitions in a CSS.
Partitions
Partitions: Supports the running of a System Control Program (SCP), such as z/OS and allows CPs, memory, and Subchannels access to channels.
Subchannels
Subchannels: A subchannel represents an I/O device to the hardware, and is used by the SCP to pass an I/O request from the SCP to the channel subsystem.
Channels
Channels: The communication path from the channel subsystem to the I/O network and the connected control units / devices.
One of the major functions is Multiple Image Facility (MIF). MIF capability enables logical partitions to share channel paths, such as ESCON, FICON, and Coupling Facility sender channel paths, between logical partitions within a processor complex. If a processor complex has MIF capability, and is running in LPAR mode, all logical partitions can access the same shared channel paths,
42
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
thereby reducing the number of required physical connections. In contrast, if a processor complex does not have MIF capability, all logical partitions must use separate channel paths to share I/O devices. For more information about CSS, refer to section 2.1 in IBM System z Connectivity Handbook, SG24-5444, which can be found at: http://www.redbooks.ibm.com/abstracts/sg245444.html?Open
HiperSockets
HiperSockets provides the fastest TCP/IP communication between consolidated Linux, z/VM, z/VSE, and z/OS virtual servers on a System z server. HiperSockets provides internal virtual LANs, which act like TCP/IP networks in the System z server. It eliminates the need for any physical cabling or external networking connection between these virtual servers. Figure 2-8 shows an example of HiperSockets connectivity with multiple LPs and virtual servers.
z/VM LP-1
Sysplex A
Linux Guest 1 Linux Guest 2 z/VM Linux Guest n TCP/IP
z/OS LP-5 z/OS LP-6 z/OS LP-7
HiperSockets FD
FE
FF
Linux LP-2 Linux LP-3 Linux LP-4
z/OS LP-8 z/OS LP-9 z/OS LP-10
FC
Figure 2-8 HiperSockets connectivity
Sysplex B
HiperSockets is implemented in the Licensed Internal Code (LIC) of a System z, with the communication path being in system memory. HiperSockets uses
43
7928ch02.fm
internal Queued Direct Input/Output (iQDIO) at memory speed to transfer information between the virtual servers. This iQDIO is an integrated function on the System z servers that provides users with attachment to high-speed logical LANs with minimal system and network overhead. Since HiperSockets does not use an external network, it can free up system and network resources, eliminating attachment costs while improving availability and performance. And also, because HiperSockets has no external components, it provides a very secure connection.
44
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Dynamic channel path management (DCM) Dynamic channel path management is designed to dynamically adjust the channel configuration in response to shifting workload patterns. It is a function in IRD, together with WLM LPAR CPU Management and Channel Subsystem I/O Priority Queuing. DCM can improve performance by dynamically moving the available channel bandwidth to where it is most needed. Prior to DCM, the available channels had to be manually balanced across the I/O devices, trying to provide sufficient paths to handle the average load on every controller. This means that at any one time, some controllers probably have more I/O paths available than they need, while other controllers possibly have too few.
45
7928ch02.fm
Because each Linux image is running in its own virtual system, the Linux images do not affect the other systems around them. By consolidating Linux servers onto one platform, hundreds or thousands of Linux instances on a single server require less energy, cooling, and floor space. Linux on System z is able to run in three modes: basic, LPAR, and z/VM guest. The most common way to consolidate distributed applications is to a single image of Linux running in one of the other two modes, shown in Figure 2-9 on page 46, depending on application footprints and resource usage.
Virtually any application that is portable to Linux can be consolidated to Linux on System z.
46
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
For more infomation on Linux on System z refer to the z/VM and Linux Operations for z/OS System Programmers Redbook, which can be found here: http://www.redbooks.ibm.com/abstracts/sg247603.html?Open
2.2.4 z/VM
z/VM is key to the software side of virtualization on the mainframe. It provides each user with an individual working environment known as a virtual machine (VM). The virtual machine uses virtualization to simulate the existence of a real machine by sharing resources of a real machine, which include processors, storage, memory, and input/output (I/O) resources. Operating systems and application programs can run in virtual machines as guests. For example, multiple Linux and z/OS images can run on the same z/VM system that is also supporting various applications and users. As a result, development, testing, and production environments can share a single physical platform. Figure 2-10 shows an example of a configuration of an LPAR with z/VM. A first-level z/VM means that it is the base operating system that is installed directly on top of the real hardware. A second-level system is a user brought up in z/VM where an operating system can be executed upon the first-level z/VM. In other words, a first-level z/VM operating system sits directly on the hardware, but the guests of this first-level z/VM system are virtualized. By virtualizing the hardware from the first level, as many guests as needed can be created with a small amount of actual real hardware.
47
7928ch02.fm
Linux01
z/VM
z/OS
Linux02
CPUs Memory
CPUs Memory
CPUs Memory
Memory
console
Memory
Hardware
CPUs
CPUs
z/VM consists of many components and facilities that bring the reliability, availability, scalability, security, and serviceability characteristics of System z servers, such as: TCP/IP for z/VM brings the power and resources of the mainframe server to the Internet. Using the TCP/IP protocol suite of TCP/IP for z/VM, multiple vendor networking environments can be reached from the z/VM system. Applications can be shared transparently across z/VM, Linux, and other environments. Users can send messages, transfer files, share printers, and access remote resources with a broad range of systems from multiple vendors. Open Systems Adapter-Express (OSA-Express),Open Systems Adapter Express2 (OSA-Express2) and Open Systems Adapter Express3 (OSA-Express3) are integrated hardware features that enable the System z
48
IBM Data Center Networking: Planning for virtualization and cloud computing
Memory
CPUs
CPUs
7928ch02.fm
platform to provide industry-standard connectivity directly to clients on local area networks (LANs) and wide area networks (WANs). The Resource Access Control Facility (RACF) Security Server for z/VM is a security tool that works together with existing functions in the z/VM base system to provide improved data security for an installation. RACF protects information by controlling access to it. RACF also controls what can be done on the operating system and protects the resources. It provides this security by identifying and verifying users, authorizing users to access protected resources, and recording and reporting access attempts. Contrasted with a discrete server implementation, z/VM-based System z solutions are designed to provide significant savings, which can help lower total cost of ownership (TCO) for deploying new business and enterprise application workloads on a mainframe.
49
z/VM uses various methods to provide devices to virtual machines. CP can dedicate, or attach, a real device to a virtual machine. This gives the virtual machine exclusive use of the entire real device. CP also can virtualize a device, which means it gives a guest a portion of a real device. This can be a portion in time, such as of a processor, or a portion of the devices storage capacity, such as of a disk drive. Network connectivity is an important concern in many environments. z/VM meets customers network needs by offering several networking options. The Control Program can dedicate network devices to virtual machines. The dedicated device can be a channel-to-channel adapter, an IBM Open Systems Adapter (OSA) that provides Ethernet connectivity, or a HiperSockets device, a kind of network adapter that connects one LPAR to another. z/VM also has its own TCP/IP stack, which guests can use as though it were an IP router. A common network option used today is the virtual switch. Here, CP equips each virtual machine with a simulated IBM OSA and connects all those simulated OSAs to a simulated LAN segment called a guest LAN. Also connected to the guest LAN is a real OSA that CP manages. With this configuration established, CP can provide packet- or frame-switching functions for the guests, just as a real switch would in a real external network. In this way, the guest LAN becomes an extension of a real external LAN segment.
50
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
the OSA port strips the Ethernet frame and forwards the IP packets to the virtual switch for delivery to the guest system based on the destination IP address in each IP packet. When operating in Ethernet mode (Layer 2), the virtual switch uses a unique MAC address for forwarding frames to each connecting guest system. Data is transported and delivered in Ethernet frames. This provides the ability to transport both TCP/IP and non-TCP/IP based application data through the virtual switch. The address-resolution process allows each guest systems MAC address to become known to hosts residing on the physical side of the LAN segment through an attached OSA port. All inbound or outbound frames passing through the OSA port have the guest systems corresponding MAC address as the destination or source address. The switching logic resides in the z/VM Control Program (CP), which owns the OSA port connection and performs all data transfers between guest systems connected to the virtual switch and the OSA port; see Figure 2-11.
NIC
NIC
NIC
Port 1
Port 3
Virtual Switch
Ethernet Switch
OSA Port
51
7928ch02.fm
Table 2-1 System z network interfaces Interface type Common link access to workstation (CLAW) Attachment type IBM System p Channel-attached routers
Description Provides access from IBM System p server directly to a TCP/IP stack over a channel. Can also be used to provide connectivity to other vendor platforms. Provides access to TCP/IP hosts by way of a CTC connection established over a FICON or ESCON channel. Provides access to TCP/IP hosts by way of a series A devices and series DX devices that function as series A devices. A variety of channel adapters support a protocol called the LCS. The most common are OSA-Express features.
Channel-tochannel (CTC)
FICON/ESCON channel
Point-to-point
HYPERchannel
Series A devices
Point-to-Multipoint
OSA-Express: 1000BASE-T Fast Ethernet Token Ring ATM Native and LAN Emulation HiperSocketsa OSA-Express: 10 Gigabit Ethernet Gigabit Ethernet 1000BASE-T Fast Ethernet Token Ring ATM LAN Emulation IUTSAMEH (XCF link)
LAN: IEEE802.3 IEEE802.3 IEEE802.5 ATM network Internal LAN LAN: IEEE802.3 IEEE802.3 IEEE802.3 IEEE802.3 IEEE802.5 ATM network Point-to-point
Provides access to TCP/IP hosts, using OSA-Express in Queued Direct I/O (QDIO) mode and HiperSockets using the internal Queued Direct I/O (iQDIO).
Provides access to directly connect z/OS hosts or z/OS LPARs, or by configuring it to utilize Coupling Facility links (if it is part of a sysplex). Enables communication between CS for z/OS IP and other servers running on the same MVS image.
52
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Here, some further considerations on the network features of the OSA, Hipersocket and Dynamic Cross-System Coupling Facility interfaces are provided. The following interfaces are supported by System z hardware and z/OS Communications Server. They deliver the best throughput and performance, as well as offer the most flexibility and highest levels of availability. These interfaces include: OSA-Express OSA-Express utilizes a direct memory access (DMA) protocol to transfer the data to and from the TCP/IP stack. It also provides the offloading of IP processing from the host. The OSA-Express Ethernet features support IEEE standards 802.1p/q (priority tagging and VLAN identifier tagging). OSA-Express also provides primary (PRIRouter) and secondary (SECRouter) router support. This function enables a single TCP/IP stack, on a per-protocol (IPv4 and IPv6) basis, to register and act as a router stack based on a given OSA-Express port. Secondary routers can also be configured to provide for conditions in which the primary router becomes unavailable and the secondary router takes over for the primary router. HiperSockets As described in HiperSockets on page 43, HiperSockets provides high-speed LPAR-to-LPAR communications within the same sever (through memory). It also provides secure data flows between LPARs and high availability, if there is no network attachment dependency or exposure to adapter failures. HiperSockets connection supports VLAN tagging. This allows to split the internal LAN represented by a single HiperSockets CHPID into multiple virtual LANs, providing isolation for security or administrative purposes. Only stacks attached to the same HiperSockets VLAN can communicate with each other. Stacks attached to a different HiperSockets VLAN on the same CHPID cannot use the HiperSockets path to communicate with the stacks on a different VLAN. When the TCP/IP stack is configured with HiperSockets Accelerator, it allows IP packets received from HiperSockets to be forwarded to an OSA-Express port (or vice versa) without the need for those IP packets to be processed by the TCP/IP stack. Dynamic Cross-System Coupling Facility (dynamic XCF) Cross-System Coupling Facility (XCF) allows communication between multiple CSs (Communications Servers) for z/OS IP stacks in a Parallel Sysplex. The XCF connectivity to other TCP/IP stacks can be defined
53
7928ch02.fm
individually, or using the dynamic XCF definition facility. Dynamic XCF significantly reduces the number of definitions needed whenever a new system joins the sysplex or when a new TCP/IP stack needs to be started up. These changes become more numerous as the number of stacks and systems in the sysplex grows. This could lead to configuration errors. With dynamic XCF, the definitions of the existing stacks do not need to be changed in order to accommodate the new stack.
Design considerations
To design connectivity in a z/OS environment, the following considerations should be taken into account: As a server environment, network connectivity to the external corporate network should be carefully designed to provide a high-availability environment, avoiding single points of failure. If a z/OS LPAR is seen as a stand-alone server environment on the corporate network, it should be designed as an end point. If a z/OS LPAR will be used as a front-end concentrator (for example, making use of HiperSockets Accelerator), it should be designed as an intermediate network or node. Although there are specialized cases where multiple stacks per LPAR can provide value, in general we recommend implementing only one TCP/IP stack per LPAR. The reasons for this recommendation are as follows: A TCP/IP stack is capable of exploiting all available resources defined to the LPAR in which it is running. Therefore, starting multiple stacks will not yield any increase in throughput. When running multiple TCP/IP stacks, additional system resources, such as memory, CPU cycles, and storage, are required. Multiple TCP/IP stacks add a significant level of complexity to TCP/IP system administration tasks. One example where multiple stacks can have value is when an LPAR needs to be connected to multiple isolated security zones in such a way that there is no network level connectivity between the security zones. In this case, a TCP/IP stack per security zone can be used to provide that level of isolation, without any network connectivity between the stacks.
54
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Routing support
z/OS Communications Server supports static routing and two different types of dynamic routing: Open Shortest Path First (OSPF) and Routing Information Protocol (RIP). z/OS Communications Server also supports policy-based routing, which determines the destination based on a defined policy. Traffic descriptors such as TCP/UDP port numbers, application name, and source IP addresses can be used to define the policy to enable optimized route selection.
55
7928ch02.fm
support) and applications so that interface failures will not impact application availability. Static VIPA A static VIPA is an IP address that is associated with a particular TCP/IP stack. Using either ARP takeover or a dynamic routing protocol (such as OSPF), static VIPAs can enable mainframe application communications to continue unaffected by network interface failures. As long as a single network interface is operational on a host, communication with applications on the host will persist. Dynamic VIPA (DVIPA) Dynamic VIPAs (DVIPAs) can be defined on multiple stacks and moved from one TCP/IP stack in the sysplex to another automatically. One stack is defined as the primary or owning stack, and the others are defined as backup stacks. Only the primary stack is made known to the IP network. TCP/IP stacks in a sysplex exchange information about DVIPAs and their existence and current location, and the stacks are continuously aware of whether the partner stacks are still functioning. If the owning stack leaves the XCF group (such as resulting from some sort of failure), then one of the backup stacks automatically takes its place and assumes ownership of the DVIPA. The network simply sees a change in the routing tables (or in the adapter that responds to ARP requests). Address Resolution Protocol takeover Address Resolution Protocol (ARP) enables the system to transparently exploit redundant physical interfaces without implementing a dynamic routing protocol in the mainframe. ARP takeover is a function that allows traffic to be redirected from a failing OSA connection to another OSA connection. If an OSA port fails while there is a backup OSA port available on the same subnetwork, then TCP/IP informs the backup adapter as to which IP addresses (real and VIPA) to take over and network connections are maintained. After it is set up correctly, the fault tolerance provided by the ARP takeover function is automatic. Dynamic routing Dynamic routing leverages network-based routing protocols (such as OSPF) in the mainframe environment to exploit redundant network connectivity for higher availability (when used in conjunction with VIPA). Internal application workload balancing Sysplex Distributor (SD) The application workload balancing decision-maker provided with the Communications Server is the Sysplex Distributor (SD). The design of the
56
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
SD provides an advisory mechanism that checks the availability of applications running on different z/OS servers in the same sysplex, and then selects the best-suited target server for a new connection request. The Sysplex Distributor bases its selections on real-time information from sources such as Workload Manager (WLM) and QoS data from the Service Policy Agent. Sysplex Distributor also measures the responsiveness of target servers in accepting new TCP connection setup requests, favoring those servers that are more successfully accepting new requests. Internal workload balancing within the sysplex ensures that a group or cluster of application server instances can maintain optimum performance by serving client requests simultaneously. High availability considerations suggest at least two application server instances should exist, both providing the same services to their clients. If one application instance fails, the other carries on providing service. Multiple application instances minimize the number of users affected by the failure of a single application server instance. Thus, load balancing and availability are closely linked. Portsharing In order for a TCP server application to support a large number of client connections on a single system, it might be necessary to run more than one instance of the server application. Portsharing is a method to distribute workload for IP applications in a z/OS LPAR. TCP/IP allows multiple listeners to listen on the same combination of port and interface. Workload destined for this application can be distributed among the group of servers that listen on the same port. External application workload balancing With external application workload distribution, decisions for load balancing are made by external devices. Such devices typically have very robust capabilities and are often part of a suite of networking components. From a z/OS viewpoint, there are two types of external load balancers available today. One type bases decisions completely on parameters in the external mechanism, and the other type uses sysplex awareness matrixes for each application and each z/OS system as part of the decision process through the Load Balancing Advisor (LBA) function. Which technique is best depends on many factors, but the best method usually involves knowledge of the health and status of the application-instances and the z/OS systems. z/OS Parallel Sysplex z/OS Parallel Sysplex combines parallel processing with data sharing across multiple systems to harness the power of plural z/OS mainframe systems, yet make these systems behave like a single, logical computing facility. This
57
7928ch02.fm
combination gives the z/OS Parallel Sysplex unique availability and scalability capabilities.
58
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
SNMPv3. z/OSCommunications Server also supports standard TCP/IP-based Management Information Base (MIB) data. Additional MIB support is also provided by enterprise-specific MIB, which supports management data for Communications Server TCP/IP stack-specific functions.
Power 780
Power 770 Power 740 Power 750 Power 720 Power 755 Power 730 Power 710 Blades
59
7928ch02.fm
2.3.1 PowerVM
The PowerVM platform is the family of technologies, capabilities, and offerings that deliver industry-leading virtualization on the IBM Power Systems. It is the new umbrella branding term for Power Systems Virtualization (Logical Partitioning, Micro-Partitioning, Power Hypervisor, Virtual I/O Server, Live Partition Mobility, Workload Partitions, and so on). As with Advanced Power Virtualization in the past, PowerVM is a combination of hardware enablement and value-added software.
60
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
The PowerVM Enterprise Edition is only available on the new POWER6-based and POWER7-based systems, and adds PowerVM Live Partition Mobility to the suite of functions. It is possible to upgrade from the Express Edition to the Standard or Enterprise Edition, and from Standard to Enterprise Editions. Table 2-2 on page 61 lists the version of PowerVM which are available on each model of POWER7 processor technology-based servers.
Table 2-2 Availability of PowerVM per POWER7 processor technology-based server model PowerVM Editions IBM Power 750 IBM Power 755 IBM Power 770 IBM Power 780 Express #7793 No No No Standard #7794 No #7942 #7942 Enterprise #7795 No #7995 #7995
It is possible to upgrade from the Express Edition to the Standard or Enterprise Edition, and from Standard to Enterprise Editions. Table 2-3 outlines the functional elements of the three PowerVM editions.
Table 2-3 PowerVM capabilities PowerVM Editions Micro-partitions Maximum LPARs Management Virtual IO Server NPIV Express Yes 1+2 per server VMcontrol IVM Yes Yes Standard Yes 10/core VMcontrol IVM, HMC Yes Yes Enterprise Yes 10/core VMcontrol IVM, HMC Yes Yes
61
7928ch02.fm
Always active on POWER5- , POWER6-, and POWER7-based servers, POWER Hypervisor is responsible for dispatching the logical partition workload across the shared physical processors. It also enforces partition security, and can provide inter-partition communication that enables the Virtual I/O Servers virtual SCSI and virtual Ethernet functions. Combined with features designed into the IBM POWER processors, the POWER Hypervisor delivers functions that enable capabilities including dedicated-processor partitions, Micro-Partitioning, virtual processors, IEEE VLAN compatible virtual switch, virtual Ethernet adapters, virtual SCSI adapters, and virtual consoles. The POWER Hypervisor is a firmware layer sitting between the hosted operating systems and the server hardware, as shown in Figure 2-13 on page 63. The POWER Hypervisor has no specific or dedicated processor resources assigned to it. The POWER Hypervisor also performs the following tasks: Enforces partition integrity by providing a security layer between logical partitions. Provides an abstraction layer between the physical hardware resources and the logical partitions using them. It controls the dispatch of virtual processors to physical processors, and saves and restores all processor state information during virtual processor context switch. Controls hardware I/O interrupts and management facilities for partitions.
62
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Partition
MEM MEM I/O Slot I/O Slot I/O Slot
Partition
MEM I/O Slot I/O Slot
Partition
MEM I/O Slot I/O Slot MEM I/O Slot
...........
P5
...........
P5
P5
P5
P5
P5
P5
POWER Hypervisor
POWERS
POWERS
POWERS
POWERS
MEM
I/O Slot
I/O Slot
I/O Slot
I/O Slot
I/O Slot
MEM
Figure 2-13 The POWER Hypervisor abstracts the physical server hardware
The POWER Hypervisor firmware and the hosted operating systems communicate with each other through POWER Hypervisor calls.Through Micro-Partitioning, the POWER Hypervisor allows multiple instances of operating systems to run on POWER5-based, POWER6- and POWER7-based servers concurrently. A logical partition can be regarded as a logical server, and is capable of booting an operating system and running a workload. Logical partitions (LPARs) and virtualization increase utilization of system resources and add a new level of configuration possibilities. This IBM Power 750 system can be configured with up to 32 cores, and the IBM Power 770 and 780 servers up to 64 cores. At the time of writing, these systems can support up to: Respectively, 32 and 64 dedicated partitions Or, up to 160 micro-partitions
63
7928ch02.fm
It is important to point out that the maximums stated are supported by the hardware, but the practical limits depend on the application workload demands.
64
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
logical partitions. The Virtual I/O Server can use both virtualized storage and network adapters, making use of the virtual SCSI and virtual Ethernet facilities. For storage virtualization, the following backing devices can be used: Direct-attached entire disks from the Virtual I/O Server SAN disks attached to the Virtual I/O Server Logical volumes defined on either of the previous disks File-backed storage, with the files residing on either of the first two disks Optical storage devices For virtual Ethernet we can define Shared Ethernet Adapters on the Virtual I/O Server, bridging network traffic from the virtual Ethernet networks out to physical Ethernet networks. The Virtual I/O Server technology facilitates the consolidation of LAN and disk I/O resources and minimizes the number of physical adapters that are required, while meeting the non-functional requirements of the server. To understand the support for storage devices, refer to the following Web site: http://www14.software.ibm.com/webapp/set2/sas/f/vios/documentation/data sheet.html Virtual I/O Server can run in either a dedicated processor partition or a micro-partition. The different configurations for Virtual I/O Server and associated I/O subsystems can be seen in Advanced POWER Virtualization on IBM System p: Virtual I/O Server Deployment Examples, REDP-4224.
65
7928ch02.fm
The virtual Ethernet switch is not fully compliant with the 802.1q specification since it does not support spanning-tree algorithms and will not participate in spanning-tree calculations. The hypervisor knows the MAC addresses of all virtual Ethernet adapters and thus can switch datagrams between the virtual Ethernet adapters if the adapters belong to the same VLAN. In Figure 2-14 on page 66, packets can be exchanged between Partition 1 and Partition 2. Partition 2 and Partition 3 can also exchange packets. But Partition 1 and Partition 3 cannot exchange packets at the virtual Ethernet switch.
Partition 1 AIX
Partition 2 Linux
Partition 3 AIX
POWER Hypervisor
VLAN 12 VLAN 6
By default, POWER Hypervisor has only one virtual switch but it can be changed to support up to 16 virtual Ethernet switches per system. Currently, the number of switches must be changed through the Advanced System Management (ASM) interface. The same VLANs can be used on different virtual switches and traffic will still be separated by the POWER Hypervisor. Benefits of the virtual Ethernet switch concept are: Partitions think they have a dedicated Ethernet adapter. For pure inter-partition communication, no physical Ethernet adapters are needed. Currently AIX-to-AIX communication over Virtual Ethernet 64 KB MTU is possible as long as there are no non-AIX LPARs in the system. In the presence of other Linux or i5OS LPARs, the MTU must be capped at 9 KB to allow for interoperability. Several VLANs may be defined on one virtual Ethernet adapter, one native VLAN (without a 802.1q tag) and up to 20 tagged VLANs.
66
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
SEA is the component that provides connectivity to an external Ethernet switch, bridging network traffic from the virtual Ethernet adapters out to physical Ethernet networks. In 2.3.8 External connectivity on page 67 we discuss SEA architecture in detail and how external Ethernet switches can be connected. All virtual Ethernet adapters in a Shared Ethernet Adapter (SEA) must belong to the same virtual switch, or SEA creation will fail.
Partition 1 AIX
Partition 2 Linux
Partition 3 AIX
POWER Hypervisor
VLAN 12
Ethernet Network
SEA provides the following functions: Datagrams from the virtual network can be sent on a physical network, and vice versa. One or more physical adapters can be shared by multiple virtual Ethernet adapters. Virtual Ethernet MAC is visible to outside systems. Broadcast and multicast datagrams are bridged. Both VLAN-tagged and untagged datagrams are bridged.
67
7928ch02.fm
SEA has the following characteristics: A Physical Ethernet adapter will not receive datagrams not addressed to itself. SEA puts the physical adapter in promiscuous mode, where it receives all packets seen on the wire. POWER Hypervisor will not give datagrams to a virtual Ethernet if they are not addressed to its MAC address. Virtual Ethernet adapters that belong to the SEA must be created as trunk adapters: when the hypervisor sees a datagram destined for a MAC address it does not know about, it gives it to the trunk adapter. SEA does IPv4 fragmentation when a physical adapters MTU is smaller than the size of a datagram. If the VIOS fails (crashes, hangs, or is rebooted), all communication between the virtual and the physical networks ceases. SEA can be placed in a redundant failover mode. To avoid single points of failure, the SEA can be configured in redundant failover mode; see Figure 2-16. This mode is desirable when the external Ethernet network can provide high availability. Also, more than one physical adapter is needed at the POWER System.
VIOS (Backup)
Adapter
VLAN 6 VLAN 12
Ethernet Network
POWER Hypervisor
Ethernet Network
PSU_temp_0522
68
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Primary and backup SEAs communicate through a control channel on a dedicated VLAN visible only internally to the POWER Hypervisor. A proprietary protocol is used between SEAs to determine which is the primary and which is the backup. One backup SEA may be configured for each primary SEA. The backup SEA is idle until a failure occurs on the primary SEA; failover occurs transparently to client LPARs. Failure may be due to: Physical adapter failure VIOS crash/hang/reboot Inability to ping a remote host (optionally) Each SEA in a failover domain has a different priority. A lower priority means that an SEA is favored to be the primary. All virtual Ethernet adapters on an SEA must have the same priority. Virtual Ethernet adapters of primary and backup SEAs must belong to the same virtual switch. SEAs also support EtherChannel (Cisco proprietary) and IEEE 802.3ad solutions, that is, several Ethernet adapters are aggregated to form one virtual adapter. The adapter aggregation is useful to address increased bandwidth demands between the server and network or in order to improve resilience with the adjacent Ethernet networkespecially when a Multichassis EtherChannel solution is implemented in the Ethernet network. SEA supports four hash modes for EtherChannel and IEEE 802.3ad to distribute the outgoing traffic: default mode - The adapter selection algorithm uses the last byte of the destination IP address (for TCP/IP traffic) or MAC address (for ARP and other non-IP traffic). This mode is typically the best initial choice for a server with a large number of clients. src_dst_port - The outgoing adapter path is selected through an algorithm using the combined source and destination TCP or UDP port values: Average the TCP/IP address suffix values in the Local and Foreign columns shown by the netstat -an command. Since each connection has a unique TCP or UDP port, the three port-based hash modes provide additional adapter distribution flexibility when there are several separate TCP or UDP connections between an IP address pair. Src_dst_port hash mode is recommended when possible. src_port - The adapter selection algorithm uses the source TCP or UDP port value. In netstat -an command output, the port is the TCP/IP address suffix value in the Local column.
69
7928ch02.fm
dst_prt - The outgoing adapter path is selected through an algorithm using the destination system port value. In netstat -an command output, the TCP/IP address suffix in the Foreign column is the TCP or UDP destination port value. SEA supports a fifth hash mode, but it is only available in EtherChannel mode. When used, the outgoing datagrams are scheduled in a round-robin fashion, that is, outgoing traffic is spread evenly across all adapter ports. This mode is the typical choice for two hosts connected back-to-back (that is, without an intervening switch). AIX supports a backup adapter capability for EtherChannel and 802.3ad. Network Interface Backup (NIB) is possible with two adapters (primary/backup).
Considerations
The following could be reasons for not using routing features at the server level: Troubleshooting and modifications of routing issues are more complex because the routing domain is spread over several administrators and thus will increase the operational expenses.
70
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Routed and gated are not as advanced as commercial routing solutions, for example IP fast rerouting mechanisms are not implemented in gated and routed. An FHRP solution could be implemented in the Ethernet network devices so that servers do not participate in the routing domain at all, when possible. When a FHRP schema is implemented in the network the server will have only a default gateway parameter configuredno other routing features are implemented at the server. In POWER5, POWER6, and POWER7 machines, all resources are controlled by the POWER Hypervisor. The POWER Hypervisor ensures that any partition attempting to access resources within the system has permission to do so. PowerVM has introduced technologies that allow partitions to securely communicate within a physical system. To maintain complete isolation of partition resources, the POWER Hypervisor enforces communications standards, as normally applied to external infrastructure communications. For example, the virtual Ethernet implementation is based on the IEEE 802.1q standard. The IBM System p partition architecture, AIX, and the Virtual I/O Server have been security certified to the EAL4+ level. For more information, see: http://www-03.ibm.com/systems/p/os/aix/certifications/index.html More information about this section can be found in PowerVM Virtualization on IBM System p: Introduction and Configuration Fourth Edition, SG24-7940, and IBM Power 770 and 780 Technical Overview and Introduction, REDP-4639.
71
7928ch02.fm
These products use Intel-based chipsets. Both Intel and AMD have developed hardware technology that allows guest operating systems to run reliably and securely in this virtual environment. x86 technology is probably the most widely used today. IBM System x servers support both Intel and AMD hardware virtualization assistance. System x servers are built with the IBM X-Architecture blueprint, which melds industry standards and innovative IBM technology . Some models even include VMware embedded in hardware to minimize deployment time and improve performance. IBM System X and BladeCenter are now part of the IBM CloudBurst offering (see 2.8.2, CloudBurst), so these platforms are also available as integrated modules for a cloud data center.
IBM BladeCenter provides end-to-end blade platform for virtualization of client, server, I/O, networking, and storage IBM Systems Director enables new and innovative ways to manage IBM Systems across a multi-system environment, improving service with
72
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
integrated systems management by streamlining the way physical and virtual systems are managed Unifies the management of physical and virtual IBM systems, delivering a consistent look and feel for common management tasks Multi-system support for IBM Power Systems, Systems x, BladeCenter, System z and Storage Systems Reduced training cost by means of a consistent and unified platform management foundation and interface Full integration of Virtualization Manager into IBM Systems Director 6.1 base functionality Consolidated management for different virtualized environments and tools includes VMware ESX, Microsoft Virtual Server and Xen virtualization as well as Power Systems Hardware Management Console (HMC) and Integrated Virtualization Manager (IVM) Track alerts and system status for virtual resources and their resources to easily diagnose problems affecting virtual resources Perform lifecycle management tasks, such as creating additional virtual servers, editing virtual server resources or relocating virtual servers to alternate physical hosts. Get quick access to native virtualization management tools through launch-in-context Create automation plans based on events and actions from virtual and physical resources, such as relocating a virtual server IBM System Director integration with VMWare Virtual Center VMware VirtualCenter client is installed on the management console and VMware VirtualCenter server is installed on a physical system with: IBM Systems Director Agent Virtualization Manager Agent for VMware VirtualCenter
Drive VMware VMotion using physical hardware status information through automated policies
73
7928ch02.fm
In contrast, a hypervisor (bare-metal) architecture installs the virtualization layer directly on a clean x86-based system. Since it has direct access to the hardware resources rather than going through an operating system, theoretically a hypervisor is more efficient than a hosted architecture and delivers greater scalability, robustness and performance. ESX Server, for example, employs a hypervisor architecture on certified hardware for data center class performance; the hypervisor runs directly on the hardware. The functionality of the hypervisor varies greatly based on architecture and implementation. Each virtual machine monitor (VMM) running on the hypervisor implements the virtual machine hardware abstraction and is responsible for running a guest operating system. Each VMM has to partition and share the CPU, memory, and I/O devices to successfully virtualize the system. Hardware support for virtualization on x86 systems is provided by Intel or AMD. In the past, limitations of the x86 architecture have posed some issues with virtualization. Both Intel and AMD have made improvements that overcome many of these limitations. VMware is currently utilizing all of these x86 virtualization techniques. Virtualization for the x86 architecture is essentially full virtualization architecture with hardware-assisted technology to overcome some of the limitations of full virtualization.
The combined characteristics of extremely compact, mature, manageable, flexible, performing, green, open, and scalable make the IBM Blade Center family of products a very attractive solution for server virtualization and consolidation
74
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
initiatives..For more information on the IBM BladeCenter refer to the IBM BladeCenter Products and Technology Redbook, which can be found here: http://www.redbooks.ibm.com/abstracts/sg247523.html?Open The IBM blade technologies cover a very broad spectrum in the market by including Intel, AMD, POWER servers, and Cell BE blades. This allows great flexibility for using disparate workloads. The IBM BladeCenter chassis portfolio (shown in Figure 2-17 on page 74) is designed for different customer needs: BladeCenter S - For SMBs or remote locations BladeCenter E - High-density and energy-efficiency for the enterprise BladeCenter H - For commercial applications, high-performance computing (HPC), technical clusters, and virtualized enterprise solutions needing high throughput BladeCenter HT - For telecommunications and rugged environments that require high performance and flexible I/O BladeCenter T - For telecommunications and rugged environments such as military, manufacturing, or medical The IBM BladeCenter network hardware supports 1 Gbps and 10 Gbps native Ethernet or the new converged enhanced Ethernet. Bladeservers can have several 10 Gbps converged network adapters or just 1Gbps Ethernet adapters. Switches can be from several vendors like BNT, Brocade or Cisco and can also be legacy Ethernet (1 or 10 Gbps) or 10 Gbps FCoE switches with up to 10 10Gbps uplink ports each. Each H chassis, for example, can have up to four switches which represents up to 400 Gbps full-duplex . All this hardware is standards-based, so standard switches (like IBM, Juniper, Brocade, or Cisco) can be used for connecting the BladeCenter switches, thereby eliminating the need to buy special proprietary switches. IBM BladeCenter also has the following storage options available, external or internal, including internal Solid State Drives: External storage: NAS, FC SAN, iSCSI SAN, SAS SAN Internal storage: Solid state (SSD), Flash, SAS, SATA Characteristics of Solid State Drives include: More than twice the MTBF of conventional drives Optimized for disk mirroring Reduce power consumption compared to conventional drives While they have smaller capacity, the performance advantage is significant and solves some previous speed bottlenecks
75
7928ch02.fm
IBM also supports RAID 5 and battery-backed cache configuration through the use of an optional ServeRAID-MR10e controller.
76
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
77
7928ch02.fm
2.5.1 Xen
Xen originated as a research project at the University of Cambridge, led by Ian Pratt, founder of XenSource, Inc and developed collaboratively by the Xen community spearheaded by XenSource and engineers at over 20 of the most innovative datacenter solution vendors, with IBM second only to Citrix as the most active contributor. The first public release of Xen was made available in 2003. XenSource, Inc. was acquired by Citrix Systems in October 2007. XenSources products have subsequently been renamed under the Citrix brand: XenExpress was renamed XenServer Express Edition and XenServer OEM Edition (embedded hypervisor), XenServer was renamed XenServer Standard Edition, and XenEnterprise was renamed XenServer Enterprise Edition. The Xen project web site has been moved to: http://xen.org Although Xen was owned by XenSource, the nature of Open Source software ensures there are multiple forks and distributions, many released by other vendors, apart from Citrix. In fact Virtual Iron, OracleVM,and Sun xVM are all based on Xen. Red Hat Inc. includes the Xen hypervisor as part of Red Hat Enterprise Linux (RHEL) software.1 At the time of writing, Xen Server v5.5 is the most widely supported version. Xen is a Type 1 hybrid hypervisor. It uses both paravirtualization and full-virtualization with device emulation (QEMU2) and/or hardware assistance. With paravirtualization, the guests operating system (Solaris, SLES, RHEL, FreeBSD) has to be modified to run on Xen. Together with QEMU, it can also provide support for the Windows operating system, which cannot be modified by third parties to support paravirtualization techniques. Figure 2-18 on page 79 displays an overview of Xen. Xen allows paravirtual guests to have direct access to I/O hardware. The scheduler is optimized to support virtual machine guests. I/O overhead is reduced in Xen, as compared to full virtualization techniques. On the other hand, hardware assistance technology processes today offer better performance than paravirtualization. Since the release of Xen Server 3.0, hardware-assisted virtualization has been supported through the use of the Intel VTx and AMD AMD-V, integrated into modern x86/x64 processors.
1
Red Hat is taking two directions. Although its current portfolio is built around the Xen hypervisor, it seems RH's strategic route will be KVM. Actually Red Hat is supporting both the paravirtualized version and the full virtualized version. The latest requires hardware support (Intel VT-x or AMD Pacifica). QEMU is a community-driven project and all Open Source hypervisors leverage it. It can emulate 9 target architectures on 13 host architectures and provide full system emulation supporting more than 200 devices.
78
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Windows hosted on Xen Server products is supported by Microsoft and the collaboration between Microsoft and Citrix is focused on driving industry standards within virtualization. This extends to interoperability between Microsoft and Xen guests, allowing the optional Citrix Essentials package to enable dynamic virtual machine migration between Microsoft Hyper-V and Citrix XenServer.
Virtualization Stack
Storage Virtualization Network Virtualization User Applications User Applications User Applications
Guest OS
Guest OS
Guest OS
SOLARIS
Windows
Direct Access
Direct Access
XenTM Hypervisor
Hardware
Preferred platform Intel VT/AMD V
From a network perspective, Xen requires an extended Layer 2 topology to allow virtual machine mobility (XenMotion). When the movement has occurred, the switching infrastructure expects the VMs MAC to be reachable through a specific interface, which has now changed. To update the external devices, the destination host sends a gratuitous ARP packet as the last step of the migration. Spanning tree protocol should not be a problem because loops are automatically prevented inside the hypervisor. Citrix Xen Server 5.5 supports up to 6 physical Ethernet NICs per physical server. Physical NICs can be bonded in XenServer so that they could act like one.
79
7928ch02.fm
Citrix XenServer v5.5 runs the 3.3 Xen kernel. The Xen 4.0 kernel is expected to incorporate networking-related features such as VLAN tagging per NIC in the VM Configuration File and a Virtual Ethernet Switch. Currently, XenServer uses virtual Ethernet bridging to allow VMs to communicate among each other and to share a physical NIC among many virtual NICs, assigned to the running VMs. MAC address assignments can be randomly assigned or allocated by the administrator. On the internal network, the bridge will be allocated a unicast MAC address which is FE:FF:FF:FF:FF:FF, while on the external network the bridge will share the MAC address of the physical NIC. Ethernet frames coming from the outside to a specific VM arrive properly because the NIC has to be set in promiscuous mode. When VLANs are enabled (the link to the physical access layer is a trunk), every distinct VLAN will have its own bridge. Security is provided basically by keeping the guest operating systems separated. Note the following considerations: It is critical to partition the memory for each guest operating system so that each VM has its dedicated memory. The VM monitor keeps all guests in their dedicated memory sandbox. Assure that there is no memory sharing when partitioning I/O devices are to be shared across different guests. Set the appropriate privilege hierarchy. Only the root user can start and stop guests and run management tasks and scheduling priority. Partitioning should prevent denial of service (DoS) attacks if, for example, there is a bug in a guest. The VM monitor needs to make sure that this will not affect the other guests. Citrix XenServer v5.5 runs the 3.3 Xen kernel, the Xen 4.0 kernel is expected to incorporate networking related features such as VLAN tagging per NIC in the VM Configuration File and a Virtual Ethernet Switch. From a management perspective, Xen has its own kernel, so it uses a separate VM monitor. Being a root user is not needed to monitor a guest and see how it is performing. Xen is the first thing that comes up in the machine. Then Xen loads its first Linux operating system (Domain 0). IBM supports both the management tools and the VM image packaging.
2.5.2 KVM
Kernel-based Virtual Machine (KVM) is a Linux kernel module that turns Linux into an hypervisor; this is known as Linux-based virtualization. It requires
80
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
hardware virtualization assistance (Intel VT-x or AMD Pacifica). It consists of a loadable kernel module, kvm.ko. This module provides the core virtualization infrastructure and a processor-specific module, kvm-intel.ko or kvm-amd.ko. KVM development, which has been carried out since 2006 by the start-up Qumranet, originated with the effort of trying to find an alternative to Xen that could overcome some limitations such as support for NUMA computing architectures. Originally, KVM was intended as a base technology for desktop virtualization and was never intended to be a stand-alone hypervisor. The KVM original patchset was submitted in October 2006. It has been included in the Linux kernel since then and Kumranet was acquired by Red Hat Inc. in September 2008. RHEV3 from Red Hat incorporates KVM. For more information about this subject, visit the following site: http://www.linux-kvm.org/ Without QEMU, KVM is not a hypervisor. With QEMU, KVM can run unmodified guest operating systems (Windows, FreeBSD, Linux). The emulation has very good performance (near-native) because it is supported by the hardware. KVM can run on any x86 platform with a Linux operating system running. The first step is to boot Linux. Then the virtual machines (VMs) are seen as Linux processes, which can be assigned with a different scheduling priority. Figure 2-19 on page 82 displays an overview of KVM. KVM is not a hosted hypervisor. Instead, it uses the hardware to get control of the machine. No modifications to the Linux operating system are required (in upstream Linux). In fact, Linux provides essential services (hardware support, bootstrap memory management, process management and scheduling, and access control), and KVM become a loadable module. It is essentially a CPU/MMU driver. Moreover, running under a Linux kernel allows for not only physical but also virtual memory paging and oversubscription. Recently (accepted for inclusion in linux kernel 2.6.32 release) the Kernel Samepage Merging (KSM) feature has been added. By looking for identical pages and merging them, this provides the memory overcommit capabilities to make efficient usage of the available physical memory and achieve more VMs per host and thus higher consolidation ratios. KVM introduces a new instruction execution mode known as Guest Mode. Usually applications run in User Mode, and an application goes into Kernel Mode when certain system operations are needed, such as writing on the hard drive.
3
RHEV is short for Red Hat Enterprise Virtualization, which is a distribution from Red Hat that is comprised of two components: RHEV-H and RHEV-M. RHEV-H (or Red Hat Enterprise Virtualization Hypervisor) is based on the KVM open source hypervisor. RHEV-M is Red Hat Enterprise Virtualization Manager, an enterprise grade server management system.
81
7928ch02.fm
KVM Guest Mode processes are run from the VMs. This allows the execution of the VMs to occur closer to the kernel, thereby avoiding User Mode context-switching. A slightly modified QEMU is used for I/O. VirtIO is an API for Virtual I/O that implements network and block driver logic. The goal behind VirtIO is to provide a single driver abstraction layer for multiple hypervisors instead of maintaining several drivers for all the different hypervisors. In fact,VirtIO uses User Mode VirtIO drivers to emulate storage and network I/O to maximize performance. This approach is called paravirtualized I/O. The different instruction execution modes in KVM are highlighted in Figure 2-19 on page 82. To overcome the limits of the paravirtualized I/O approach, KVM can leverage hardware-assisted virtualization for PCI pass-through on Intel and AMD platforms and SR-IOV adapters for hypervisor-bypass, allowing to obtain near-native I/O performance.
Traditional OS
Virtual Machine
Memory & CPU Paravirtualized I/O
QEMU
Linux User Mode Linux Kernel Mode
From a network perspective, KVM requires an extended Layer 2 topology to allow virtual machine mobility. Spanning tree protocol should not be a problem because loops are automatically prevented inside the hypervisors. Linux management tools can be used to manage KVM. KVM supports VLANs in the same way that the Linux kernel supports them. These technologies reduce the overhead on the VM monitor through extended page table support. Moreover, KVM supports virtual machine mobility. With regard to virtual networking (or guest networking), KVM internal network can be configured in different modes4:
82
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
User Networking: the VM simply accesses the external network and the Internet. Private Virtual Bridge: a private network between two or more VMs that is not seen by the other VMs and the external network Public Bridge: the same as Private Virtual Bridge, but with external connectivity. KVM can also be seen as a virtualization driver, and in that sense it can add virtualization support of multiple computing architectures, and it is not confined only in the x86 space: IA64, S390 and Embedded PowerPC. IBM supports KVM in terms of management tools, VM image packaging, and hardware platforms (System x, BladeCenter and System z ). KVM permits hybrid mode operation; regular Linux applications can run side-by-side with VMs, achieving a very high rate of hardware efficiency. Moreover, being based on Linux, KVM supports all the hardware that Linux supports. VM Storage access, for example, is performed through Linux. KVM uses Linux and is a self-contained, non-intrusive patchset. IBM supports RHEL 5.4 with KVM in production environments today, and will continue to work with clients and others to enhance KVM over time.
83
7928ch02.fm
Existing applications
App App App App App
Future applications
App App App
VMware vSphere 4
Availability Application Services
VMotion Storage vMotion HA Fault Tolerance Disaster Recovery
Security
vShield Zones VMSafey
Scalability
DRS Hot Add
vCompute
ESX ESXi DRS
vStorage
VMFS Thin Provisioning
vNetwork
Distributed Switch
Internal cloud
External cloud
The vSphere 4 kernel has been developed for 64-bit processor. Also, vSphere utilizes the virtualization acceleration feature on microprocessors, such as Intel VT and AMD-V. This will enhance virtualization performance and scalability compared with ESX Version 3 which is built as 32 bit and software based. vSphere virtualizes and aggregates the underlying physical hardware resources across multiple x86 servers providing pools of virtual resources to the data center. vSphere is a hypervisor based solution that is installed on the bare metal of the x86 server and creates an abstraction layer between the physical CPU, memory, storage and network to the virtual machines. 5 There are two types of vSphere software. The vSphere ESX software has a service console that is compatible with Red Hat Enterprise Linux (RHEL) 5.2.
5
84
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Third party applications can be installed into the service console to provide additional management for the ESX server. The other type is called ESXi and it does not have a service console. VMware has published a set of APIs that can be used by third parties to manage this environment. This environment is deemed more secure by not having this Linux environment to maintain. ESXi is maintained using the VMkernel interface. Figure 2-4 on page 85 describes some of the performance and capacity enhancements of vSphere as it relates to virtual machines and the network.
Table 2-4 Performance and capacity enhancements of vSphere Maximum Number of CPUs per VM Memory per VM Network throughput per Host IOPS per Host ESX 3.5 4 64 GB 9 Gbps 100,000 ESX 4.1 8 255 GB 40 Gbps 200,000+
ESX overview
All virtual machines (VMs) run on top of the VMkernel and share the physical resources. VMs are connected to the internal networking layer and gain transparent access to the ESX server. With virtual networking, virtual machines can be networked in the same way as physical machines. Figure 2-21 shows an overview of the ESX structure.
85
7928ch02.fm
Virtual Machine
Virtual Machine
Virtual Machine
Virtual Machine
Service console
CPU
Memory
86
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Virtual NIC Virtual NIC, also called Virtual Ethernet Adapter, is a virtualized adapter that emulates Ethernet and is used by each VM. Virtual NIC has its own MAC address, which is transparent to the external network. Virtual switch port A virtual port corresponds to a physical port on a physical Ethernet switch. A virtual NIC connects to the port that you define on the virtual switch. The maximum number of virtual switch ports per ESX server is 4096. The maximum number of active ports per ESX server is 1016. Port group A port group specifies port configuration options such as bandwidth limitations and VLAN tagging policies for each member port. The maximum number of port groups per ESX server is 512. Uplink port Uplink ports are ports associated with physical adapters, providing a connection between a virtual network and a physical network. Physical NIC Physical NIC is a physical Ethernet adapter that is installed on the server. The maximum number of physical NICs per ESX server varies depending on the Ethernet adapter hardware. Figure 2-22 on page 88 shows an overview of how ESX virtual network components relate to each other.
87
7928ch02.fm
Virtual Switch
Virtual Port
Uplink Port
Physical NIC -1
Physical NIC -2
Physical NIC -3
VLAN tagging
802.1Q tagging is supported on a virtual switch. There are three types of configuration modes: Virtual switch tagging (VST) Define one port group on a virtual switch for each VLAN, then attach the virtual machines virtual adapter to the port group instead of to the virtual switch directly. The virtual switch port group tags all outbound frames and removes tags for all inbound frames. It also ensures that frames on one VLAN
88
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
do not leak into a different VLAN. Use of this mode requires that the physical switch provide a trunk. This is the most typical configuration. Virtual machine guest tagging (VGT) Install an 802.1Q VLAN trunking driver inside the virtual machine, and tags will be preserved between the virtual machine networking stack and external switch when frames are passed from or to virtual switches. External switch tagging (EST) Use external switches for VLAN tagging. This is similar to a physical network, and a VLAN configuration is normally transparent to each individual physical server.
QoS
The virtual switch shapes traffic by establishing parameters for three outbound traffic characteristics: Average bandwidth, burst size, and peak bandwidth. You can set values for these characteristics for each port group. For the virtual Distributed Switch, the same traffic classifications can be placed on the inbound traffic.
High availability
In terms of networking, you can configure a single virtual switch to multiple physical Ethernet adapters using NIC teaming. A team can share the load of traffic between physical and virtual networks and provide passive failover in the event of a hardware failure or a network outage. You can set NIC teaming policies (virtual port ID, source MAC hash or IP hash) at the port group level. High availability of VM is described later in this section.
89
7928ch02.fm
Figure 2-23 Virtual switch configuration of a server with four physical NICs
90
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
virtual port Groups are applied to the uplinks and not the vmnics on individual hosts. Each vmnic on each host is mapped to virtual uplinks, permitting teaming and failover consistency, irrespective of vmnic assignments.
vCenter
Virtual Uplink
Virtual Uplink
Virtual Uplink
Distributed Switch
VM
VM
VM
With vCenter management, the distributed switch is configured like a logical single switch that combines multiple virtual switches on different ESX servers. vDS does not provide switch-to-switch interconnection such as STP. Access to corresponding VLANs should be prepared by physical network layers prior to configuring the vDS. Uplinks on each vDS should be in the same broadcast domain. Consider vDS as a template for the network configuration.
Nexus 1000V
VMware provides the Software Development Kit (SDK) to third-party partners to create their own plug-in to distributed virtual switches. Nexus 1000V is one of the
91
7928ch02.fm
third-party enhancements from Cisco. After 1000V is installed, it acts as substitution for the virtual Distributed Switch. Note the following major differences: Switching function Virtual port channel Link Aggregation Control Protocol (LACP) Load balancing algorithm enhancement, source and destination MAC address, IP address, IP port and hash Traffic control DSCP and ToS marking Service Class (CoS) Security Access control list DHCP snooping Dynamic ARP inspection (DAI) Management Same command line interface (CLI) as other Cisco devices Port analyzer (SPAN and ERSPAN) Netflow v9 support RADIUS and TACCS support Nexus 1000V has two major components: the Virtual Ethernet Module (VEM) running on each ESX hypervisor kernel, and, as in vDS, a Virtual Supervisor Module (VSM) managing multiple clustered VEMs as a single logical switch. VSM is an appliance that is integrated into vCenter management. Cisco implements VN-Link on both Nexus 1000v and other Nexus hardware switch. VN-Link offloads switching function from virtual switch. It defines association with virtual NIC and virtual port called VNtag ID. For more information on VN-Link, see: http://www.cisco.com/en/US/netsol/ns894/index.html Communications between VMs are processed on external hardware switch even if those VMs are on same segment on same virtual switch. Currently this function is Cisco proprietary; Nexus 1000v and Cisco Nexus 2000/4000/5000 are required. Cisco has proposed this technology to IEEE.
92
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
There are other, similar technologies. Virtual Ethernet Port Aggregator (VEPA) and Virtual Ethernet Bridging (VEB), which can offload switching function from a virtual switch, are also proposed to IEEE.
93
7928ch02.fm
Cluster
Physical Server Physical Server Physical Server Physical Server Physical Server Root Resource Pool: 5x (4.0 GHz, 4 GB) (CPU = 24 GHz Mem = 20 GB)
Virtual Machine
Virtual Machine
Virtual Machine
Virtual Machine
The VMware vCenter management server monitors the utilization of resources and balances computing resources across resource pools. Resource allocation is based on predefined rules. This capability is based on VMotion, which makes it possible to move an entire VM environment to another physical ESX server.
94
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
VMotion overview
This function moves a running VM from one physical server to another with minimum impact to users. Figure 2-26 shows an overview of VMotion.
VCenter ESX Server A Management Console Virtual Machine V3 Virtual Machine Physical Network V3 ESX Server B
Virtual Machine V1
Virtual Machine V2
Virtual Machine V4
Virtual Machine V5
VMWare Kernel
VMWare Kernel
In this overview, the ESX servers (ESX Server A and ESX Server B) share the storage (for example, the SAN or iSCSI volume). When VM3 is VMotion from ESX Server A, it is moved, which means the memory contents of the virtual machine is copied to ESX Server B. After the transfer is completed, VM3 is activated on ESX Server B. This technology provides for a dynamic infrastructure which includes features for: Dynamic resource optimization in the resource group High availability of VM Easy deployment of VM VMotion only works in an L2 environment. In other words, source and target ESX servers should be located within the same broadcast domain.
VM Direct I/O
VMDirectPath is a new capability provided in vSphere for direct assignment of physical NIC/HBA to a VM for guest.VMDirectPath is designed for VMs that require the dedicated network bandwidth. But virtual machines that use Direct I/O cannot perform additional VM functions such as VMotion, fault tolerance,and suspend/resume. Note that this function requires specific network adapters listed on the compatibility list provided by VMWare.
95
7928ch02.fm
96
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Careful capacity planning and testing should be performed. At this time, only virtual machines with one virtual CPU are supported.
97
7928ch02.fm
APP OS
APP OS
APP OS
APP OS
vmkernel TCP/IP
Virtual Port vDS Data Plane Component on each ESX/ESXI Host
vMotion
ISCSI
NFS
Guest VMs
Management
FT
Shaper applies an absolue limit ( if configured) for each traffic type over he NIC team
Scheduler prioritized traffic leaving the host according to relative "shares" values configured for each traffic type on the vDS
IPv6 enhancements
vSphere 4.1 is undergoing testing for U.S. NIST Host Profile compliance that includes requirements for IPsec and IKEv2 functionality (with the exception of MLDv2 plus PKI and DH-24 support within IKE).
98
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
In vSphere 4.1, IPv6 is supported for: Guest virtual machines ESX/ESXi management vSphere client vCenter Server vMotion IP storage (iSCSI, NFS) - experimental Note: IPv6 is not supported for vSphere vCLI, VMware HA and VMware FT logging. IKEv2 is disabled by default.
Taken from Cloud Director Installation and Configuration Guide, EN-000338-00 found at http://www.vmware.com/pdf/vcd_10_install.pdf
99
7928ch02.fm
vCenter
ESX/ESXi ESX/ESXi ESX/ESXi ESX/ESXi ESX/ESXi
vCenter vCenter
vShield Manager vShield Manager
vShield Manager
Figure 2-28 shows a vCloud Director cluster comprised of four server hosts. Each host runs a group of services called a vCloud Director cell. All hosts in the cluster share a single database. The entire cluster is connected to three vCenter servers and the ESX/ESXi hosts that they manage. Each vCenter server is connected to a vShield Manager host, which provides network services to the cloud. Table 2-6 provides information about the limits in a vCloud Director installation.
Table 2-6 Limits in a Cloud Director installation Category Virtual Machines ESX/ESXi Hosts vCenter Servers Users Maximum number 10,000 1000 10 5000
100
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
The vCloud Director installation and configuration process creates the cells, connects them to the shared database, and establishes the first connections to a vCenter server, ESX/ESXi hosts, and vShield Manager. After installation and configuration is complete, a system administrator can connect additional vCenter servers, vShield Manager servers, and ESX/ESXi hosts to the Cloud Director cluster at any time. vCloud Director includes three different techniques to isolate network pools: VLAN backed - a range of VLAN IDs and a vNetwork distributed switch are available in vSphere. The VLAN IDs must be valid IDs that are configured in the physical switch to which the ESX/ESXi servers are connected. vSphere port groups - Unlike other types of network pools, a network pool that is backed by port groups does not require a vNetwork distributed switch. This is the only type of network pool that works with Cisco Nexus 1000V virtual switches. vCloud Director Network Isolation (VCDNI) - An isolation backed network pool does not require pre-existing port groups in vSphere but needs a vSphere vNetwork distributed switch. It uses portgroups which are dynamically created. A Cloud isolated network spans hosts, provides traffic isolation from other networks. This technique uses MAC-in-MAC encapsulation. To learn more about the value of IBM System x, BladeCenter and VMware, see IBM Systems Virtualization: Servers, Storage, and Software, REDP-4396.
2.5.4 Hyper-V R2
Windows Server 2008 R2 includes Hyper-V R2, which is a hypervisor-based architecture (bare metal hypervisor) that is a very thin software layer (less than 1 MB in space). It was released in the summer of 2008. The free, stand-alone version of Hyper-V (Microsoft Hyper-V Server 2008) was released in October. The two biggest concerns with Hyper-V have been addressed in the R2 release: It fully supports failover clustering. It now includes live migration, which is the ability to move a virtual machine from one physical host to another without service interruption. Hyper-V is a Type 1 hybrid hypervisor; that is, a thin hypervisor provides hardware virtualization in conjunction with a parent partition (privileged virtual machine), which provides virtual machine monitor (VMM) functionality. It leverages both paravirtualization and full virtualization because it is the hypervisor that mediates between the hardware and the unmodified operating systems. It can run only on 64-bit servers, but can host both 32-bit virtual machines and 64-bit virtual machines.
101
7928ch02.fm
It is capable of virtualizing multiple Windows and non-Windows operating systems (only SUSE Linux Enterprise Server 10 is officially certified for paravirtualization at this time) on a single server. It requires hardware-assisted virtualization (Intel VT or AMD Pacifica) and hardware Data Execution Prevention (DEP). A primary virtual machine (parent partition) runs only Windows Server 2008 and the virtualization stack, and has direct access to the physical machines hardware and I/O. The other VMs (children) do not have direct access to the hardware. Also Hyper-V implements a virtual switch. The virtual switch is the only networking component that is bound to the physical network adapter. The parent partition and the child partitions use virtual network adapters (known as vNICs), which communicate with the virtual switch using Microsoft Virtual Network Switch Protocol. Multiple virtual switches can be implemented, originating different virtual networks. There are three types of virtual networks in Hyper-V: External Virtual Network: The VMs can access the external network and the Internet. This must be tied to a specific physical network adapter. Internal Virtual Switch: This permits VM-to-VM communication but not external connectivity. It can also facilitate communications between the host operating system and the guests. Private Virtual Network: The same as an internal virtual switch, but they cannot access the host operating system. VLAN IDs can be specified on each Virtual Network Virtual Switch except for the Private Network Virtual Switch. Figure 2-29 on page 103 shows an overview of Hyper-V. The purple VMs are fully virtualized. The green VMs are paravirtualized (enlightened partitions). Paravirtualized drivers are available for guests; guests not implementing paravirtualized drivers must traverse the I/O stack in the parent partition, degrading guest performance. The VMbus is a logical channel that connects each VM, using the VSC (Virtualization Service Client) to the parent partition which runs the VSP (Virtualization Service Provider). The VSP handles the device access connects from the child partitions (VMs).
102
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
WS08 Parent
Partition VM Worker Processes WMI Provider VM Service WS08 Kernel IHV Drivers
WS08 Natively
SLES 10 SP1
Windows 2000
Server not Enlightened w/o Integration Components
RHEL 4
User Mode
IC IC
VSC VSC Hypercall Adapter VSC
IC
Kernel Mode
VSP
VMBus VMBus VMBus Emulation Emulation
At least one VLAN must be created to allow the VMs to communicate with the network (VLAN creation is supported by Hyper-V). The physical NIC is then configured to act as a virtual switch on the parent partition. Then the virtual NICs are configured for each child partition to communicate with the network using the virtual Switch. Each virtual NIC can be configured with either a static or dynamic MAC address. At this time NIC teaming is not supported. Spanning tree protocol should not be a problem because loops are prevented by the virtual switch itself. If it is small or medium-sized, the entire environment can be managed by the Hyper-V Manager. In an enterprise scenario, Microsoft System Center Server Management Suite Enterprise (SMSE) is the recommended choice. In the future, this should also allow the management of Citrix XenServer and VMWare ESX v3 hosts, apart from Hyper-V hosts. One significant advantage to the Microsoft approach is that SMSE is not limited to managing virtual environments because it was designed to manage all systems, including physical and virtual.
103
7928ch02.fm
For more information on implementing Hyper-V on System x, see Virtualization cookbook: Microsoft Hyper-V R2 on IBM System x at this link: http://www-01.ibm.com/common/ssi/cgi-bin/ssialias?infotype=SA&subtype=W H&htmlfid=XSW03046USEN&appname=STG_BC_USEN_WH
104
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
File
SONAS
For clients requiring: Massive I/O, backup or restores Consolidation, or scale large numbers of clusters
XIV
For clients requiring: Breakthrough ease of use and management High utilization with automated performance (hot spot) management Virtually unlimited snapshots Advanced self-healing architecture
Enterprise
For clients requiring: Advanced disaster recovery with 3way mirroring and System z GDPS support Continuous availability, no downtime for upgrades Best-in-class response time for OLTP or transaction workloads, flash optimization Single system capacity scalable to the PB range
N series
For clients requiring: NAS storage Simple two-site high availability
DS5000
Storwize V7000
For clients requiring: 10s of TB of rack-mounted storage with sophisticated software functions Breakthrough ease of use and management Non-disruptive migration from or virtualization of, existing disk
Midrange
For clients requiring: Good cost/performance, generalpurpose storage Need to add capacity to existing configuration 10s of TB of capacity
Entry
DS3000
For clients requiring: SMBs or branch office locations; cost sensitive; start at as small as 100s GB, up to low TB capacity
Storage Optimizers
SVC
For clients requiring optimization of storage costs: Storage technologies that add function, performance, ease of use or efficiency to new or existing storage
For more information on IBM Storage Solutions please refer to the IBM System Storage Solutions Handbook, SG24-5250, which can be found here: http://www.redbooks.ibm.com/abstracts/sg245250.html?Open
2.6.1 Storage Area Networks (SAN) and SAN Volume Controller (SVC)
SAN-attached storage connects storage to servers with a Storage Area Network using ESCON or Fibre Channel technology. Storage area networks enable to share homogeneous storage resources across the enterprise. For many companies, however, information resources are spread over various locations and storage environments with products from different vendors. With this in mind, the best storage solution takes advantage of the existing investment and provides growth whenit is needed.
105
7928ch02.fm
For more detailed information about this subject, see SAN Volume Controller in IBM Systems Virtualization: Servers, Storage, and Software, REDP-4396 which can be found at: http://www.redbooks.ibm.com/abstracts/redp4396.html?Open As illustrated in Figure 2-31, a SAN has the following characteristics: Remote, shared storage access Private network for storage Storage protocols Centralized management
Client
LAN
Network Server
Network Server
Backup Server
Disk Storage
SAN
Tape
Based on virtualization technology, SAN Virtual Controller (SVC) supports a virtualized pool of storage from the storage systems attached to a SAN. This storage pool helps to tap unused storage capacity and make the business more efficient and resilient.
106
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
SVC helps to simplify storage management by presenting a single view of storage volumes. Similarly, SVC is an integrated solution supporting high performance and continuous availability in open system environments, as shown in Figure 2-32.
The solution runs on clustered storage engines, based on System x servers and open standards-based technology. Industry-standard host bus adapters (HBAs) interface with the SAN fabric. SVC represents storage to applications as virtual disks, created from the pool of managed disks residing behind the storage engines. Storage administrators can scale performance by adding storage engines and capacity by adding disks to the managed storage pool. The SAN Volume Controller allows to do the following: Manage storage volumes from the SANs Virtualize the capacity of multiple storage controllers Migrate data from one device to another without taking the storage offline Increase storage capacity utilization and uptime Virtualize and centralize management Support advanced copy services across all attached storage Use virtual disks and management
107
7928ch02.fm
Support local area network (LAN)-free and server-free backups Manage up to 8 petabytes (PBs) of total usable storage capacity Apply copy services across disparate storage devices within the network Respond with flexibility and speed Experience high availability in terms of SVC nodes, mirrors, and fault tolerance SVC is designed to support the delivery of potential benefits to meet a clients business requirements. These benefits include, but are not limited to: Keeping applications running Supporting continued access to data via data migration capabilities Allocating more capacity to an application automatically Reducing cost and complexity Saving costs of midrange storage while achieving the benefits of enterprise storage Managing systems from different vendors and automation Providing a single place to manage multiple, different systems Providing a common set of functions Increasing staff productivity Creating a virtualized pool of heterogeneous storage environments Reducing administration efforts Generating additional operational savings Utilizing storage assets more efficiently Combining storage capacity from many disk arrays into a single storage resource Centralizing storage management Applying copy services and replication across disparate storage arrays
108
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
operational simplicity and energy efficiency, support a low cost of ownership, and increase reliability to provide significant operational efficiencies. For further details, see Virtualization Engine TS7520: Virtualization for open systems in IBM Systems Virtualization: Servers, Storage, and Software, REDP-4396 which can be found at: http://www.redbooks.ibm.com/abstracts/redp4396.html?Open
109
7928ch02.fm
Massive parallelism and sophisticated distributed algorithms that yield superior power and value and very high scalability (both with a scale out and with a scale up model). Compatibility benefits due to the use of standard, off-the-shelf components. The system can leverage the newest hardware technologies without the need to deploy a whole new subsystem Maximized utilization of all disks, true distributed cache implementation coupled with more effective cache bandwidth, and practically zero overhead incurred by snapshots Distributed architecture, redundant components, self-monitoring and auto-recovery processes; ability to sustain failure of a complete disk module and three more disks with minimal performance degradation For more information on IBM XIV Storage System refer to theIBM XIV Storage System: Architecture, Implementation, and Usage, SG24-7659, which can be found here: http://www.redbooks.ibm.com/abstracts/sg247659.html?Open
110
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Client
File I/O
File I/O
NAS
Figure 2-33 Network Attached Storage overview
111
7928ch02.fm
2.6.8 N-Series
The IBM System Storage N series provides a range of reliable, scalable storage solutions for a variety of storage requirements. These capabilities are achieved by using network access protocols such as Network File System (NFS), Common Internet File System (CIFS), HTTP, iSCSI and FCoE as well as storage area network technologies such as Fibre Channel (FC), SAS and SATA, utilizing built-in Redundant Array of Inexpensive Disks (RAID) technologies. The N series is a specialized, thin server storage system with a customized operating system, similar to a stripped-down UNIX kernel, hereafter referred to as Data ONTAP. The IBM System Storage N series Gateways, an evolution of the N5000 series product line, is a network-based virtualization solution that virtualizes tiered, heterogeneous storage arrays, allowing customers to utilize the dynamic virtualization capabilities available in Data ONTAP across multiple tiers of IBM and vendor-acquired storage. For more information on N Series Storage Systems refer to the IBM System Storage N Series Hardware Guide, SG24-7840, which can be found here: http://www.redbooks.ibm.com/abstracts/sg247840.html?Open
112
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
dynamically brings new capabilities to the business that can be exploited to deliver higher efficiencies. The ability to move virtual machines from physical host to physical host while the virtual machines remain operational is another compelling capability. One of the key values that IBM systems management software can provide is to mask the complexities that are introduced by virtualization. Businesses are embracing virtualization because it brings value and enhances capabilities for business continuity and disaster recovery. The ability to use business policy-based process automation for orchestrating, provisioning, workload, and service level management in line with business goals will drive higher levels of virtualization adaptation. In dynamic infrastructure and cloud computing environments, the support of automated provisioning of servers, storage and network is a key enabler to support services in a highly-virtualized infrastructure. Managed provisioning is the interlock between the service management supported by Tivolis Service Automation Manager and the resources management supported by single platform applications (such as Systems Director or Tivoli Network Manager). Figure 2-34 on page 114 illustrates the solution overview of a service-oriented architecture to deploy provisioning services to the virtualized managed nodes at the resources level. The solution is comprised of the following elements: Self-service provisioning of virtualized network, storage, and server resources User-initiated scheduling and reservation through a service catalog model Self-service provisioning of a preconfigured application Customized deployment workflows implementing client-specific processes and enforcing specific policies Integrated Service Management of the Provisioning Environment
113
7928ch02.fm
Capacity Management
Monitoring
Virtualized Infrastructure
114
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Internal
Managed Services
Appliances
SaaS
Cloud (Public/Private/Hybrid)
Consume Services
Self-Service Portal Dashboard Consolidated Reporting
Workflow Modeling Data Modeling Tool Configurat ion Develop & Test
Transpor tation
People
F acilities
Remote
Applications Information
System
Storage
N etw ork
Voice
Security
IBM Service Management offerings provide a complete solution, automating the full life cycle of service requests, incidents, and trouble tickets from their creation through to the environmental changes that they produce. IBM Service Management products are integrated to capture incoming requests, incidents, and problems; route them to the correct decision-makers; and expedite resolution with enterprise-strength server and desktop provisioning tools. They do this while keeping an accurate record of all the Configuration Items in a federated management database and a real-time picture of the deployed infrastructure, thus matching hardware and software services with the business needs that they fulfill. By automating change, configuration, provisioning, release, and asset management tasks, IBM Service Management software helps to reduce cost and eliminate errors.
115
7928ch02.fm
Figure 2-36 illustrates a view of the overall architecture of IBM Service Management (ISM). It is based on the Tivoli Process Automation Engine (TPAe), which provides a common runtime and infrastructure (shown in green) for the various ISM components. TPAe provides a process runtime, as well as a common user interface and configuration services that can be leveraged by other components. On top of TPAe, Process Management Products (PMPs) can be deployed. The PMPs provide implementations of processes around the ITSM service life cycle. The Change Management PMP, for example, provides a good practices framework for implementing the change management process according to ITSM and ITIL. IT management products that are outside the scope of TPAe, so-called Operational Management Products (OMPs), are accessible by means of Integration Modules, so their functionality can be used by PMPs inside TPAe.
Tivoli Service Automation Manager (TSAM) allows to quickly roll out new services in the data center to Support dynamic infrastructure and cloud computing services. The TSAM component is also based on the Tivoli Process Automation Engine (TPAe), implementing a data model, workflows and
116
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
applications for automating the management of IT services by using the notion of Service Definitions and Service Deployment Instances. The set of applications provided by TSAM can be collectively referred to as the TSAM Admin User Interface (UI), which provides access to all TSAM-specific functionality. It should be noted that this UI is not aimed at end users, because some level of detailed knowledge about TSAM itself and about the concrete types of services automated using TSAM is required. The TSAM Admin UI is therefore more likely to be used by designers, operators, and administrators of services. A more end user-centered interface for requesting and managing IT services automated through TSAM can be provided by exposing certain functionality as service catalog offerings using the Service Request Management (SRM) component. By means of service offerings and offering UIs, the level of complexity exposed for any kind of service can be reduced to a level consumable by end users, hiding many of the details that are exposed by the TSAM Admin UI. TSAM and SRM rely on the CCMDB component to access information about resources in the managed Data Center. As illustrated in Figure 2-37, TSAM interlocks with Tivoli Provisioning Manager (TPM) to drive automated provisioning and deployment actions on the managed IT infrastructure. That is, while TSAM concentrates on the process layer of IT Service Management and hooks automation into these processes, TPM can be used for really implementing these automated actions in the backend.
117
7928ch02.fm
118
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
OMNIBUS ObjectServer
RE ECTU RCHIT RIC A
GENE
Omnibus Probe
Security Events
Network Events
System Events
Storage Events
Transaction Events
Mainframe Events
OMEGAMON
ITM
ITM
Tivoli Netcool/OMNIbus software is available on a variety of platforms, such as Microsoft Windows Vista, Sun Solaris and Linux on IBM System z. Operations staff can launch in context directly from Tivoli Netcool/OMNIbus events to detailed Tivoli Network Manager topology and root cause views, as well as other views from the Tivoli Monitoring family, Tivoli Service Request Manager, Tivoli Application Dependency Manager and many more. Tivoli Netcool/OMNIbus can serve as a manager of managers that leverages existing investments in management systems such as HP OpenView, NetIQ, CA Unicenter TNG, and many others. The current Netcool/OMNIbus version 7.3 includes, as the strategic Event Management Desktop, the Web user interface previously provided by Netcool/Webtop. Netcool/OMNIbus v7.3 also includes Tivoli Netcool/Impact for the query and integration of data stored and generated by IBM Tivoli Software products only, and not for other data management or automation purposes with data sources that are not Tivoli products. For example, Tivoli Netcool/Impact can be used to integrate with information generated by IBM Tivoli Netcool/OMNIbus or correlated with information from other licensed Tivoli product data stores, but cannot be used to integrate with non-Tivoli, custom or 3rd party data sources, read/write data to such data sources, or any resources managed by Tivoli products.
119
7928ch02.fm
Netcool/OMNIbus V7.3 has implemented significant improvements and enhancements in the following areas: Performance and scalability High performance Active Event List (AEL) Ultra-fast event filtering through user-defined indexes and automatic filter optimization Manage events from multiple silos (ObjectServers) in a single AEL Improved multi-threading and event consumption rates Tiered Architecture (ESF) configuration included in the default configurations and documentation Enhanced event flood protection and event rate anomaly detection including new event generation at the probe ObjectServer gateways with smarter resynchronization for improved disaster recovery performance Deployment integration and interoperability Improved operator workflows Ability to view and work with events from multiple ObjectServers in a single event list New Active Event List configuration options providing familiar look and feel New dashboard portlet to create monitor box views, supporting drag-and-drop configuration New graphics engine for improved look and feel Improved user interactions with the map, including customizable mouseover effects Improved performance with more than 50 K events supported in a single AEL Replacement of entities with system filters and views Consistent look and feel based on Tivoli guidelines
Map Refresh
Usability/Consumability
Tivoli Netcool/OMNIbus provides single pane of glass visibility to help leverage and extend the native capabilities provided by the Tivoli common portal interface with cross-domain, multivendor event management, thereby enabling centralized
120
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
visualization and reporting of real-time and historical data across both IBM and third party tools. This information is consolidated in an easy-to-use, role-based portal interface, accessed through single sign-on, so that all the monitoring data and management information needed can be retrieved from one place.
121
7928ch02.fm
Figure 2-39 on page 122 provides an overview of Tivoli Provisioning Manager functionality.
IBM Tivoli Provisioning Manager 7.1.1 software uses the Tivoli process automation engine that is also used by other IBM Service Management products (such as the Change and Configuration Management Database and Tivoli Service Request Manager) to provide a common look and feel and seamless integration across different products.
TPM simplifies and automates data center tasks. The main capabilities are summarized here: It provides operating system imaging and bare metal provisioning. TPM offers flexible alternatives for quickly creating and managing operating systems cloned or scripted installs such as dynamic image management, single instance storage, encrypted mass deployments, and bandwidth optimization.
122
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
It provides software distribution and patch management over a scalable, secure infrastructure. TPM can automatically distribute and install software products defined in the software catalog without creating specific workflows or automation packages to deploy each type of software. It provides automated deployment of physical and virtual servers through software templates. It provides provisioning support for different system platforms. It provisions new virtual machines on VMWare ESX servers. It provisions Windows OS and applications onto a new ESX virtual machine. It creates new LPARs on pSeries. It provisions new AIX OS onto a pSeries LPAR through NIM. It integrates with the IBM TotalStorage Productivity Center to provide a storage capacity provisioning solution designed to simplify and automate complex cross-discipline tasks. It supports a broad range of networking devices and nodes, including firewalls, routers, switches, load balancers and power units from leading manufacturers (such as Cisco, Brocade Networks, Extreme, Alteon, F5 and others). It provides compliance and remediation support. Using compliance management, the software and security setup on a target computer (or group of computers) can be examined and then compare that setup to the desired setup to determine if they match. If they do not match then noncompliance occurs, and recommendations about how to fix the noncompliant issues are generated. It provides report preparation. Reports are used to retrieve current information about enterprise inventory, activity, and system compliance. There are several report categories. Each category has predefined reports that can be run from the main report page or customized, in the wizard, to suit the business needs. Also the report wizard can be used to create new reports. It supports patch management. TPM provides out-of-the-box support for Microsoft Windows, Red Hat Linux, AIX, Solaris, HP Unix, and SUSE Linux Enterprise Sever using a common user interface. Accurate patch recommendations for each system are based on vendor scan technology.
123
7928ch02.fm
124
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Enterprise Management
Advanced Managers
V MCont rol Net work Cont rol Ac tive E nergy Manager
RE ST AP I
CLI
Platform Management
Storag e Managemen t Network Management Power Systems Management* BC / X Systems Ma nagement* z System Mana gement* STG Ot hers IHV Others (CI SCO,
CAS et c)
Platform Management
Console Assembl y Management Server Asse mbl y Age nt Assembly (Common & Platfo rm) Hosti ng R untime (LWI) Southbound Interfaces
SS H DCOM SLP SNM P CIM
Resource Manageability
Vi rtual Servers
(CIM, S LP , ot her)
Sto ra ge Resources
(CI M, SNMP, SLP )
N etwork Resource s
(CIM , S NMP, SLP , ot her)
IBM Systems Director provides the core capabilities that are needed to manage the full lifecycle of IBM server, storage, network, and virtualization systems: Discovery Manager discovers virtual and physical systems and related resources. Status Manager provides health status and monitoring of system resources. Update Manager acquires, distributes, and installs update packages to systems. Automation Manager performs actions based on system events. Configuration Manager configures one or more system resource settings. Virtualization Manager creates, edits, relocates, and deletes virtual resources. Remote Access Manager provides a remote console, a command line, and file transfer features to target systems.
125
7928ch02.fm
Apart from these basic features, there are platform-specific plug-ins for every IBM platform. Enhancements improve ease of use and deliver a more open, integrated toolset. Its industry-standard foundation enables heterogeneous hardware support and works with a variety of operating systems and network protocols. Taking advantage of industry standards allows for easy integration with the management tools and applications of other systems. Optional, fee-based extensions to IBM Systems Director are available for more advanced management capabilities. Extensions are designed to be modular, thereby enabling IT professionals to tailor their management capabilities to their specific needs and environments. Version 6.1, launched in November 2008, provides storage management capabilities. Version 6.1.1, which shipped starting in May 2009, also includes basic (SNMP-based) network management and discovery capabilities (Network Manager). IBM System Director Network Manager provides platform network management for IBM Ethernet switches7: Network devices Discovery and Inventory Thresholds definition and monitoring Troubleshooting tools such as ping and traceroute A set of default groups of network resources to allow users to view status of a group of related devices or perform tasks across a set of devices Using Figure 2-40 on page 125 as a reference, Systems Director integrations can be categorized into the following four macro areas: Through northbound interfaces This includes, for example, using REST APIs towards Tivoli products such as TEP (ITM) or to leverage ITNM network discovery capabilities, and also towards non-IBM products such as CA Unicenter, HP OpenView and Microsoft Systems Management Server. Through southbound interfaces This includes, for example, accessing physical network resources using SNMP. With other platform management tools Such tools provide Launch-in-context capabilities with Cisco, Brocade, and Juniper fabric management tools and API-based integration in the future.
7
For more information about the supported Network Devices, refer to this site: http://publib.boulder.ibm.com/infocenter/director/v6r1x/topic/director.plan_6.1/fqm0_r_supported_h ardware_and_software_requirements.html
126
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
With other System Director advanced managers These include, for example, the BladeCenter Open Fabric Manager (BOFM) or VMControl. From a networking point of view, the pivotal element that provides integration with Tivoli products and the rest of the Data Center Infrastructure is Systems Director Network Control. IBM Systems Director Network Control builds on Network Management base capabilities by integrating the launch of vendor-based device management tools, topology views of network connectivity, and subnet-based views of servers and network devices. It is responsible for: Integration with Tivoli Network Management products Broaden Systems Director ecosystem Enhanced platform management Virtualization Automation The differences between the basic Network Management functionalities (included in Director free of charge) and the Network Control enhancements (which are fee-based) are listed inTable 2-7.
Table 2-7 Network management functionality of IBM Systems Director Task or feature IBM Systems Director Network Management Yes Yes Yes Yes Yes Yesa Yes Yes Yes Yes No No IBM Systems Director Network Control Yes Yes Yes Yes Yes Yesa Yes Yes Yes Yes Yes Yes
Network system discovery Health summary Request access (SNMP, Telnet) Collect and view inventory View network system properties Network-specific default groups View network problems and events Network monitors and thresholds Event filters and automation plans Network diagnostics (ping, traceroute) Vendor-based management tool integration Network topology collectionb
127
7928ch02.fm
Task or feature
a. IBM Systems Director Network Control provides an additional Subnets group. b. This is not supported on Linux for System p.
For more information about IBM Systems Director, refer to the Systems Director 6.1.x Information Center page: http://publib.boulder.ibm.com/infocenter/director/v6r1x/index.jsp
128
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Third-party tools such as CiscoWorks or HP OpenView can be in-context launched for device-specific monitoring, configuration, and troubleshooting. Accurate monitoring and root cause analysis ITNM can be configured to actively monitor the network for availability and performance problems. ITNM uses the network topology model to group related events and identify the root cause of network problems, thereby speeding mean time to repair. Event correlation uses a network topology model to identify the root cause of network problems. Network events can be triggered by setting personalized thresholds for any SNMP performance metrics including complex expressions such as bandwidth utilization. Figure 2-41 on page 130 illustrates the Network Manager processes. The discovery agents discover the existence, configuration and connectivity information for network devices, updating the network connectivity and information model (NCIM). At the same time, OMNIbus probes receive and process events from the network. The polling agents poll the network for conditions such as device availability and SNMP threshold breaches. If a problem or resolution is detected, the polling agents generate events and send them to OMNIbus. The Event Gateway provides the communication functionality for ITNM to enrich and correlate events from OMNIbus.
129
7928ch02.fm
Network Manager
Event Gateway RCA Engine
NCIM
OMNIbus
Omnibus Probes
Polling Agents
Discovery Agents
Topology-based root cause analysis (RCA) consists of the event stream being analyzed in the context of the discovered relationships (physical and logical). Root cause analysis can be performed on any events from any source. Dynamic root cause analysis recalculates the root cause and symptoms as new events arrive. The RCA process can include simple hierarchical networks or even complex meshed networks, with remote (OSPF and BGP, for example), intra (such as card failure), and intra-device dependencies (such as cable failures). ITNM enriches events from Netcool probes and the Netcool Knowledge Library. SNMP traps and other events from Netcool probes are enriched from an extensive set of rules known as the Netcool Knowledge Library. ITNM further enriches these events with topology information to provide the user with useful contextual information including navigating to topology maps to see the device in various network contexts, and using diagnostic tooling. Events can be configured to alert the operator when network services, such as VPNs, have been effected providing the operator with information on priority in order to perform triage.
130
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
IBM Tivoli Network Managers latest release at the time of this writing is 3.8.0.2 and it is available as a stand-alone product or integrated with Netcool/OMNIbus (O&NM - OMNIbus and Network Manager). These are the most recently included features: IPv6 support Support for new technologies and products: Juniper NetScreen, MPLS TE, Multicast PIM, Cisco Nexus 5000/7000, and Brocade 8000 An option to use Informix to host the network topology model ITNM can be integrated with different Tivoli products such as ITM, TBSM, Netcool/OMNIbus, TPM, Impact, and others. Several scenarios illustrating these integrations are presented in the next section. A comprehensive overview of possible ITNM integrations is shown in Figure 2-42.
For more information about ITNM, refer to the IBM Tivoli Network Management information center at: http://publib.boulder.ibm.com/infocenter/tivihelp/v8r1/index.jsp?toc=/c om.ibm.netcool_OMNIbus.doc/toc.xml
131
7928ch02.fm
132
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
The policy-based compliance management function of Tivoli Netcool Configuration Manager provides a rules-based tool for checking and maintaining network device configuration compliance with various sets of policies. You can configure compliance checks to check for the presence or absence of specific commands or data in a device's configuration or a response from a query to a device. Based on the results of the compliance check, the tool either reports the results or, if desired, initiates a configuration change to bring a device back into compliance. You can organize and group related compliance checks into higher-level policies. You can schedule compliance checks to execute on a dynamic or scheduled basis. Likewise, you can set up compliance checks to trigger automatically as a result of a configuration change on a device. The following capabilities are available in the policy-based compliance management function of Tivoli Netcool Configuration Manager: Policy-based compliance management uses reusable building blocks for creating compliance checks: Definitions: the lowest level component that contains the configuration data to be checked Rules: composed of definitions and the actions to take if a device configuration passes or fails Policies: composed of rules, which can be grouped together across various device types and compliance checks. Processes: group one or more policies to execute against a set of devices on a scheduled basis The function allows you to run compliance checks in either a dynamic, scheduled, or automatically triggered manner. The function automatically fixes out-of-compliance device configurations to get them back in compliance. Policy-based compliance management allows you to configure compliance reports for automatic generation and distribution. The integration of Tivoli Network Manager, Tivoli Netcool/OMNIbus, and Tivoli Netcool Configuration Manager provides a closed loop network management problem resolution in one single solution. Figure 2-43 on page 134 shows the closed loop problem resolution provided by the integration of Tivoli Network Manager, Tivoli Netcool/OMNIbus, and Tivoli Netcool Configuration Manager.
133
7928ch02.fm
ITNM
RCA
DETECT
FIX TNCM enables automated correction of common problems through automated remediation
ITNCM
Policy Compliance
The integration scenario described in this documentation provides the following benefits: The integration reduces inefficiencies associated with separate products to monitor different aspects of the same network. By sharing specific device information and preserving device groupings and hierarchies between Tivoli Network Manager, Tivoli Netcool/OMNIbus and Tivoli Netcool Configuration Manager, you reduce the need to run separate discoveries. The integration ensures both Tivoli Network Manager and Tivoli Netcool Configuration Manager have the same view of the network, and one that is constantly updated as the network changes. This ensures the administrators for the two products have a consistent view of network outages and enables them to isolate root cause of outages easily. Device configuration problems can be difficult to isolate and identify. By integrating with Tivoli Netcool Configuration Manager, a unified view of events is created that helps operators isolate the problems caused by changes to network device configuration. Tivoli Netcool Configuration Manager maintains a backup of network configuration, including audit trails. Accidental misconfigurations of a network
134
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
device are easy to spot, isolate, and rectify, by simply rolling back the changes via Tivoli Netcool Configuration Manager. The integration provides the ability to implement network policies and enforce compliance by utilizing the capability of Tivoli Netcool Configuration Manager to make a change to large number of devices in one go, while ensuring the changes are accurate without manual intervention. This reduces the time to value of network management implementation Within a single application, ITNCM forms the foundation to the network and network services resource configuration, and a combined multi-service, multivendor device configuration management and service activation supporting platform. In addition, the ITNCM application combines elements of internal workflow, inventory of the infrastructure under management and automation of the activation of the business services to be deployed on the production infrastructure. Consequently it is possible that ITNCM can be used to launch services onto the network infrastructure in a low risk, low cost, efficient manner, providing a faster return of investment and a rapid service turn-up to customers on the infrastructure without the need to implement huge complex and expensive operational support systems (OSS)/business support systems (BSS) infrastructures. OSS/BSS integration is a key element in the ITNCM solutions architecture. With ITNCM's network automation foundation in-place and the Web Services API, ITNCM has the ability to integrate the network and network services, seamlessly into the business workflow, and customer management systems at a later stage. Built upon the patented Intelliden R-Series Platform, ITNCM offers network control by simplifying network configurations, standardizing best practices and policies across the entire network, and automating routine tasks such as configuration and change management, service activation, compliance and security of mission critical network. Figure 2-44 exemplifies this.
Simplify
configuration management to enable lower level engineers to do more
Standardize
best practices while delivering converged services
Automate
service activation through flexible vendor-agnostic network aware definitions
135
7928ch02.fm
ITNCM provides flexibility to support a breadth of network changes for heterogeneous devices. By automating the time-consuming, error-prone configuration process, ITNCM supports the vast majority of network devices available from Cisco, Juniper and Brocade and many other vendors' equipment. Programmable templates allow you to make bulk changes to multiple devices simultaneously, and changes are driven by pre-defined policies that determine who can make what changes where in the network. The key to policy-based change management is a configurable workflow engine for automating error-prone manual and scripted network configurations tasks. It also provides repeatability and a full audit trail of all network changes. Key configuration management features: Change management implements security approvals and scheduled configuration changes Automated conflict detection alerts you to potentially conflicting updates Change detection identifies and reports on configuration changes A complete audit trail tracks all changes, both through ITNCM and outside of the system
136
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
The three main areas of integration for network management are listed here: Integration with Event Management This allows the incorporation of network-related events into an overall event dashboard, for example ITNM-Netcool/OMNIbus integration. Integration with Systems Management This allows the inclusion of networking infrastructure management and provisioning within systems and storage management and provisioning tools, for example the Systems Director Network Manager module. Integration with Service Management This allows relating networking management and provisioning with IT service management provisioning. Also the automation of recurring tasks can be performed theoretically leveraging this kind of integration. The network infrastructure provisioning has to be orchestrated with the server and storage provisioning, for example the TBSM-ITNM integration.
137
7928ch02.fm
Tivoli development started an integration initiative that provides a guideline about how to converge the products to a common set of rules to allow the products to work together. IBM is implementing this initiative with each product release to enhance the overall integration. The integration initiatives cover the following tracks: Security integration initiatives Security integration enables Tivoli products to integrate on security aspects, such as authentication and single sign-on, shared user registry support, centralized user account management, consistent authorization, audit log consolidation, and compliance reporting. Navigation integration initiatives Navigation initiatives allow seamless user interface transition from different Tivoli products when the context is needed. This seamless integration involves integrated user interface and launch in context (LIC) abilities. Data integration initiatives Data integration allows data structures to be exchanged to ensure the management context is available across different Tivoli products. This data integration includes event transfer and management resource consolidation. Task integration initiatives Task integration allows a Tivoli management application to use a facility that is provided by a separate Tivoli product. Therefore, they do not need to provide an overlapping functionality. Reporting integration Reporting integration provides centralized management reporting across various Tivoli products. This reporting integration is realized by using the Tivoli Common Reporting. Agent management Agent management allows self-monitoring of various Tivoli products using IBM Tivoli Monitoring agents. In addition, different levels of integration can be achieved regarding a certain aspect. For example, for Navigation Integration this can be provided by: Basic navigation, which is the ability to move seamlessly between views provided by multiple related products. Launch, where one product console can be launched from another. Launch in context, where the launched console comes up in the same context that the user had in the launching console. A user might be looking at an event about a problem with a computer system and launch in context to another product console. When it comes up, it displays further information
138
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
about that computer system. Single Sign On integration should also be enabled to provide a seamless navigation integration. Shared console is an even deeper level of integration. The same console has panels with information from multiple products. When the user changes contexts in one panel, the other panels switch to the same context. The Tivoli Integrated Portal can perform this function. Generally speaking, the following integration technologies can be leveraged to accomplish these tasks: By using Application Programming Interfaces (APIs), for example RESTful By using IDML files exchange, using Discovery Library Adapters (DLAs) to translate data structures for different DBs, for example. DLAs are an easy-to-develop, lightweight solution that allows for rapid integration between management products, customer data, and other third-party data sources. These IDML files are created by DLAs on a periodic frequency (set by the client) and then sent to a common location (set by the client) for multiple management products to consume the same set of IDML files. Using Tivoli Integration Composer as a gateway between ex-Micromuse and ex-MRO products such as Maximo. For more information about these topics, see the IBM Redbooks publication Integrating Tivoli Products, SG24-7757: http://www.redbooks.ibm.com/abstracts/sg247757.html?Open
139
7928ch02.fm
2.8.1 iDataplex
iDataPlex is an innovative, high density, scale out x86computing solution. Its design goals are listed here below: Optimized both mechanically as a half-depth server solution and component-wise for maximum power and cooling efficiency Designed to maximize utilization of data center floor space, power, and cooling infrastructure with an industry standard-based server platform Easy-to-maintain solution with individually serviceable servers, front access hard drives/cabling Customer-specific computer, storage, or I/O needs and delivered preconfigured for rapid deployment Common tools across the System x portfolio for management at the node, rack or data center level These design goals go far beyond a single server or a single rack level; they are goals for the entire data center. With the new philosophy and the new design, iDataPlex solution promises to address the data center challenges at various levels: An innovative rack design achieves higher node density within the traditional rack footprint. Various networking, storage, and I/O options are optimized for the rack design. An optional Rear Door Heat eXchanger virtually eliminates traditional cooling based on computer room air conditioning (CRAC) units. An innovative flex node chassis and server technology are based on industry standard components. Shared power and cooling components improve efficiency at the node and rack level. Intelligent, centralized management is available through a management appliance. Each of these innovations is described in the following paragraphs.
140
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
48 23.6
82.4
However, the iDataPlex rack is shallower in depth compared to a standard 42U server rack as shown in Figure 2-47 on page 142. The shallow depth of the rack and the iDataPlex nodes is part of the reason that the cooling efficiency of iDataPlex is higher than the traditional rack design (20% less cooling required, 63% less fan power), because air travels a much shorter distance to cool the internals of the server compared to airflow in a traditional rack.
141
7928ch02.fm
600 mm
(840mm with RDHx)
446 mm x 520 mm
446 mm x 520 mm
444 mm x 700 mm
444 mm x 700 mm
1050 mm
Figure 2-47 Comparison of iDataPlex with two standard 42U racks (top view)
An iDataPlex rack provides 84U (2 columns of 42U) of horizontal mounting space for compute nodes and 16U (2 sets of 8U) of vertical mounting space (sometimes called pockets) for power distribution units and networking equipment. This makes a total of 100U usable space on two floor tiles.
142
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch02.fm
Storage node
This approach provides maximum flexibility for data center solutions to incorporate a combination of configurations in a rack.
143
7928ch02.fm
2U Chassis
iDataPlex Rear Door Heat Exchanger
dx320 server
dx360 server
Storage Tray
I/O Tray
Switches
PDUs
3U Chassis
On the left side is a 2U (top) and a 3U chassis (bottom). These can be equipped with several different server, storage, and I/O components. The populated chassis are mounted in an iDataPlex rack (middle). This rack was specifically designed to meet high-density data center requirements. It allows mandatory infrastructure components like switches and power distribution units (PDUs), as shown on the right, to be installed into the rack without sacrificing valuable server space. In addition, the iDataPlex solution provides management on the rack level and a water-based cooling option with the Rear Door Heat eXchanger.
144
IBM Data Center Networking: Planning for virtualization and cloud computing
Draft Document for Review February 10, 2011 6:42 pm Table 2-8 Supported Ethernet switches Switch BNT RackSwitchTM G8000F BNT RackSwitch G8100F BNT RackSwitch G8124F Cisco 2960G-24TC-L Cisco 3750G-48TS Cisco 4948-10GE Force10 S50N Force 10 S2410CP SMC 8024L2 SMC 8126L2 SMC 8708L2 SMC 8848M Ports 48 24 24 24 48 48 48 24 24 26 8 48 Layer L2/3 L2 L2 L2 L2/3 L2/3/4 L2/3 L2 L2b L2 L2 L2c
7928ch02.fm
Port speed 1 Gbpsa 10 Gbps 10 Gbps 1 Gbps 1 Gbps 1 Gbpsa 1 Gbpsa 10 Gbps 1 Gbps 1 Gbps 10 Gbps 1 Gbpsa
a. 10 Gb uplink modules available b. Has L2/3/4 Quality of Service features c. Has limited L3 capabilities
If the plan is to use copper-based InfiniBand and Ethernet cabling, then mount both switches horizontally, if possible. This is due to the number of cables that go along the B and D columns; all the copper cables take up so much space in front of the vertical pockets that proper cabling to vertically-mounted switches is no longer possible.Figure 2-50 on page 146 illustrates the recommended cabling.
145
7928ch02.fm
Switch
Switch
PDU
PDU
PDU
42 U
42 U
Switch
42 U
Switch
42 U
PDU
PDU
PDU
Figure 2-50 Recommended cabling for Ethernet-only (left) and Ethernet plus InfiniBand (right)
For more information, see the Redbooks publication Implementing an IBM System x iDataPlex Solution, SG24-7629: http://www.redbooks.ibm.com/redbooks/pdfs/sg247629.pdf
2.8.2 CloudBurst
IBM CloudBurst is a prepackaged private cloud offering that brings together the hardware, software, and services needed to establish a private cloud. This offering takes the guesswork out of establishing a private cloud by preinstalling and configuring the necessary software on the hardware and leveraging services for customization to the environment. Just install the applications and begin leveraging the benefits of cloud computing, such as virtualization, scalability, and a self-server portal for provisioning new services. IBM CloudBurst includes a self-service portal that allows users to request their own services, automation to provision the services, and virtualization to make system resources available for the new services. This is all delivered through the integrated, prepackaged IBM CloudBurst offering and includes a single support interface to keep things simple. IBM CloudBurst is positioned for enterprise clients looking to get started with a private cloud computing model. It enables users to rapidly implement a complete cloud environment including both the cloud management infrastructure and the cloud resources to be provisioned.
146
IBM Data Center Networking: Planning for virtualization and cloud computing
PDU
PDU
7928ch02.fm
Built on the IBM BladeCenter platform, IBM CloudBurst provides preinstalled capabilities essential to a cloud model, including: A self-service portal interface for reservation of compute, storage, and networking resources, including virtualized resources Automated provisioning and de-provisioning of resources Prepackaged automation templates and workflows for most common resource types, such as VMware virtual machines Service management for cloud computing Real-time monitoring for elasticity Backup and recovery Preintegrated service delivery platforms that include the hardware, storage, networking, virtualization, and management software to create a private cloud environment faster and more efficiently IBM CloudBurst is a prepackaged and self-contained service delivery platform that can be easily and quickly implemented in a data center environment. It provides an enhanced request-driven user experience, and aids in an effort to help drive down costs and accelerating time to market for the business. The major service management capabilities of IBM CloudBurst include: IBM Tivoli Service Automation Manager Enhanced Web 2.0 interface image management capability. Resource reservation capabilities allowing environments to be scheduled. The delivery of integrated IBM Tivoli Usage and Accounting capability to help enable chargeback for cloud services to optimize system usage. This enables the creation of resource usage and accounting data to feed into Tivoli Usage and Accounting Manager, which allows for tracking, planning, budgeting, and chargeback of system resource usage. A high availability option using Tivoli Systems Automation and VMware high availability that provides protection against unplanned blade outages and that can help simplify virtual machine mobility during planned changes. Integration with Tivoli Monitoring for Energy Management that enables monitoring and management of energy usage of IT and facility resources, which can assist with efforts to optimize energy consumption for higher efficiency of resources, in an effort to help lower operating cost. An optional secure cloud management server with IBM Proventia Virtualized Network Security platform from IBM Security. IBM Proventia protects the CloudBurst production cloud with Virtual Patch, Threat
147
7928ch02.fm
Detection and Prevention, Proventia Content Analysis, Proventia Web Application Security, and Network Policy enforcement. Detects and blocks network attacks and unauthorized network access. Enables cloud computing service providers to deliver segmented security in multi-tenant virtual environments. Integrates virtualized security with traditional network protection to reduce complexity of security operations. Support for the ability to manage other heterogeneous resources outside of the IBM CloudBurst environment (this requires the purchase of additional licenses).
148
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
Chapter 3.
149
7928ch03.fm
150
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
Data Center Network-wide virtualization- these techniques extend the virtualization domain to the whole Data Center Network and even to the WAN in case there are multiple Data Centers.
One-to-many
Physical Switch
Many-to-one
Virtualized Server
VM1 APP OS VM2 APP OS
Logic al S W5
Hypervisor / VMM
vNICs, NIC Sharing, SR-IOV, vSwitching 802.1q, MPLS Node partitioning techniques. Virtual routers, switches, FWs VLANs, MPLS L3 VPNs
NIC teaming, NIC bonding Link aggregation, Port Channel, Tunneling Node aggregation techniques: Stacking, Virtual Chassis, VSS Internet VPNs
Network Nodes
Some of these virtualization features are based on open standards (such as 802.1q VLANs), but many others are still tied to specific vendor implementations, so interoperability is not always guaranteed. The virtualization features that were introduced are targeted mostly at the forwarding and control planes. However, management plane and services plane virtualization techniques are also emerging in the marketplace. Management plane virtualization enables multi-tenant data center network infrastructures because different virtual networks can be managed independently without conflicting with each other. Services Plane virtualization (see 3.4.5, Network Services Deployment Models on page 175) allows you to create virtual services contexts that can be mapped to different virtual networks, potentially in a dynamic fashion. This is a pivotal functionality in a dynamic infrastructure because it provides
151
7928ch03.fm
simplified scaling, reduced time-to-deploy and more flexibility compared to the appliance-based deployment model. Another level of abstraction that is quite common in networking is the concept of an overlay network. This can be defined as a virtual network defined on top of an physical network infrastructure. An example of an overlay network is the Internet in its early stages, which was an overlay on the plain old telephone service (POTS). The same approach can be used for large scale content delivery and for video surveillance applications over IP for example. Other networking-related virtualization features can refer to the ways network addresses are used in the data center. Both layer 2 (MACs) and layer 3 (IP) addresses virtualization techniques are available and can bring management issues to the table.
3.2 Impact of server and storage virtualization trends on the data center network
The current landscape in the data center is characterized by several business issues and leading edge technology trends. From a business perspective: New economic trends necessitate cost-saving technologies, such as consolidation and virtualization, like never before. Mergers and acquisition activities bring pressure for rapid integration, business agility, and fast delivery of IT services. The workforce is getting increasingly mobile; pervasive access to corporate resources becomes imperative. Data security is becoming a legal issue. Companies can be sued and fined, or incur other financial damages for security breaches and for non-compliance. Energy resources are becoming a constraint and should be considered a requirement for an enterprise-grade network design. Users dependency on IT services puts increasing pressure on the availability and resilience of the infrastructure, and competition and on-demand for real-time access to services and support.. Those business drivers are causing new trends to develop in data center architectures and the network must adapt with new technologies and solutions, as follows: The new landscape brings the focus back to the centralized environment. The distributed computing paradigm may not always match the needs of this dynamic environment. Virtualization and consolidation together put more traffic on single links since the one machine-one application correlation is
152
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
changing. This can cause bottlenecks and puts pressure on the access layer network. For example, the End of Row (EoR) model may be unfit to sustain this new traffic pattern and so the Top of Rack (ToR) model is rapidly becoming attractive. The access switch sprawl can be effectively managed by consolidating several access nodes into a bigger, virtual one. Figure 3-2 shows the consolidation and virtualization of server and storage.
SRVR SRVR SRVR
NIC
NIC
NIC
NIC
NIC
NIC
LAN NIC/SW
SAN HBA/SW
SAN SW SAN SW
SAN SW SAN SW
Virtualization brings a new network element into the picture. The virtual switch in the hypervisor cannot always be managed efficiently by traditional network management tools, and it is usually seen as part of the server and not part of the network. Moreover, these virtual switches have limited capabilities compared with the traditional network hardware (no support for multicast and port mirroring, plus limited security features). However, virtual switches are becoming much more capable over time. Hypervisor development is opening up new opportunities for more feature rich virtual switches. Figure 3-3 on page 154 shows networking challenges due to virtual switches and local physical switches.
153
7928ch03.fm
SRVR
SRVR
SRVR
C I N
C I N
C I N
C I N
C I N
SAN HBA/SW
LAN L AN AC CE SS ACCESS SW SW
SA N SW SAN SW
LAN LA N AC CE SS ACCESS SW SW
SA N SW SAN SW
The server and its storage are being increasingly decoupled, through SANs or NAS, thus leading to the possibility of application mobility in the data center. The 10 Gb Ethernet is reaching its price point for attractiveness. This increased capacity makes it theoretically possible to carry the storage traffic on the same Ethernet cable. This is also an organizational issue because the Storage and the Ethernet networks are usually designed managed by different teams. Figure 3-4 on page 155 shows storage and data convergence and its impact on the network infrastructure.
154
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
SAN SW
SAN SW
SAN SW
SAN SW
SAN SW
Figure 3-4 Storage and Data Convergence impact on the network infrastructure
The consolidation of the distributed IT environment into single distributed data centers struggles with an increasingly mobile workforce and distributed branch offices. Distance, and thus latency, increases for remote access users. New technologies are available to address this issue and bring better response times and performance to this scattered user base. This becomes even more critical for applications that exchange significant amounts of data back and forth, because both bandwidth and latency must be optimized on an application and protocol level, but optimizing network utilization is not usually a priority during application development.
155
7928ch03.fm
Figure 3-5 shows how increasing distance may impact application response time. This diagram does not show security layers which also add end to end latency.
Designing a traditional data center network usually implied deploying firewalls between the users and the applications, but this is rapidly changing because network and security must be designed together with a comprehensive architecture in order to ensure that security appliances do not decrease end to end performance. The high-level design should include secure segmenting based on business needs and priorities. Security services should be coupled with the applications even if these are mobile across data centers. Additional technologies can be used to provide this secure segmentation, for example VRFs. At the same time remote users must be authenticated and access is to be controlled and virtualized to enter the IP core and use the centralized, virtualized IT resources. Figure 3-6 illustrates access control and defined policies, as well as support for remote and mobile application access.
156
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
SRVR
SRVR
LAN NIC/SW
USER
157
7928ch03.fm
Server virtualization patterns that allow many applications to run concurrently, yet independently, on a single physical server. The increasing availability of WAN services at lower prices. Application servers can be positioned in a central data center, accessed across a WAN and still be responsive enough for business requirements. Server consolidation and virtualization combined clearly increase the average traffic volume per server dramatically, thus increasing the risk of network congestion. The main effects of consolidating several geographically dispersed data centers into a single logical one are listed here: Infrastructure scalability requirements become more important. The architecture of a single, very large data center is very different from the one supporting multiple, smaller data centers. Single faults affect a larger part of the infrastructure. Hardware problems occurring on a single physical machine may affect multiple services that rely on the logical servers inside the physical host. From a network point of view, the forwarding plane becomes more complex. Additional resiliency is required using, for example, load sharing solutions to avoid chokepoints and single points of failure. In addition to recentralization of the IT infrastructure, users now also need to access these resources from increasingly scattered locations. Laptop computers, mobile devices and wireless connectivity (such as WiFi and cellular networks) allow users to meet their business requirements in an increasingly connected-yet-dispersed fashion. Users that access centralized applications from branch offices or remote locations expect LAN-like performance regardless of the connection they are relying on. However, this requirement can be difficult to fulfill and also very expensive if it is addressed only by acquiring more WAN bandwidth. This approach may also be ineffective in some cases due to, for example, the TCP transmission window mechanism. Furthermore, with applications growing richer in content and application developers often overlooking the need to optimize network consumption, it becomes challenging to offer LAN-like performance to remote users accessing centralized IT applications. This is especially true because low priority applications share the same WAN resources used by business-critical applications and congestions may occur. This is why it is very important to classify applications according to a consistant QoS model and apply it end to end.
158
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
This scenario also impacts the latency that users experience when accessing applications, given that data travels over a longer distance and more appliances are involved in the traffic flow. This is particularly harmful for applications that rely on a large number of small packets travelling back and forth with a small amount of data (chattiness). In addition to WAN slowing down the application access to users, another critical requirement in this scenario is the availability of the centralized IT infrastructure. A single fault, without a proper single point of failure avoidance and disaster recovery (or even business continuity) strategy, would affect a very large population with a dramatic negative impact on productivity (service uptime and revenues are increasingly tightly coupled). Increased application availability can be provided by: Leveraging increased HA features of single systems and components From a network perspective, this also means being able to upgrade software and hardware components without turning off the entire system. This becomes even more important if the system leverages virtualization technologies because several logical systems may experience downtime. Providing redundancy at every level in the network Interface level (NIC teaming, NIC sharing, and so on) LAN Link level (either spanning tree-based or with an active/active model) Device level (node aggregation) WAN Access level (Dual/Multi-homed Service Provider Connectivity that can be implemented using BGP or dedicated appliances) Site level (disaster recovery, global load balancing). This can also be implemented with an Active-Active scheme (DNS-based or using dedicated devices) to maximize resource utilization. As consolidation initiatives arrive at a concrete implementation phase, bandwidth costs may still be an important item in the business case for emerging economies. IT managers might have the need to properly manage their connectivity to make sure critical-application traffic gets priority over traffic that can be delayed or that is not time sensitive. In more mature markets, prices for bandwidth are decreasing but traffic shaping is still a possible method to approach these situations and prevent slow response times during high usage conditions and avoid increasing bandwidth for some locations including the data center links. Concurrent access to centralized applications and sources of data from several branch offices, from an application resources perspective, might not pose any problems, but when it comes to managing data center connectivity for the WAN, in order for the properly planned, this even if the commited rate of the connectivity itself looks to be enough for average traffic. Traffic shapers have to be considered when the need is to allocate precisely the right amount of bandwidth resources to the right service and thus have a mean to guarantee the
159
7928ch03.fm
operationals level agreement with the business units. From a data center perspective this implies that a certain slice of bandwidth is dedicated to the single branch office to deliver the mission critical services. Classification of traffic needs to be done in advance as the number of applications that travel the wide area network (WAN) can be very vast and traffic shapers can also limit the bandwidth of some kind of applications (e.g. ftp transfer, p2p applications) so that all services are available but performances can vary over different conditions. This approach gracefully avoids services disruption by eliminating services from contending for the same network resources. Classification needs to evolve with traffic patterns and applications that change over time. This aspect is critical for the management of the solution and does not have to turn into something overwhelming for the IT department. A sound financial analysis is advisable in order to understand what should be done strategically over time to guarantee the service levels to all branch offices and other external users.
Another technique that can be used to mainly optimize bandwidth resources is WAN optimization. WAN Optimization Controllers (WOCs) are applicances deployed in a symmetrical fashion, both in the data center and in the remote
160
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
office, creating a tunnel between the sites that need to be accelerated. The goal of WOCs is to use bandwidth more efficiently and to free up some space for new applications or flows to be delivered over the wide area network. WOC techniques can also reduce latency at the session level, not at the link level which depends on other factors the WOC cannot change. This is important as WOCs do not accelerate traffic on the link but are appliances that sit in the middle of the session to understand the packet flows, avoid data duplication and combine and reduce packets of the application sessions in order for a single IP datagram sent over the tunnel to carry more packets of the same or different sessions, thus reducing the overall latency of a single session. An additional approach is to deploy application front-end special-purpose solutions known as Application Delivery Controllers (ADCs). This asymmetric WAN Optimization approach is provided by deploying feature-rich and application-fluent devices that are capable of handling Layer 4 through Layer 7 information, together with the traditional load balancing features like server health monitoring and content switching. As stated above, this layer sits in front of the application to mainly offload servers from: Terminating the SSL sessions, leaving more processor cycles to the application Serving repetitive and duplicate content to the users by instructing the users browser to use the dynamic or static objects that will be cached locally. This approach has the benefit of lowering the use of disk, memory, CPU resources in both a virtualized or traditional environment. For example, the consequence in a virtualized data center is more VMs per physical server due to reduced resource utilization. Another aspect of ADCs is that they are tied with the application layer and, as application traffic, are adapted to VM migrations, even between data centers or to the cloud. Downtime and user disruption are therefor avoided by accelerated data transmission and flexibility in the expansion of the overall service delivery capacity of the service delivery. Other features to consider when approaching ADCs solutions are compression that allows less data to be sent over the WAN and TCP Connections management. Note that this process can be transparent to both server and client applications, but it is crucial to redirect traffic flows in real-time if needed. It is also important that the choice of where to deploy the envisioned solution is meeting the performance and above all cost objectives. These functions, just like the other network services, can be implemented using different deployment models (see 3.4.5, Network Services Deployment Models on page 175 for more details). For example, software-based ADCs are entering the market from different vendors. A hybrid deployment model can also be used, leveraging hardware
161
7928ch03.fm
platforms for resource-intensive operations such as SSL termination, and offloading other functions such as session management or load balancing to software-based ADCs. Each of the described approaches can address different aspects of the overall design and does not necessary mean that one excludes the other. In fact the combination of features can efficiently solve the identified issues in bandwidth management.
162
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
oversubscription is an issue with this approach since it is not dealt with in a rack by rack fashion and it is difficult and expensive to scale if the number of racks exceed a certain threshold. On the other hand there are less switches to manage compared with the ToR model, but cabling is more difficult and less flexible. Emerging models that mix the best of the ToR and EoR worlds are emerging (for instance virtual End of Row switching- vEoR). For example using unmanaged (that can act like remote line cards) or managed appliances (fixed form factor switches that can be logically aggregated) in dedicated racks instead of chassis-based switches. This approach aims at minimizing the number of management points and to create a single, big, logical switch that can satisfy the scalability and flexibility requirements driven by consolidation and virtualization. Blade Switching - Using blade center modules that perform switching functions for the Blade Servers and aggregate traffic all the way to the data center network core. These switches have most of the features of the ToRs and EoRs but are specific for a certain blade server chassis so theres limited flexibility and also less choice. Virtual Switching (vSwitching) - Using part of the hypervisor resources to perform Ethernet switching functions for intra-server VM-to-VM communications. These techniques have limited functionalities compared to physical switches and are typically hypervisor-specific, as shown throughout Chapter 2, Servers, storage and software components on page 31. The latency induced by having multiple switching stages (and also, depending on the design, perhaps multiple security stages) from the server to the data center network core and and the availability of high port density 10 GbE core switches is driving the trend towards the delayering of the data center network that can achieve both significant cost benefits (less equipment to manage) and service delivery improvement (less latency turns into improved application performance for the end users). On the other hand servers equipped with 10 GbE interfaces induce higher bandwidth traffic flowing to the core layer so oversubscription on the access layer uplinks may become detrimental for the service performance. The development of IEEE standards for 40 GBE and 100 GBE will alleviate this potential bottleneck going forward and help to simplify the data center network design.
163
7928ch03.fm
Because of the increasing virtualization of servers, the availability of more powerful multi-core processor architectures, and higher bandwidth I/O solutions, it is obvious that the access layer will face huge growth in terms of bandwidth demands. On the other hand, this trend faces limitations in highly virtualized environments because I/O operations experience significant overhead passing through the operating system, the hypervisor and the network adapter. This may limit the ability of the virtualized system to saturate the physical network link. Different technologies are emerging to mitigate this. For example, bypassing the hypervisor using an a Single Root I/O Virtualization (SR-IOV) adapter allows a guest VM to directly access the physical adapter, boosting I/O performance significantly. This means that the new network access layer becomes the adapter, which incorporates switching capabilities. However, that there is no one-size-fits-all solution. For example, there may be different requirements for north-to-south traffic (client to server) and east-to-west traffic (server to server). Because IT services are delivered using an increasing number of different servers that can reside in the same data center (favored by application development methodologies like SOA), the interaction between these hosts (east-to-west traffic) may play a fundamental role for the overall performance, even more important than traffic patterns coming in and out of the data center (north-to-south traffic). Sometimes the I/O subsystem can be the bottleneck of systems virtualization performance, but for other CPU-intensive workloads (for example, analytics), the I/O subsystem performance may not be an issue. The outcome of these trends is that the access layer is evolving towards Top of Rack (ToR) topologies rather than pure End of Row (EoR) topologies. A ToR topology is more scalable, flexible, and cost-effective, and it consumes less energy than an EoR topology when dealing with the increased bandwidth demands at the access layer. In the long term it is desirable to have 10 Gbps capability as close as possible to the servers, especially when a converged access switch solution is installed at the access layer. A converged server can utilize the converged 10 Gbps adapter to burst large files over the Ethernet, if the SAN traffic is low, and the downstream Ethernet network must be able to properly deliver the required bandwidth throughout the network. Because the traffic pattern varies from data center to data center (actually it is the architecture of the applications and how users are accessing the applications
164
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
that will generate requirements on the design of the access layer) it is difficult to provide baselines to develop a design. The modular ToR approach can provide a sufficient solution for todays bandwidth demands. Because bandwidth can be added where it is needed, it is important that port utilization (both server and uplink ports) is monitored with proper management tools so that bandwidth can be added to the data center infrastructure before bandwidth constraints are recognized by end users. The growth of traffic volume generated at the access layer is tightly coupled with the increase of latency, which is originated by having additional switching stages (virtual switches or blade switches, for example). To lower the latency introduced by the data center network, different architectural patterns are emerging. One possible option is to adopt a two-tiered network by collapsing the core and aggregation layer. Another alternative is to adopt a cut through ToR switching architecture for the access layer. This switching architecture in the Ethernet world is not something new (it has been popular for Infiniband networks), but it is gaining traction again due to the reduced latency brought by these kind of switches. Manufacturers are releasing devices with this architecture, which has differences compared to the traditional Store and Forward model: Cut through switches forward the Ethernet frames as soon as the Destination MAC address is read. No error correction/CRC means that a requirement for cut through switches is that a very low error rate is needed because otherwise physical or data link layer errors will be forwarded to other segments of the network. The hosts NIC must perform the discard function instead. Virtual Output Queueing architectures are needed to avoid head of line blocking, and thereby avoid having the cut through switch behave like Store and Forward due to port contentions. Using windowed protocols such as TCP can exacerbate this situation by increasing response times.
165
7928ch03.fm
Node aggregation
Node aggregation techniques are similar to server clustering in some ways because many physical entities are logically grouped as one, single entity. The network impact of this approach is that there are fewer logical nodes to monitor and manage, thus allowing you to disable the spanning tree protocol aggregating links across different physical switches.. At the same time, risk is reduced because nodes are still physically redundant, with no single point of failure. In fact the network becomes a logical hub-and-spoke topology with a proprietary control plane replacing the spanning tree protocol. Figure 3-8 on page 166 illustrates node aggregation techniques.
Aggregation layer
2:1
Access layer
These techniques can be separated into two types: 2:1 aggregation This is performed at the core or aggregation layer, providing an extended backplane and simplifying dual-homing topologies. This allows for active/active configurations and link aggregation techniques. The logical network becomes a point-to-point topology, so spanning tree protocol can be disabled. By using a logical hub-and-spoke topology you can use all the available bandwidth with multichassis link aggregation techniques that logically group multiple links across different appliances into a single one.
166
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
This also allows to disable default gateway redundancy (e.g. VRRP), providing the same level of resilience and availability. The L3 processing and routing tables are also simplified because they are shared by two physical nodes. N:1 aggregation This is performed at the aggregation or access layer and allows many physical switches (especially Top of Rack, or ToR) to be seen as one. This is very important because it simplifies the ToR model, which can add a high degree of complexity even if it better suits the need for an easy-to-manage yet dynamic data center. N:1 aggregation can be used in the same physical rack by stacking and interconnecting many access servers. Or it can be used across different racks to virtualize the whole access/aggregation layer. Furthermore, with this approach the aggregation layer can be reduced if not eliminated. More traffic can be confined within the access layer. Sample scenarios of this virtualized environment are many ToR switches that act like a single ToR, or multiple blade switches that act as a single ToR. In these scenarios, multiple uplinks can be combined into an aggregated link to provide more capacity with the same resilience. If one link fails, the overall capacity is reduced but the overall availability is preserved. Note that the classic L2 topology (Spanning Tree based) is replaced by a private topology that requires some degree of understanding in order to properly design the network, and it may not always be possible to extend these functionalities in a multi-vendor network environment.
Node partitioning
Node partitioning techniques allow you to split the resources of a single node while maintaining separation and providing simplified (and thus improved) security management and enforcement. Partitioned nodes go beyond virtualizing the forwarding and control planes, such as you have with VLANs. This type of virtual node also virtualizes the management plane by partitioning the nodes resources and protecting them from one another. In this scenario a multitenant environment can be implemented without needing dedicated hardware for each tenant. Another possible use case of this technology is to collapse network zones with different security requirements that were physically separated on the same hardware, thereby improving efficiency and reducing the overall node count. This model is analogous to the server virtualization techniques in which a hypervisor allows multiple operating system instances to run concurrently on the
167
7928ch03.fm
same hardware. This approach allows for cost-effective, responsive, and flexible provisioning but needs a low-latency network for location-independence. Typical deployment scenarios are virtual routers, switches, and firewalls that can serve different network segments without having to reconfigure the whole environment if new equipment is added inside the data center. This can be extremely helpful for firewalls because these virtual firewall solutions reduce the need for dedicated in-line firewalling that can become bottlenecks, especially if the link speed is very high.
168
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
also extended to the virtualized data center layer to ensure compliance with the enterprise security framework. Correct level of traffic performance by domain/service/type: interconnecting data centers results in even more business transactions contending for network resources. A possible methodology to adopt to address this aspect could be the following: gain complete visibility of all the traffic flows; guided by business requirements, set the right level of priorities mainly from an availability and performance point of view; enable solutions that can adapt network resources (e.g. bandwidth on WAN links) depending on the agreed needs (e.g. traffic managers) or that can reduce traffic on the WAN links by avoid deduplication of data patterns (WAN Optimization Controllers). monitor the overall service performances
There are three major areas that need to be considered in order to plan the interconnection between Data centers: Layer 2 extension: some applications need to rely on layer 2. If high availability or workload mobility are among the drivers for interconnecting data centers, the layer 2 of the single data center needs to be extended to all the data centers that are relevant to the service. Layer 3 extension : some applications (i.e. backups) can rely on layer 3 and therefore adequate IP connectivity needs to be provisioned between data centers. Storage Area Network extension: data within the enterprise is considered a critical asset. Not only the replication, but the availability of data is essential to achieve significant efficiency in the overall journey of building a single distributed data center. Figure 3-9 demonstrates a sample scenario.
169
7928ch03.fm
Examples of the drivers behind layer 2 extensions, creating larger bridged domains, are: The need for applications that use "hard-coded" IP parameters that cannot be easily modified to be available from different locations. That is, the same subnet needs to be available in different remote locations. Data center consolidation initiatives or the feasibility of outsourcing models can result in a portion of the server farm to be moved to another location Layer 2 heartbeat mechanism to deploy high availability server clusters Virtual machine mobility There are network, server load balancing, and security devices that require high availability, as well as heartbeat and First Router Hop Protocols that need bridged connectivity. There are several different approaches available today for extending layer 2 networks and others that are in draft version in the standardization bodies. Each approach needs to be carefully evaluatedin terms of how they meet different requirements such as availability features, implementation processes, cost and management. Here, we outline three approaches:
170
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
1. The first approach is to create a switching domain, made of multiple switches that operate as a unified, high bandwidth device. Multi-Chassis-Link aggregation techniques are an example of this approach: switches that are in different data centers can be linked over optical links (Over dark fiber or a protected WDM based network) over large distances. This technique is especially suitable for dual site interconnection and security on the link can be added as an option. Also the technique has the additional feature of leveraging link aggregation (formerly known as 802.1ad, which has recently been moved to a stand-alone IEEE 802.1AX-2008 standard). Another important aspect when designing the interconnection between data centers at this level is the speed of links and related standards. With the objective of providing a larger, more durable bandwidth pipeline, IEEE have announced the ratification of IEEE 802.3ba 40Gb/s and 100Gb/s Ethernet, a new standard governing 40 Gb/s and 100 Gb/s Ethernet operations that considers distances up to 40 km. For data centers that need to be interconnected within this distance, this option can extend the core switching infrastructure. 2. The second major approach is to deploy Virtual Private LAN Service (VPLS ). The primary purpose of VPLS is to extend Layer 2 broadcast domains across WANs/Metro Area Networks (MANs) using a Multiprotocol Label Switching (MPLS) backbone. Enterprises that decide to deploy VPLS have found the deployment easy, a low latency variation, and a very scalable solution. In fact VPLS can scale to a very large amount of MAC addresses. Additional features that can be deployed aim to manipulate virtual local area networks (VLANs) within a bridge domain or a virtual private LAN service (VPLS) instance. VLAN translation is an example of such features. 3. The third approach is based on internet standards that are not yet finalized. Cisco sponsored Overlay Transport Virtualization (OTV) and the efforts of the IETF L2VPN working group (also known as E-VPN) are examples of potential alternatives to consider in the short-medium term. An IETF draft has been submitted to describe OTV. At the time of this publication, the draft can be found at the following link: http://www.ietf.org/id/draft-hasmit-otv-01.txt OTV can be considered as a very recent alternative for extending Layer two domains as it relies on a "MAC in IP" technique for supporting L2 VPNs over, mainly, an IP infrastructure. OTV can be positioned at the L2/L3 boundary at the DC aggregation and promises operational simplicity. Additional requirements that need to be carefully examined when considering OTV are: MAC addresses scalability VLAN scalability Convergence time Additional VLAN features (e.g. VLAN translation VLAN re-use)
The IETF L2VPN working group is the same group that is working on VPLS.
171
7928ch03.fm
While alternatives are available and the business relevance is increasing today, there can be some aspects that need to be considered and addressed in all the components just stated but that are particularly severe for Layer 2 extensions: Latency - Since there is no market solution today that can address the latency derived from the speed of light, this aspect might not always be an issue when data centers are within shorter distances. Latency becomes an important factor to be considered when data centers are to be "actively" interconnected over longer distances. Bandwidth - Workload mobility and data replication require high-speed connectivity that is able to support a large amount of traffic. Moving virtual machines across data centers means that the snapshot of the virtual machine needs to cross the links between the data centers. If the end to end link between the sites does not comply to the vendor specification, the switch off of a VM and the restarting of a new one can fail. The real challenge is data synchronization across data centers, with the objective of having the most recent data as close as possible to the application. In both situations, techniques such as WAN Optimization Controllers can help in reducing the amount of bandwidth by avoiding deduplication of data. Quality of Service complemented by traffic shapers could also be an alternative for preserving the overall operational level agreements (OLAs). Organization (ie: DC team and WAN team)- Alternative selection criteria is also based on the enterprise organization structure and the skills associated within each alternative. For example, the fact that VPLS is based on an MPLS infrastructure implies that skills have to be in place, either internal or external to the organization if VPLS will be selected as a Layer 2 extension. Data flows of multi-tier application - When data centers are servicing users, the changes in traffic flows that are built within the overall network infrastructure need to be well understood and anticipated when moving a service. Especially when the service is built upon a multilevel application approach (i.e. front end service - application service - data service), moving single workloads (i.e. front end service layer) from one data center to another might result in increased latency and security flow impacts in the overall business transaction. In many situations, security needs to be applied within the flow and this layer needs to be aware of transactions that are going through different paths that must be created or that the path is changing within interconnected data centers. Fast Convergence - The longer convergence times of spanning tree protocol (STP) are well known within the industry. When extending layer 2 within data centers, the simple STP approach can result in unpredictable results as convergence is concerned and resynchronization of a large STP domain can take too much time for business requirements. Because MPLS relies on the underlying IP network, some critical configuration parameters of both IP and MPLS have to be analyzed. Moreover, the convergence time that an IGP,
172
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
EGP, and First Hop Router Protocol (for example, VRRP) with default parameters provide is usually not suitable for today's business requirements. In an MPLS network, two main types of fast convergence architectures can be used: MPLS Fast Rerouting Traffic Engineering (FRR TE) IP Fast Rerouting (IP FRR) In order to achieve a sub-second convergence time, the following aspects have to considered: Failure detection Label Distribution Protocol session protection mechanisms IGP tuning BGP tuning First Hop Router Protocol (FHRP) tuning System scheduler tuning Scalability(MAC/VLANs/Data Centers)- Large enterprises have the need to extend a certain number of VLANs across data centers. Also, the number of MAC addresses that the network must be aware of can grow significantly in the order of thousands of entries. Each node at the edge should have the information of where the different MAC addresses are in order to send the frames to the correct data centers.The number of sites that need to be connected is also relevant to the technology choice. For example, a large number of sites that are connected can have a layer 2 domain that is too big for STP to handle. Policy Synchronization - The security framework includes policies related to how the services are protected. Interconnecting data centers are an opportunity for policy enforcement to match in all locations. On the other hand, it represents a challenge where enterprises strive to create an homogeneous set of policies and do not necessarily own the infrastructure but they are renting it or they are outsourcing a service or part of it. Also encryption and security play an important role: the merger of layer 2 and layer 3 can have important implications on the current security framework. When the MPLS services are outsourced to a service provider, security from the service provider has a very important role in the chain. Encryption can be an option to think about. In the same conditions PE routers are shared among mutiple clients. Both aspects and in general, internal audit requirements, might impact the overal solution approach. Management - Determining the impact of deploying new technologies and other IT related services and analyzing how the management infrastructure will support the related architectures aims at network efficiency and overall reduction in the cost of ownership. Understanding the gaps in the current
173
7928ch03.fm
network mangement processes or tools is advisable. For instance, MPLS connectivity management might address the need for dynamic and automated provisioning, as well as resource mangement for tunnels. Also, troubleshooting differs from that for a routed network, therefore diagnostic capabilities are an important aspect of all the management processes. Storms or loops - A major difference between a router and bridge is that a router discards a packet if the destination is not in the routing table. In contrast, a bridge will flood a packet with unknown destination address to all ports except the port it received the packet from. In routed networks, IP routers decrease the TimeToLive value by one at each router hop. The TTL value can be set to a maximum of 255, and when the TTL reaches zero, the packet is dropped. This of course does not solve loops consequences in a routing domain but considering that in a bridged network there is no such mechanism, the available bandwidth can be severely degraded until the loop condition is removed. When interconnecting data centers at layer 2, the flooding of broadcasts or multicasts or the propagation of unknown frames needs to be avoided or strictly controlled so that, a problem in one data center is not propagated into the another one. Split brain - This situation might occur if both nodes are up but there is no communication (network failure) between sites. Of course this situation needs to be avoided by carefully planning the redundancy of paths at different levels (node, interface, link) Cloud computing models imply additional thinking for the interconnection of data centers. The alternatives offered by the market -private, public or hybrid cloud all have in common the need for a careful networking strategy to ensure availability, performance and security. Especially in multi-tenant environments where service providers offer shared services, it is essential to consider which aspects of service levels and security have strong relationships with the networking and therefore must be addressed in a complementary manner. Interaction with cloud models may imply different locations for the applications and the related connectivity with respect to end-users. The techniques behind the interconnection with the Cloud should not only deliver reliable and secure data between different points, but must, above all, adapt to the variability of resources that can be dynamically and automatically relocated geographically and in real time, inside and out the cloud. An example is the availability in the cloud of additional processing power to meet peak in the workloads on the same day or the temporary allocation of additional storage for specific projects. The network must be designed with flexibility in the ability to modify and extend the service to the rest of the cloud in a way that is completely transparent to end users.
174
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
In designing and implementing Cloud models, the networking component should be built in synergy with and sustain all other system components of Cloud computing, such as applications, processes, delivery and automation systems.
175
7928ch03.fm
example, some functions that are resource-intensive still need to be performed in hardware to avoid performance bottlenecks. On the other hand, application delivery functions can be broken down further and then split depending on the best deployment model for each function. Using application delivery as an example, SSL termination may be a better fit for a hardware platform, while other more application-specific functions may be virtualized. The characteristics of each approach can be expressed in the following terms: Performance: Generally indicated in how many packets are processed per second by the appliance or other similar performance metrics. This can impact the performance of the entire network (if the appliance becomes a bottleneck) or the hypervisor, in the case of a software appliance. Provisioning effort: Defined as the time and effort required to get the network service up and running. An appliance approach needs to consider all of the aspects of a dedicated hardware in terms of hardware provisioning and set up that might be shared among other services with software-only models. Scalability: In networking terms, the ability to grow incrementally. Scalability can be horizontal (scale out model), adding network nodes in the systems or vertical (scale up model), adding resources to a single network node. Management integration: Defines the possibilities in terms of systems management integration. Functional richness: Defines the different functions a single appliance or cluster carries out. Point of enforcement: defines where network services can be deployed. Location: Defines where the appliance is located and with how much flexibility it can be relocated. Cost: Defines the required financial impact of each deployment model. Physical hardware appliances have a broader applicability because they perform their functions as long as the protocols they rely on are running and supported throughout the network. Some applications that cannot be virtualized or are chosen not to be virtualized still need to rely on this type of deployment model for network services. Performance can be very high, depending on the traffic load for each appliance, but it may come at a very high price. Moreover, physical appliances may incorporate several functions that are not required in most deployments. Virtualized appliances, in contrast, are generally based on modular hardware, with different blades or service modules that can be deployed as needed for specific functions or to allow growth with a scale up model.
176
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
Virtual Appliances are a better fit for cloud computing networking infrastructures because of their infinite scalability and the ability to closely integrate with the virtual infrastructure because the services can easily be moved to another physical host.
177
7928ch03.fm
As it can be seen in Figure 3-10, network security is only a component of the overall framework and can be seen as the enforcement at the network level of the enterprise security policies. In a highly virtualized data center, for example, security boundaries are no longer defined only by physical appliances but those move inside the server and storage platforms (see Figure 3-11). Also the server-to-server communications often occur within the same physical box and traditional network security appliances such as firewalls cannot enforce the security policies within the hypervisor. This new environment shift is depicted in Figure 3-11. A functional view of network security can be categorized in the following areas: Remote access (SSL/IPSec VPN) - both site to site and to allow remote users to access Enterprise Resouces outside of the perimeter. Threat management - Network level - network boundaries security enforcement (firewalling, IPS, IDS, Anti-Spam) layers 2 through 4 in the OSI model. Threat management - Application level security enforcement (application firewalling) layer 7 in the OSI model. Functions like URL filtering fall into this category.
178
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
Resource access management and identity management to control and audit who is entitled to access the network resources in order to monitor, manage and configure the different appliances.
Before After
Physical NICs
Virtual Switch
Virtual Switch
Management VM
Physical Network
Virtual Network
Bear in mind that all these functional components can be deployed using the different models that have been presented in the previous section. We will now briefly describe some relevant IBM products in the virtualized data center network environment. The IBM security product portfolio covers professional and managed security services and security software and appliances. To learn about comprehensive security architecture with IBM Security, see Enterprise Security Architecture using IBM ISS Security Solutions, SG24-7581.
179
7928ch03.fm
180
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
Management
SVM
Policy
VM
VM
Applications
Response Engines
Applications
Applications
OS
Hardened OS Kernel
OS
OS
Kernel
Kernel
Kernel
VMsafe
Hypervisor Hardware
Each SVM on virtualized servers is managed and configured through Site Protector (SP), which is the central manager of SVM. VSS has the following functions: Inter-VM IPS and Firewall Virtual Machine Observer (VMO) Antirootkit (ARK) Network Access Control (NAC) Discovery License and Update Management (LUM)
181
7928ch03.fm
With DV filters, all NICs of VMs are monitored. An overview of a DV filter is shown in Figure 3-13. A DV filter does not require any configuration changes to the internal network.
SVM
IPS
VM
VNIC
VM
VNIC
VNIC
vSwitch
DV Filter
DV Filter
vSwitch
vSwitch
PNIC
PNIC
Because of performance and resource consumption considerations, accelerated mode can be configured. Unlike a DV filter, accelerated mode monitors vSwitch inbound and outbound traffic. Thus, inter-VM traffic within vSwitch is not monitored. This mode is illustrated in Figure 3-14. In both methods, the same IPS function is supported, but traffic coverage is different.
182
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
SVM
IPS
VM
VNIC VNIC
VM
VNIC
VNIC
vSwitch
vSwitch
PNIC
VSS also provides protocol access control like a normal IPS appliance from ISS. It filters network packets based on protocol, source and destination ports, and IP addresses. It can be enabled or disabled globally or for specific VMs. Because the VMware internal network does not support layer 3, this function does not support L3 security zoning, which normal firewall appliances can.
Discovery
The discovery module collects inventory information about VMs, such as operating system name and version, open TCP ports, and so on. A vulnerability check is not performed in this version.
Antirootkit (ARK)
A rootkit is a program that is designed to hide itself and other programs, data, and/or activity including viruses, backdoors, keyloggers and spyware on a computer system. For example, Haxdoor is a known rootkit of the Windows platform, and LinuxRootkit targets Linux OS. Generally, it is difficult to detect a rootkit that rewrites kernel code of OS. However, in a virtualized environment, it is possible that the hypervisor can monitor OS memory tables from outside OS. VMsafe supports a vendor tool that inspects each memory table on VM. VSS can detect malicious software inside OS by inspecting each kernel memory space.
183
7928ch03.fm
NIC sharing
The most basic of the new connectivity standards simply assigned an operating system to share available network resources. In its most basic format, each operating system has to be assigned manually to each NIC in the platform. Logical NIC sharing allows each operating system to send packets to a single physical NIC. Each operating system has its own IP address. The server manager software generally has an additional IP address for configuration and management. A requirement of this solution is that all guest operating systems have to be in the same Layer 2 domain (subnet) with each guest operating system assigned an IP address and a MAC address. Because the number of guest operating systems that could reside on one platform was relatively small, the MAC address could be a modified version of the NICs burned-in MAC address, and the IP addresses consisted of a small block of addresses in the same IP subnet. One additional IP address was used for the management console of the platform. Features to manage QoS and load balancing to the physical NIC from the guest operating systems were limited. In addition, any traffic from Guest OS 1 destined
184
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
to Guest OS 2 traveled out to a connected switch and then returned along the same physical connection; see Figure 3-15 on page 185. This had the potential of adding extra load on the Ethernet connection.
M a n a g e m e n t C o n s o le G u est O S 1 G ue st O S 2 G ue st O S 3 G u est O S 4
10
10
.1
.1
.1 .
.1
1 .2
10
.1 .
1 .3
1 0 . 1 .1 .4
1 0 .1 . 1 . 5
185
7928ch03.fm
Platform Management
Guest OS
802.1q Trunk
802.1q Trunk
Virtual
Figure 3-16 Virtual switching
Physical
A Layer 3 implementation of this feature allows traffic destined for a server that resides on the same platform to be routed between VLANs totally within the host platform, and avoids the traffic traversing the Ethernet connection, both outbound and inbound. A Layer 2 implementation of this feature effectively makes the physical Layer 3 switch or router a Switch-on-a-stick or a One-armed router. The traffic from one guest operating system destined for a second guest operating system on that platform traverses the physical NIC and the Ethernet two times. The challenge that this presents to the network architecture is that now we have a mix of virtual and physical devices in our infrastructure. We have effectively moved our traditional access layer to the virtual realm. This implies that a virtual NIC (vNIC) and a virtual switch all have the same access controls, QoS capabilities, monitoring, and other features that are normally resident and required on access-level physical devices. Also, the virtual and physical elements may not be manageable from the same management platforms, which adds complexity to network management.
186
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
Troubleshooting network outages becomes more difficult to manage when there is a mixture of virtual and physical devices in the network. In any network architecture that employs virtual elements, methods will have to be employed to enable efficient monitoring and management of the LAN.
NIC teaming
To eliminate server and switch single point-of-failure, servers are dual-homed to two different access switches. NIC teaming features are provided by NIC vendors, such as NIC teaming drivers and software for failover mechanism are used in the server systems. In case of the primary NIC card fails, the secondary NIC card takes over the operation without disruption. NIC teaming can be implemented with the options Adapter Fault Tolerance (AFT), Switch Fault Tolerance (SFT) or Adaptive Load Balancing (ALB). Figure 3-17 on page 188 illustrates the most common options SFT and ALB. The main goal of NIC teaming is to use two or more Ethernet ports connected to two different access switches. The standby NIC port in the server configured for NIC teaming uses the same IP and MAC address (in case of Switch Fault Tolerance) of the failed primary server NIC. When using Adaptive Load Balancing, the standby NIC ports are configured with the same IP address but using multiple MAC addresses. One port receives data packets only and all ports transmit data to the connected network switch. Optionally, a heartbeat signaling protocol can be used between active and standby NIC ports. There are other teaming modes in which more than one adapter can receive the data. The default Broadcom teaming establishes a connection between one team member and the client, the next connection goes to the next team member and that target and so on, thus balancing the workload. If a team member fails, then the work of the failing member is redistributed to the remaining members.
187
7928ch03.fm
NIC teaming with Switch Fault Tolerance One Ethermet port is active second port is standby Using one IP and M address AC
NIC team with Adaptive Load Balancing ing One Ethermet port receives and all ports transmit Using one IP and multiple MAC addresses
VRRP
VRRP
x
Ethernet Interface 0: 10.1.1.30 MAC A Ethernet Interface 1: 10.1.1.30 M A AC 10.1.1.30 MAC A
x
Ethernet Interface 0: Ethernet Interfaces 1 to N: 10.1.1.30 MAC B, C, N
Figure 3-17 NIC teaming with Switch Fault Tolerance Adaptive Load Balancing
188
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
The configuration of that generic platform is composed of the following components (see Figure 3-18 on page 190): Processor - general purpose, embedded, or specialized processing element Memory - general purpose or embedded PCIe Root Complex (RC) - one or more RC can be supported per platform PCIe Root Port (RP) - one or more RP can be supported per RC. Each RP represents a separate hierarchy per the PCI Express Base Specification. PCIe Switch - provides I/O fan-out and connectivity PCIe Device - multiple I/O device types, for example network, storage, and so on System Image - software that is used to execute applications or services, To increase the effective hardware resource utilization without requiring hardware modifications, multiple SI can be executed. Therefore, a software layer (Virtualization Intermediary - VI) is interposed between the system hardware and the SI. VI controls the underlying hardware and abstracts the hardware to present each SI with its own virtual system. The intermediary role for I/O virtualization (IO VI) is used to support IOV by intervening on one or more of the following: Configuration, I/O, and Memory operations from a System Image, and DMA, Completion, and interrupt operations to a System Image.
189
7928ch03.fm
SI 1
SI 2
SI 3
SI N
Processor Memory
PCI device
190
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch03.fm
Will the converged network traffic to each guest operating system oversubscribe the physical NIC? Do any of the guest operating systems require Layer 2 separation? Does each guest operating system require the same security or access restrictions? Who will be responsible for the end-to-end network architecture? Enterprise naming standards, MAC address standards, and IP address standards need to be applied to the virtual devices. Separate Layer 2 domains may exist in the virtual design. They need to adhere to the enterprise architecture standards. Are there organizational issues that need to be addressed pertaining to services management? Historically, different groups own management of the servers and the network. Will organizational roles have to be redefined or reorganized to allow efficient management of the dynamic infrastructure? What combination of tools will be used to manage the infrastructure? Currently, there are no tools that perform end-to-end network management across virtual, logical, and physical environments. Several network management systems have plans for this functionality but none have implemented it at this time. Can the selected platform and hypervisor support a virtual switch? If so, can this virtual switch support QoS? Does it participate in Layer 3 switching (routing)? What routing protocols? Does it do VLAN tagging? Can it assign QoS tags? All of these considerations will affect the network architecture required to support NIC virtualization.
191
7928ch03.fm
192
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
Chapter 4.
193
7928ch04.fm
Services portfolio, focusing on how Enterprise Networking Services can help clients navigate through the available options and achieve their business goals minimizing risks and future-proofing the technology choices.
194
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
195
7928ch04.fm
With a plan and a solid functional design, the probability of success is raised: a more responsive network with optimized delivery, lower costs, increased ability to meet application service level commitments, and a network that supports and fully contributes to a responsive IT environment.
196
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
considerations around continuing to support aspects of the legacy data center networking environment along with anything new.
197
7928ch04.fm
appealing or appealing," and 38% and 30% giving a similar rating for hybrid cloud computing and public cloud computing, respectively. By definition, cloud computing uses the network to gain access to computing resources-the network becomes the medium for delivery of enormous computing capability and hence plays a critical role. This critical role means that it is crucial to "get the network right" in terms of the right levels of performance, security, availability, responsiveness and manageability for the selected cloud computing deployment model. Enterprises that adopt cloud computing have a range of deployment models from which to choose, based on their business objectives. The most commonly discussed in the press is public cloud computing where any end user with a credit card can gain access to IT resources. On the other end of the spectrum are private cloud computing deployments where all IT resources are owned, managed and controlled by the enterprise. In between, there is a range of options including third-party managed, third-party hosted and shared or member cloud services. It is also possible to merge cloud computing deployment models to create hybrid clouds that use both public and private resources. Each cloud computing deployment model has different characteristics and implications for the network. That said, organizations select their cloud deployment models based on business requirements and needs. Whether a company opts to buy from a cloud provider or build its own private cloud computing environment, the enterprise network design must be revisited to validate that security, reliability and performance levels will be acceptable. In particular, the networking attributes for a private cloud data center network design are different from traditional data center network design. Traditionally, the data center network has been relatively static and inflexible and thought of as a separate area of IT. It has been built out over time in response to point-in-time requirements, resulting in device sprawl much like the rest of the data center IT infrastructure. The data center network has been optimized for availability, which typically has been achieved via redundant equipment and pathing, adding to cost and sprawl as well. The attractiveness of a private cloud deployment model is all about lower cost and greater flexibility. Low cost and greater flexibility are the two key tenets the network must support for success in cloud computing. This means the network will need to be optimized for flexibility so it can support services provisioning (both scale up and down) and take a new approach to availability that does not require costly redundancy. An example of this could be moving an application workload to another server if a NIC or uplink fails versus providing redundant links to each server.
198
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
Private cloud adoption models require a new set of network design attributes and these are demonstrated in Figure 4-1.
Traditional Network Design Optimized for availability Slow to adapt Costly Rigid, inflexible Infrastructure silos Dedicated Device sprawl Location dependence
Cloud Network Design Optimized for availability New approach to availability Cost more important Variable, metered Integrated infrastructure Shared Consolidation/Virtualization Location independence
In addition, network design can no longer be accomplished as a stand-alone exercise. The network design must encompass key requirements from other areas of the data center IT infrastructure, including server, storage, security and applications. The central design principles for the network must be based on present and planned application workloads, server and storage platform virtualization capabilities, IT services for internal and external customers, and anticipated schedules for growth and investment. From an organizational standpoint, the network design must address existing boundaries between the network, servers and storage and establish better interfaces that allow operational teams to work effectively in virtualized and dynamic environments. This requires the networking team to develop a greater operational awareness of server and storage virtualization technologies, capabilities and management.
199
7928ch04.fm
A newer protocol, IPv6, also developed by the IETF, is available. IPv6 makes offers a bigger address space to accommodate the budding need for new IP addresses. IPv6 also promises improvements in security, mobility and systems management. While IPv4 and IPv6 can coexist, ultimately the network of every organization that uses the public Internet will need to support IPv6 - once the available IPv4 address space for the world is exhausted, the ability to route of network traffic from and to hosts that are IPv6-only will become a requirement. No matter the reason a company introduces IPv6, deployment of IPv6 will take focus, planning, execution and testing and require operational changes, all of which will impact servers, other devices in the data center and applications in addition to the network itself. The Number Resource Organization (NRO)4, the coordinating body for the five Regional Internet Registries (RIRs) that distribute Internet addresses, says that as of September 2010 just 5.47% of the worldwide IPv4 address space remains available for new allocations (see Figure 4-2). When the remaining addresses are assigned, new allocations will be for IPv6 addresses only.
44 42 41 39 39
34 32 30
26 26
22 16 14
- ----Allocated 81.64%
Mar Jun Sep Dec Mar Jun Sep DecMar Jun Sep Dec Mar Jun Sep Dec Mar Jun Sep
Network equipment providers ship new equipment that is IPv6-ready. Telecommunications carriers can, in many cases, handle IPv6 traffic or have backbone upgrade projects underway. However, for a specific enterprise, the network, network-attached devices, the tools used to manage data centers and other IT resources, and the applications that are used to run the business need to be checked for their readiness to support IPv6 and plans made and executed
4 5
Internet Number Resource Status Report - September 2010, Number Resource Organization http://www.nro.net/documents/presentations/jointstats-sep10.pdf Global IP Addressing Statistics Sheet, American Registry for Internet Numbers (ARIN), September 2010 https://www.arin.net/knowledge/stats.pdf
200
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
for any upgrades. Also, each enterprise must determine its strategy for the coexistence of IPv4 and IPv6 - tunnelling, translation, dual stack or a combination as both protocols will need to be supported and interoperate until IPv4 can be retired. Depending on the geography, the IP address range allocated to an organization, and any workarounds in place to conserve IP addresses, such as Network Address Translation (NAT) or Classless Inter-Domain Routing (CIDR), planning and readiness for IPv6 may be considered prudent or essential. However, it is risky to wait until IP addresses and workarounds are exhausted or important customers or supply chain partners - or government bodies - require IPv6 as a condition of doing business and implementation must be hurried or performed in a makeshift fashion. This is especially true in an environment like the data center that has limited change windows and cannot afford unplanned downtime or performance degradation.
4.2 Assuring service delivery : The data center network point of view
As described in Chapter 1., Drivers for a dynamic infrastructure on page 1, the business requirements that impact the CIOs decision making process in a global economy significantly influence the IT environment and hence the data center network, which must support broader IT initiatives. The purpose of the data center network in this new context can be summarized with one simple goal : assuring service delivery. This objectives impact on the data center network can be broadly categorized along two main axis: Service Drivers The data center network must support and enable broader strategic IT initiatives such as server consolidation and virtualization, cloud computing and IT optimization. In this sense the network must ensure performance, availability, serviceability and shorten the time required to setup new services. The data center network must achieve the service delivery objectives at the right cost. In this sense , the data center network must be consolidated and virtualized itself in order to achieve the same cost savings (both on OPEX and CAPEX) gained through server and storage consolidation and virtualization.
Cost drivers
The structure of this section and the flow of the presented content is described by Figure 4-3 on page 202.
201
7928ch04.fm
Burgeoning solutions
These business and service objectives cause non-trivial challenges for data center network managers and architects since the traditional data center network infrastructure is not up to the task of satisfying all these new requirements. In fact many technology-related challenges arise in this context. In order to overcome the limitation of the data center networks static, traditional physical model, emerging standards and new architectural alternatives (described in 4.2.2, Burgeoning solutions for the data center network on page 208) can be exploited. The drawback of this path, however, is that while alleviating todays challenges, additional points of concern emerge. These will be presented in section 4.2.3, Additional points of concern on page 210. The only way to overcome these is to set high level but yet actionable guidelines to follow, as shown in figure above. These will be presented and discussed in section 4.2.3.
202
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
Data center perimeter security services WAN optimization services Virtualized L1 L2 & L3 network services L1,
In nfrastructure Se ervices
SAN
Manage ement
Access Security
Access Security
SAN
Storage
Memory CPU
Memory CPU
Storage
Data C t D t Center A
Data Center D t C t B
This diagram can be used as a guide to document client - server flows and impacts on the overall data center network infrastructure. In fact for clients accessing the data center services, the IT infrastructure looks like a single distributed data center. The geo-load balancing components takes care of redirecting the client to the most appropriate data center (if these are active / active) or it may redirect clients in case of disaster recovery situation. The Wide Area Network (WAN) acceleration layer optimizes the delivery of services for the remote clients. It is important to highlight that these functions (WAN optimization and perimeter security as well) may be different if the clients are accessing via the Internet or Intranet. So in order to document these differences, different diagrams should be used. The interconnection shown between data centers (there are two in Figure 4-4 but there can be more) has to encompass traditional Layer 1 services (for example, dense wavelength division multiplexing or DWDM connectivity for storage extension) and Layer 3 services (for example via multi-protocol label switching or MPLS) but also, a Layer 2 extension may be needed, driven by server virtualization requirements.
203
7928ch04.fm
The trend towards virtualized resource pools for servers and storage has a two-fold implication from a networking point of view: Some network-specific functions may be performed within the virtualized server and storage pools, as has been described in Chapter 2. For example access switching and firewalling functions may be performed within the hypervisor and the storage infrastructure may leverage dedicated Ethernet networks Some server and storage specific functionalities may significantly impact the network infrastructure that has to be able to support these virtualization driven features, such as virtual machine (VM) mobility. As shown in Figure 4-4, network specific functions such as local switching and routing, application delivery and access security may be performed at the server virtual resource pool level. While access security and switching are needed in most situations, application delivery may be optional for some specific application requirements (but can be performed both at a physical or at a virtual level). Infrastructure management (shown in green in the diagram) should encompass the whole infrastructure end-to-end and in this context the integration between system and network management becomes more meaningful, for example. Figure 4-5 shows examples of how the diagrams can be used in different situations : Example A shows a typical client-server flow where access security, application delivery and local switching functions are all performed in hardware by specialized network equipment Example B shows a client-server flow where switching is also performed virtually at the hypervisor level and access security and application delivery functions are not needed Example C shows an example of a VM mobility flow between Data Center A and Data Center B
204
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
Management
SAN
SAN
SAN
Management
SAN
Storage
Storage
Storage
Storage
Example A
Single Distributed Data Center
Data Center Edge Connectivity
Example B
SAN
Management
SAN
Storage
Storage
Example C
Given the complex situation we just described, we now provide a list of the most relevant of the networking challenges in the data center, from a technology point of view, that need to be addressed in order to obtain a logical, functional view of the infrastructure as shown in the figures above. Note that the order does not imply relative importance since the impact of these challenges, the specific environment, processes and requirements. The first set of challenges described here are related to the service drivers as they impact the data center because the network must support other strategic IT initatives. Layer 2 extension The widespread adoption of server virtualization techologies drive a significant expansion of the layer 2 domain, also bringing the need to extend layer 2 domains across physically separated data centers in order to
205
7928ch04.fm
stretch VLANs to enable VM mobility using technologies such as VMWare VMotion or POWER Live Partition Mobility. These are very challenging requirements to satisfy in order to achieve the service delivery objectives since they directly impact the Scalability (in terms of new MAC addresses that can be included in a Layer 2 domain) and Flexibility (in terms of that time and effort needed to deploy new services) of the Data Center Network. Control Plane Stability The stability of the control plane in a typical layer 2 data center network is typically controlled by the spanning tree protocol - STP or one of its many variants. This approach, however, does not have the robustness, flexibility and efficiency that is required to assure service delivery to the business. For example the typical convergence time required to recover from a link failure is not in synch with the needs of todays network-dependant and high paced business environment. This limitation hence has a huge impact on the availability and the reliability of the data center network. Another drawback of the STP is that in order to avoid loops in a layer 2 network, a tree topology must be enforced by disabling a subset of the available links. So usually 50% of the available links are idle and the efficiency of the capacity usage is sub-optimal at best. This can be mitigated by balancing the available VLANs across different STP instances but it goes without saying that being able to exploit 100% of the available capacity would be a huge improvement from a data center network performance point of view.
Obtain VM-level network awareness As described in detail throughout Chapter 2., Servers, storage and software components, the VM is the new building block in the data center and the importance of physical NICs for the network architecture fades when compared to the virtual networking realm inside the server platforms. On the other hand, it is difficult to manage both the virtual and the physical network environment with a consistant set of tools and orchestrate changes in an end-to-end fashion. This trend puts a lot of pressure on the serviceability and the manageability of
206
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
the data center network and can also impact its availability if changes are not agreed across different functional teams in charge of the infrastructure management. End to end network visibility Server consolidation and virtualization initiatives demand more bandwidth per physical machine and the same is true on a WAN scale when consolidating scattered data centers into fewer ones. This and the fact that it is very difficult to obtain end-to-end visibility of the network flows (different teams in charge of managing virtual resources inside servers, Blade switches, LAN switches, WAN routers) have the risky consequence that it is becoming increasingly more difficult to spot and remove network bottlenecks in a timely and seamless fashion. Clearly this has a significant impact on the performance and even the availability of the enterprise network if the quality of service (QOS) model is not designed and enforced properly.
Support new, network-demanding services New services and application that have very demading network requirements, such as video, cannot be accomodated easily in traditional data center network environments. So the challenge is being able to expoit these new technologies that are demanded by the business while minimizing the CAPEX expenditures that are needed and the risk of heavily changing the data center network infrastructure, whose stability has been traditionally enforced by it being static. These challenging new requirements impact both the performance and the capacity planning non-functional requirements.
The next set of data center network challenges are related to the cost drivers because the network must bring operating expenses (OPEX) and capital expense (CAPEX) cost savings, exploiting automation, consolidation and virtualization technologies leveraged in other IT domains such as storage and servers. Appliance sprawl Todays data center network environments are typically over-sophisticated and characterized by dedicated appliances that perform specific tasks for specific network segments. Some functions may be replicated so consolidation and virtualization of the network can be
207
7928ch04.fm
leveraged to reap cost savings and achieve greater flexibility for the setup of new services without the need to procure new hardware. This appliance sprawl puts pressure on the scalability and manageability of the data center network. Heterogeneous Management tools The plethora of dedicated hardware appliances has another consequence that impacts the operating expenses more than the capital expenses. In fact, different appliances use different vertical, vendor and model specific tools for the management of those servers. This hetereogeneity has a significant impact on the manageability and serviceability of the data center network. Another challenge that is driven by cost reduction and resource efficiency is the ability to share the physical network resources across different business units, application environments or network segments with different security requirements by carving out logical partitions out of a single network resource. The concern about isolation security and independence of logical resources assigned to each partition limits the widespread adoption of those technologies. So logically sharing physical network resources has a significant impact on the security requirements but may also limit performance and availability of the data center network.
208
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
time, the ability to manage the network as one single logical switch drives cost savings because the network is easier to manage and provision. Standards-based fabrics, leveraging emerging standards such as IETF TRILL (Transparent Interconnect of Lots of Links) and IEEE 802.1aq6 - Shortest Path Bridging(SPB). Early manifestations of these fabrics are appearing on the market and the basic idea is to merge the plug-and-play nature of a ethernet layer 2 network with the reliability and self configuring characteristics of a layer 3 IP network. The challenges that these standards aim to solve are the same as the vendor fabrics but also guaranteeing interoperability across different network equipement vendors. The price attractiveness of 10 gigabit ethernet (GbE) network interface cards and network equipment ports make it easier to sustain increased traffic loads at the access layer virtualizing and consolidating server platforms. The imminent standardization of 40 GBE and 100 GbE makes this even more true in the near future in order to be able to handle the increased server traffic at the core layer. Fiber Channel over Ethernet (FCoE) disks and network appliances make it possible to converge storage and data networks in order to achieve cost savings by sharing converged network adapters, collapsing Ethernet and Fiber Channel Switches in the same physical appliance and simplifying cabling. The FCoE standard is providing the foundation for the forthcoming converged data center fabric, but other protocols are also needed in order to deliver a fully functional end-to-end, DC-wide unified fabric. The IEEE standards body is working in the Data Center Bridging task group7 (DCB) to provide a framework to ensure lossless transmission of packets in a best-effort Ethernet network. New deployment models for network services such as multipurpose virtualized appliances and virtual appliances can be exploited in order to consolidate and virtualize network services such as firewalling, application acceleration and load balancing speeding up the time needed to deploy new services and the improving the scalability and the manageability of the data center network components. Regulatory and industry-specific regulations together with IPv4 address exhaustion drive the adoption of IPv6 networks. In order to bridge the gap between physical and virtual network resources, standards are being developed by IEEE (802.1Qbg and 802.1Qbh) to orchestrate the network state of the virtual machines with the port settings on physical appliances thus enabling network-aware resource mobility inside the data center. Again these are standards based solutions that can be compared
6 7
For more information on IEEE 802.1aq refer to http://www.ieee802.org/1/pages/802.1aq.html . For more information on IETF TRILL refer to http://datatracker.ietf.org/wg/trill/charter/ For more information on IEEE DCB refer to http://www.ieee802.org/1/pages/dcbridges.html
209
7928ch04.fm
with vendor-specific implementations such as Cisco Nexus 1000v and BNT VMReady. Network resources can also be logically aggregated and partitioned in order to logically group or share physical appliances to improve the efficiency and the manageability of the data center network components. In order to overcome the limits of the vendor specific network management tools, abstraction layers can enable network change and configuration management features in a multivendor network environment. These tools, more common in the server and storage space, can also provide linkage to the service management layer in order to orchestrate network provisioning with the setup of new IT services, speeding up the time to deploy metrics and reducing capital and operating expenses associated with the growth of the Data Center infrastructure.
210
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
for the data center network managers and architect headaches. In fact a network services and application aware data center network requires the ability of linking the data plane understanding also with control plane and management plane services and also a broader understanding of the overall IT environment. The consolidation and virtualization of server platforms and network resources has to be balanced carefully in order to adapt to the security policies. The tradeoff between resource efficiency and guaranteeing secure isolation is a significant point of concern when collapsing services with different security requirements on the same physical resources. The high level scenario of what has been just described is shown in Figure 4-6 on page 211. This has a twofold implication : the hardware vendors should be able to prove this isolation (both from a security and from a performance independance standpoint) and this has to be proven for auditing, adding also a regulatory compliance dimension to this issue.
Physical zoning pattern Virtual zoning pattern
VM VM VM
VM VM VM VM Security Zone
VM VM
S VM VM VM VM VM VM VM VM V Security Zone M
Virtualized Server Virtual security appliances
Virtualized Server
Core Layer
Core Layer
Internet
Restricted Secured
Internet
Figure 4-6 Collapsing different security zones on shared server and network hardware
Business: The cost reduction imperative driven by the global economy has put a lot of stress on enterprises IT budgets, so projects and initiatives cannot
211
7928ch04.fm
clearly demonstrate the value for the business have very little chance of being approved and launched by the decision makers. In this context Network and IT Managers are struggling to obtain tools and methodologies that can show clear return on investment to their business executives. Networking industry consolidation and the forces driving new alliances among IT and Networking vendors may impact the existing business relashionships together with the integration and interoperability governance that have been previously in place. Organization: IT organizations can no longer be grouped into independant silos. The interdependencies between different teams managing and developing new solutions for the data center are just too many to rely on the traditional organizational model. Just to name a few examples, application characteristics that cannot be ignored by network architects and managers, virtual switch management boundaries blur and storage and data converged networks project responsibilities have to be carefully balanced between the storage and the network teams. The data center professionals mindset must include the new evolutionary step. T-shaped skills (The vertical bar on the T represents the depth of related skills as expertise in a same field, whereas the horizontal bar is the ability to collaborate across disciplines with experts in other areas, and to apply knowledge in areas of expertise other than their own) must be developed and the over-specialization paradigm has to be complemented by skills in adjacent disciplines (for example hipervisors, storage and security). So it is a matter of new skills that must be sought after both from a technological standpoint and from a communication and teamworking point of view.
212
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
Build plans that leverage networking technologies and solutions for business advantage Facilitate consensus across the organization Mitigate risk by providing guidance and suggestions that help enhance availability, resiliency, performance and manageability of networking infrastructure The definition of these principles should also avoid: Delay in the time to market of service delivery because the current networking infrastructure may not support new business initiatives Improper use of the enterprise IT organization structure that could lead to unefficiencies that might arise from new projects that are not deployed in a timely or prioritized fashion Delays in justifying technology investments Duplicated or suboptimal investments around tools and resources In the following sections we present three key evolutionary imperatives that help to overcome the required shift in terms of mindset.
213
7928ch04.fm
The network strategy and planning must therefore be driven by the business requirements and guided by a set of common design principles. In addition, the network must provide efficient, timely collection and access for the information services and must enable the protection of enterprise assets, while facilitating compliance with local, country, and international laws. Enterprise should identify the scope, requirements and strategy for each identified networking initiative inside the enterprise unique business environment, obtain visibility of the current IT infrastructure (for example performing an assessment of the current environment), analyse gaps and set the actionable roadmap to define the potential for the effects such a strategy could have on the organization, the networking and IT infrastructures. We have seen how envisioning a single distributed data center network that provides reliable access to all authorized services, provides efficiencies in management and cost, and enables more consistent service delivery needs, goes through a well defined process that aims at making educated and consistent decisions on network infrastructure and network application growth. The methodology that needs to be enforced, tightly links to business objectives and overall IT governance and needs to consider all the specific interdependencies that provide a framework for the development of the scope and direction of the data center network. Enterprises should assess the current status of the networking environment and uncover the networking future state requirements by analyzing both business and IT objectives. This methodology often needs to leverage reference models, best practices and intellectual capital. The analysis will then identify technology solutions that address a migration process that leverages the current IT investments in a sustainable operating environment. A fundamental aspect is to understand how IT integrates with the business domain. Aspects that should be defined include: Business environment and objectives Current IT environment and plans Current technology environment Current networking environment: Current principles & standards Budget & IT Investments process IT strategy and plans Business satisfaction with the existing network
214
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
The next steps aim at unleashing the potential of a sound transformation that matters to the enterprise: Determine existing versus required state Identify the gaps between the enterprise's current environment and the new networking strategy; Build a portfolio of initiatives that overcome the gaps Prioritize the different identified initiatives and the networking and overall IT dependencies Determine initiative overall impact (including on the organization) and transition strategy Plan an actionable roadmap in conjunction with the business ROI expectation in terms of costs versus the value derived. Develop possible solution approaches for each planned initiative and select which approach to pursue; Schedule initiatives Complete transition plan With the reiteration of this process the enterprise has set the strategy and the path to a comprehensive networking design that encompasses and unifies networking data forwarding, networking security and networking services.
215
7928ch04.fm
Processes : for example centralizing processes that span across the organization (for example procurement, external support, crisis management) improve consistancy and speeds up lead time. Also cost savings can be achieved by eliminating duplicated processes optimizing resource efficiency both physical and intellectual. Tools : for example reducing the number of vertical tools that are required to manage and provision the infrastructure can bring cost benefits in terms of required platforms and skills and also improve service delivery by leveraging consistant user interfaces. So the goal of standardization in this context can be broadly categorized as follows: Lower transition costs to new technologies or suppliers by leveraging standard-based protocols. But also vendor-specific features that are de-facto standards can be leveraged in this context, not just IEEE or IETF ratified ones. This also implies that procurement processes across different technology domains (for example network and security equipment) should be centralized and coordinated in order to avoid duplication and ensure a common strategy and interoperability. Lower coordination costs (which also encompass skills and time) by decoupling the functional layer with the implementation layer. Tools that provide this abstraction layer are pivotal in order to use standard interfaces (such as APIs) that application developers and other infrastructure domains can understand, regardless of the low-level implementation of the specific function. Lower management costs and improved service delivery by using common interfaces and operational policies so that advanced functions like correlation and analytics can be implemented without requiring complicated and time consuming ad-hoc customizations. External and internal regulatory compliance is easier to achieve and demonstrate because changes, inter-dependencies and control are easier to trace back and fix. On the other hand, there are also barriers to the standardization process as it has been previously described. Standards might offer limited functionality, lower adoption rates and less skills available than other more customized alternatives which may be already in use. It is not always easy to coordinate across different organizational departments. It is very difficult to have an over-encompassing view of the overall infrastructure and its interpendencies because applications, technology
216
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
domains and even data centers may be handled independently and not in a sychronized, consistant fashion. The upfront capital investments that may be required and also the risks of shifting away from what is well known and familiar are also barriers in this context. Network professionals typically operate through command line interfaces (CLIs) so a shift in terms of tooling, mindset and skills is required in order to integrate network management and configuration into more consumable interfaces such as the one used in the storage and server arenas. The over-sophisticated and highly customized and heterogeneous IT world struggles with the end goal of a homogeneous resource pool-based IT environment without actionable, well-understood and agreed-opon milestones and initiatives that might depend on tools and technologies that offer limited functionalities compared to whats required. Standardization is a journey that requires coordination and understanding of common goals across people, tools, technologies and processes. On the other hand, it is a key enabler for automation, cost and service delivery optimization and is something that should be always kept in mind when integrating new functionalities and technologies inside an dynamic data center infrastructure.
217
7928ch04.fm
to at least one of the enterprise's key business requirements. The principles are also developed to help explain why certain products, standards, and techniques are preferred and how these principles respond to the needs of Enterprise Security, Servers, Applications, Storage and so on. A Network Conceptual Design: this is the first step in crafting a network solution to support business requirements. It describes the relationship between the building blocks or functions that comprise the network and services required to meet business needs. This design shows also the interactions between networking and other IT domains. At this level, technology is not specified but support of the business purpose is shown. Business requirements are then mapped to the network through guiding principles, conceptual network design diagrams and application flows. A Specified Design: this process refines the conceptual design by developing the technical specifications of the major components of the solution in terms of interfaces, capacity, performance, availability, security and management and how they relate to other IT domains. This approach gives detailed specification of the topology, sizing, and functionality of the network. At this point decisions are made, sorting through architectural and functional alternatives. A Physical Design: this documents the most intricate details of the design that are required for the implementation. At this level of the design, it is important to: Identify product selection criteria; Select products that meet specific requirements; Select connection technologies; Rationalize and validate the design with the architectual governance; Develop infrastructure and facilities plans for the network; Develop a detailed bill of materials; Prepare a network infrastructure implementation plan that lists also organizational impacts and dependencies for the master plan that the architectual board need to craft. Once the Physical Design is complete, the enterprise will have a level of detail appropriate to execute the inititiave. In Figure 4-7 on page 219, the possible interactions among the IT infrastructure teams that are needed for a service oriented DCN design are depicted, along with examples, that the network domain might experience with other IT and non-IT domains.
218
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
Applications
Industry Regulations
Network
Storage Facilities
Zoning, device access,port security protocols,... SNMPNETCONFWeb-Interface Network Provisioning & AutomationTools
Service Management
Networking is in a truly unique position to connect together the pieces of the IT environment and drive the transformation towards efficiency at the right cost for the business. While we see industry regulations as external yet mandatory to follow, for each of the planned IT initiatives, the network team, as shown in Figure 4-7, can leverage and work with: Application team to determine how the network should treat the flows and assure the required availability. It is also important to prove to the business the cost that the network must sustain in order to satisfy all the requirements. For example, the expected level of availability of a certain application has to be matched with similar (if not higher) network availability. This holds true both for client to server and for server to server traffic flows. Server team to define the right level of synergy to avoid layer and functionalities duplications and delays in the application flow processing. For example the virtual switching functions implemented in the hypervisors are important requirements to document at this stage to define clearly the infrastructure management boundaries.
Security
Server
Framework
219
7928ch04.fm
Storage team to explore and implement ways to improve the overall transaction quality and cost while maximazing data security. For example data and storage convergence initiatives should be architected and implemented with the storage, server and network teams acting as a single entity. Security team to protect data with the right impact on the operations service agreements following the guiding principles, constraints, and requirements that govern, control or influences the final solution design and delivery. In this sense the network must be the manifestation of the enterprise security policies (for example zoning, port access security, ...) by adopting the broader security framework. For example, network and security features may be implemented on the same appliance thus bringing the need to synchronize architecture and operations between the network and security teams. Site& facilities team to understand the consequences that new initiatives have on cabling, cooling, and rack space. For example, even if rack space is not an issue, power contraints may be a barrier for a data center augmentation. Service management team to assess the current and planned portfolio of management tools and processes for the IT environment and identify any gaps or overlaps in the coverage provided by that portfolio. For example, service provisioning and automation tools may need specific configurations on the network side both on the physical devices and on a network management level. Deep skills and extensive experience are necessary to detail and specify the complete enterprise architecture. It is necessary to link network services and security to other aspects of IT and business organization at key design decision points, including systems management, applications development, organizational change, testing and business continuity. The approach we recommend enables the enterprise to deliver a robust and resilient solution that aligns with and supports the ever changing business and industry regulations environment.
220
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
to support the changing business requirements and take advantage of the tremendous industry innovation. When equipped with a highly efficient, shared, and dynamic infrastructure, along with the tools needed to free up resources from traditional operational demands, IT can more efficiently respond to new business needs. As a result, organizations can focus on innovation and aligning resources to broader strategic priorities. Decisions can be based on real-time information. Far from the "break/fix" mentality gripping many data centers today, this new environment creates an infrastructure that provides automated, process-driven service delivery and is economical, integrated, agile, and responsive. What does this evolution mean for the network? Throughout the evolution of the IT infrastructure, you can see the increasing importance of stronger relationships between infrastructure components that were once separately planned and managed. In a dynamic infrastructure, the applications, servers, storage and network must be considered as a whole and managed and provisioned jointly for optimal function. Security integration is at every level and juncture to help provide effective protection across your infrastructure-and across your business. Rapid innovation in virtualization, provisioning and systems automation necessitates expanding the considerations and trade-offs of network capabilities. Additionally, the ultimate direction for dynamic provisioning of server, storage and networking resources includes automatic responses to changes in business demands-such as end-user requests, business continuity and energy constraints-so your current network design decisions must be made within the context of your long-term IT and business strategies. IBM Global Technology Services (GTS) has a suite of services to help you assess, design, implement and manage your data center network. Network Strategy, Assessment, Optimization and Integration Services combine IBM's IT and business solutions expertise, proven methodologies, highly skilled global resources, industry-leading management platforms and processes, and strategic partnerships with other industry leaders to help you create an integrated communications environment that drives business flexibility and growth. IBM Network Strategy, Assessment and Optimization Services help identify where you can make improvements, recommend actions for improvements and implement those recommendations. In addition, you can: Resolve existing network availability, performance or management issues Establish a more cost-effective networking and communications environment Enhance employee and organizational productivity Enable and support new and innovative business models.
221
7928ch04.fm
IBM Network Integration Services for data center networks help you position your network to better meet the high-availability, high-performance and security requirements you need to stay competitive. We help you understand, plan for and satisfy dynamic networking demands with a flexible, robust and resilient data center network design and implementation. Whether you are upgrading, moving, building or consolidating your data centers, our goal is to help improve the success of your projects. To help you deliver a reliable networking infrastructure we offer services that include: Network architecture and design based on your data center requirements A comprehensive design that integrates with your data center servers, storage, existing networking infrastructure and systems management processes Project management, configuration, implementation, network cabling and system testing of server-to-network connectivity, routers, switches, acceleration devices and high-availability Internet Protocol (IP) server connections Options for proactive network monitoring and management of your enterprise network infrastructure to help achieve optimal levels of performance and availability at a predictable monthly cost
222
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
Help me do it
Plan
Assess existing network Develop actionable recommendations Deliver a transition roadmap
Design
Develop the network architecture Create a detailed design Make a plan for implementation
Implement
Build network according to design specified Test network in limited production environment Deploy network in full production environment
Operate
Monitor and manage the network remotely Provide equipment maintenance services
Project management
In all phases shown in Figure 4-8, we will work with you to: Plan Understand your current IT and networking environment. Collect and document your requirements. Identify performance and capacity issues. Determine your options. Compare your current environment to your plans. Make recommendations for transition
Design Develop a conceptual-level design that meets the identified solution requirements. Create a functional design with target components and operational features of the solution. Create a physical design to document the intricate details of the solution, including vendor and physical specifications, so that the design may be implemented. Deliver a bill of materials and a plan for implementation. Implement
223
7928ch04.fm
Review the implementation plans. Perform site readiness survey. Procure the equipment. Develop installation procedures, solution testing and certification plans. Stage the solution. Implement and test the solution. Train your in-house support staff. Review information about your networking environment Configure remote monitoring and management of your devices Test and start the service Provide proactive 24x7 monitoring Manage and report based on the level of service you have selected Deliver these services for a monthly fee based on the number of devices under management
224
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
business partnership relationships at varying levels and help both organizations maximize the value and objective attainment that each can realize from this relationship.
225
7928ch04.fm
226
IBM Data Center Networking: Planning for virtualization and cloud computing
7928ch04.fm
procurement, site preparation, cabling configuration, installation, system testing, and project management. IBM has the deep networking skills and extensive experience necessary to assist you in detailing and specifying a data center network architecture. Our Network Integration Services team links network services and security to other aspects of your IT and business organization at key design decision points, including systems management, applications development, organizational change, testing and business continuity. Our approach enables us to deliver a robust and resilient solution that aligns with and supports your changing business. A strong partner ecosystem with suppliers like Cisco, Juniper, F5, and Riverbed also enables our network and security architects to support the right combination of networking technology for you.
227
7928ch04.fm
228
IBM Data Center Networking: Planning for virtualization and cloud computing
7928bibl.fm
Related publications
The publications listed in this section are considered particularly suitable for a more detailed discussion of the topics covered in this book.
IBM Redbooks
For information about ordering these publications, see How to get Redbooks on page 231. Note that some of the documents referenced here may be available in softcopy only. Best Practices for IBM TEC to Netcool Omnibus Upgrade, SG24-7557 Communications Server for z/OS V1R9 TCP/IP Implementation Volume 1 Base Functions, Connectivity and Routing, SG24-7532 Communications Server for z/OS V1R9 TCP/IP Implementation Volume 2 Standard Applications, SG24-7533 Communications Server for z/OS V1R9 TCP/IP Implementation Volume 3 High Availability, Scalability and Performance, SG24-7534 Communications Server for z/OS V1R9 TCP/IP Implementation Volume 4 Security and Policy-Based Networking, SG24-7535 IBM Systems Virtualization Servers, Storage, and Software, REDP-4396 IBM System z Capacity on Demand, SG24-7504 IBM System z Connectivity Handbook, SG24-5444 IBM System z Strengths and Values, SG24-7333 IBM System z10 BC Technical Overview, SG24-7632 IBM System z10 EC Technical Guide, SG24-7516 IBM System z10 EC Technical Introduction, SG24-7515 Introduction to the New Mainframe - z/VM Basics, SG24-7316 Migrating to Netcool Precision for IP Networks - Best Practices for Migrating from IBM Tivoli NetView, SG24-7375 PowerVM Virtualization on IBM System p Introduction and Configuration 4th Edition, SG24-7940 z/VM and Linux Operations for z/OS System Programmers, SG24-7603 GDPS Family - An Introduction to Concepts and Capabilities, SG24-6374
229
7928bibl.fm
Clustering Solutions Overview - Parallel Sysplex and Other Platforms, REDP-4072 Cloud Security Guidance IBM Recommendations for the Implementation of Cloud Security, REDP-4614 IBM Scale Out Network Attached Storage Architecture, Planning, and Implementation Basics, SG24-7875 IBM zEnterprise System Technical Guide, SG24-7833 IBM zEnterprise System Technical Introduction, SG24-7832 IBM XIV Storage System: Architecture, Implementation, and Usage, SG24-7659 IBM System Storage Solutions Handbook, SG24-5250 Integration Guide for IBM Tivoli Netcool/OMNIbus, IBM Tivoli Network Manager, and IBM Tivoli Netcool Configuration Manager, SG24-7893 IBM System Storage N Series Hardware Guide, SG24-7840 IBM System Storage N series Software Guide, SG24-7129 IBM b-type Data Center Networking: Design and Best Practices Introduction, SG24-7786 IBM b-type Data Center Networking: Product Introduction and Initial Setup, SG24-7785 IBM j-type Data Center Networking Introduction, SG24-7820 TCP/IP Tutorial and Technical Overview, GG24-3376 IBM BladeCenter Virtual Fabric Solutions, REDP-4673 Cloud Security Guidance IBM Recommendations for the Implementation of Cloud Security, REDP-4614 Enterprise Security Architecture using IBM ISS Security Solutions, SG24-7581 IBM BladeCenter Products and Technology, SG24-7523
Online resources
These Web sites are also relevant as further information sources: ICS SalesOne Portal http://w3-03.ibm.com/services/salesone/S1_US/html/htmlpages/networki ngsvcs/ns_splash.html
230
IBM Data Center Networking: Planning for virtualization and cloud computing
7928bibl.fm
Virtual Ethernet Port Aggregator Standards Body Discussion, Paul Congdon, 10th November 2009 www.ieee802.org/1/files/public/docs2008/new-congdon-vepa-1108vo1.pdf First Workshop on Data Center - Converged and Virtual Ethernet Switching (DC CAVES) http://www.i-teletraffic.org/itc21/dc-caves-workshop/ VSI Discovery and Configuration - Definitions, Semantics, and State Machines http://www.ieee802.org/1/files/public/docs2010/bg-sharma-evb-VSI-dis covery-0110-v01.pdf Trivial TLP Transport (T3P) - Proposed T3P DU and State Machines http://www.ieee802.org/1/files/public/docs2010/bg-recio-evb-t3pr-011 0-v01.pdf
Related publications
231
7928bibl.fm
232
IBM Data Center Networking: Planning for virtualization and cloud computing
Index
Numerics
10 gigabit ethernet 209 802.1q trunk 185
C
Capacity estimate 24 capacity, definition of 25 CAPEX 201 capital expenses 208 Challenges 2 Appliance sprawl 207 Control Plane Stabilit 206 End to end network visibility 207 Heterogeneous Management tools 208 important challenges 195 Instituting a virtual data center 196 Introducing high levels of network automation 196 Layer 2 Extension 205 Network resources consolidatio 208 Obtain VM-level network awareness 206 Optimal use of network resources 206 Supporting business resilience and disaster recovery 196 technology related 202 Channel Sub System (CSS) 42 Channel Subsystem Priority Queuing 44 chassis-based switch 163 Cisco FabricPath 208 Cisco Nexus 1000v 210 Classless Inter-Domain Routing 201 client-server flow 204 Cloud computing 2, 67 other system components 175 cloud computing 7, 10, 174, 197 cost benefits 197 many definitions 7 Unified Ontology 8 cloud computing uses the network 198 cloud deployment models 198 cloud solution 8 cluster 94 collaborate across disciplines 212 Collapsing different security zones on shared server and network hardware 211 commited rate 159 Configuration management 25, 2829 consistant set of tools 206 consolidation and virtualization of the network 207 converge storage and data networks 209 convergence 24 corporate firewall 14 Cost drivers 201 cost drivers 207 cost reduction 208, 211 cost savings 208209 current data center network 197
A
A Closer Look at Private Cloud TCO 197 abstraction layer 22 abstraction layers 210 Accelerate 212 access controls 27 access layer 153, 163 increased bandwidth demands 164 access switch sprawl 153 achieve greater flexibility 208 actionable roadmap 214 active/active configuration 166 active/active model 159 Adapter Fault Tolerance (AFT) 187 Adaptive Load Balancing (ALB) 187 adding greater redundancy 195 adding network bandwidth 195 additional points of concern 210 address network performance issues 195 ad-hoc customizations 216 Advanced POWER Virtualization editions 60 alleviate the challenges 208 alternatives to help manage risks and reduce costs 197 analyse gap 214 appliance sprawl 208 application acceleration 209 application characteristics 212 application firewalling 178 application session 161 approaches 193 architectural decisions 195 Avoiding complexity 196 Avoiding risk 196
B
BNT VMReady 210 Brocade Virtual Cluster Switching VCS 208 Build 213 Business 211 business 210 business and service objectives 202 business model 1, 5 business requirement 158, 193, 198 too much time 172 business requirements that impact the CIOs decision making 201
D
Data Center A 204 B 204 Bridging 209 Consolidation 149, 157, 195 Network Access 150 Network-wide virtualization 151 data center 23, 1112, 22, 149, 169, 193 application mobility 154 current landscape 152 data synchronization 172 layer 2 172 network-aware resource mobility 209 networking challenges 205 new building block 206 new solutions 212 other devices 200 physical server platform 2223 systems resources 16 tremendous strain 17 virtual machines 172 data center network 196 service delivery objectives 201 support and enablement 201 data center network architectures 208 data center network design 193 data center network has been optimized 198 Data center network logical architectural overview diagram 203 DCM (Dynamic channel path management) 45 DCN 193 Dealing with a massive increase 196 dedicated hardware appliances 208 dedicated partition 41 de-facto standard 216 delay, definition of 26 delay-sensitive applications 26 delivery of services for the remote clients optimization 203 demading network requirements 207 dense wavelength division multiplexing (DWDM) 203 deployment model 152, 175, 198 attractiveness of a private cloud 198 required financial impact 176 deployment models New 209 Developing a plan for the network and associated functional design 195 Developing a Strategic Approach 213 digitally-connected world 18 dynamic infrastructure 18 disaster recovery 203 Dispelling the vapor around cloud computing 197 DV filter 181 DVIPAs (Dynamic VIPAs) 56 DWDM 203 Dynamic channel path management (DCM) 45 Dynamic Cross-System Coupling Facility (dynamic XCF)
dynamic environment 150, 152 Dynamic Infrastructure 23 strategy 2 dynamic infrastructure 1, 1819, 151, 163, 197, 221 efficient management 191 emerging trend 22 open, scalable, and flexible nature 15 pivotal functionality 151 dynamic routing 56 Dynamic VIPAs (DVIPAs) 56
E
Emerging economy 1 End of Row (EOR) 153 End of Row (EoR) 153 end user 7, 15, 163, 195 Enterprises that adopt cloud computing 198 Ethernet connection 185 Ethernet frame 165 Ethernet mode (Layer 2) 51 Ethernet port 187 Ethernet world 165 Evaluating vendor technologies 197 Evolution of the data center factors influencing ix evolutionary imperatives 193 example hypervisors 212 expected service delivery 212 extended backplane 166 extensive customization 10 external regulations 28 External switch tagging (EST) 89
F
Facilitate 213 facilities team 220 fact VPLS 171 Failure management 25 FCoE Fiber Channel over Ethernet 209 Fiber Channel Forwarder 162 over Ethernet 209 Switch 209 Fiber Channel over Ethernet FCoE 209 firewalling 209 First Hop Router Protocol (FHRP) 173 FRR TE 173
G
geo-load balancing 203 Gigabit Ethernet 52 global economy 211 Global integration indicators 1 Global Technology Services (GTS) 221
networking guideline principle 212 governance 212 governance of the data center network 212 guest VM 164 guiding principle 217
H
hardware platform 161, 176 Health Insurance Portability and Accountability Act (HIPAA) 18, 28 higher-value activity 10 HiperSockets 43 homogeneous resource 217 hypervisor API 180 Hypervisor development 153 hypervisor level 204 hypervisor resource 163
I
I/O (IO) 188 I/O operation 164 I/O performance 164 IBM Integrated Communications Services portfolio 193 IBM Security Framework 177 network component 180 product 180 IBM Security, see (ISS) 183 IBM Systems Director 124 IBM Tivoli Network Manager (ITNM) 128 IBM Virtualization Engine TS7700 109 IDS (Intrusion Detection Services) 58 IEEE 802.1Qbg 209 802.1Qbh 209 IEEE 802.1aq 209 IEEE standard 163 IETF 209 IETF draft 171 IETF L2VPN 171 IETF TRILL 209 IFL (Integrated Facility for Linux) 46 IGP 172 impact on the security requirements 208 implementing new network equipmen 195 improve the efficiency 210 improving network security 195 inability to perform accurate and timely root-cause analysis 195 increasing capacity by upgrading switches 195 independant silo 212 independence of logical resources 208 individual SLAs 12 Infiniband network 165 Infrastructure management 204 Integrated Facility for Linux (IFL) 46 Integrated Virtualization Manager, IVM 64 interconnection between data centers 203
Internet address 200 Internet Number 200 American Registry 200 internet standard 171 interoperability governance 212 interoperability issue 13 Inter-VM IPS 181 Intrusion Detection Services (IDS) 58 IO VI 189 IP address 183, 200 range 201 standard 191 IP address range allocated to an organization 201 IP assignment 175 IP core 156 IP resolution 175 IP router 174 IPS/IDS. VPN 175 IPv4 address exhaustion 209 IPv4 address space Number remaining 200 IPv4 and IPv6 can coexist 200 IPv6 200 address spaces 200 network support 200 planning and readiness 201 promises 200 IPv6 and the data center 199 IPv6 networks adoption of 209 isolation security 208 IT infrastructure 203 IT simplification 210
J
jitter, definition of 26 Juniper Stratus 208
K
key enabler 217
L
L2/L3 boundary 171 LAN is shifting 210 LAN-like performance 158 Layer 2 (data link layer) 50 layer 2 domain 208 Layer 3 (network layer) 50 layer 3 IP network 209 legacy data center 197 License and Update Management (LUM) 181 Live Partition Mobility 64 logical partitions 208 Logical Unit Number (LUN) 37 logically aggregate 210 logically group 210
M
main enablers 157 manage the network as one single logical switch 209 manageability of the data center network components 209 manageability of the data center network components. 210 management cost 215 management of servers 208 MapReduce method 6 Market Landscape for Data Center Networking 193 master plan 218 merge the plug-and-play nature of a ethernet layer 2 networ 209 Microsoft System Center Virtual Machine Manager 2007 (SCVMM) 103 Migration planning 196 Mitigate 213 mobile device 67 moving an application workload to another server 198 MPLS 203 MPLS service 173 multipathing 208 multiple SI 189 multi-protocol label switching (MPLS) 203 multipurpose virtualized appliances 209 multi-tenant architecture 8 multi-tenant environment 174 multi-tenant platform 12 multi-vendor networking environment 197 Multi-vendor networks 197
N
navigating vendor-specific alternatives 210 Network Access Control (NAC) 181 Network Address Translation 201 network architecture 2425, 28, 186187, 206 Network Challenges relevant 205 Network challenges risks 195 Network changes most common types of 195 Network critical role 198 network device 25, 29 network infrastructure 23, 28, 152, 154, 196, 202 Data Convergence impact 155 educated and consistent decisions 214 network interface card (NIC) 23, 159, 198 Network requirements 195 Network resources 210 network service 161, 175, 209 deployment model 176 different categories 175
access security 204 application delivery 204 local switching and routing 204 Network virtualization different levels 184 first manifestations 150 network virtualization 3, 33, 149 technology 149 network-attached device 199 networking attributes for a private cloud data center network design 198 Networking industry consolidation 212 Networking investments 195 Networking vender 212 Networks - essential to the success of cloud computing initiatives 197 network-specific function 204 new inputs 193 new issues 210 New products 196 New products and new technologies 196 new services 208 new skills 212 new technologies 196 NFR non-functional requirements 202 non-blocking 208 non-functional requirements NFR 202 non-functional requirements (NFR) 202 non-functional requirements (NFRs) 23 north-to-south traffic 164 different requirements 164 NRO Number Resource Organization 200 Number Resource Organization NRO 200 Number Resource Organization (NRO) 200
O
oad balancing 209 Open Shortest Path First (OSPF) 55 Open Systems Adapter-Express (OSA-Express) OSA-Express (Open Systems Adapter-Express) 48 operating expenses 208 operating system 15, 164 OPEX 201 optimized for flexibility 198 options 212 Organization 212 organization 210 OSA-Express 53 OSI model 178 layer 7 178 outside OS 183 OS memory tables 183 overall initiative 195 Overlay Transport Virtualization (OTV) 171
P
packet services delivery 210 Parallel Sysplex 57 Parallel Sysplex clustering 45 partitioned 210 Payment Card Industry (PCI) 18, 28 Perceived barrier 12 performance and availability 208 physical appliances 209 physical server platform 22 physical switch 153 plain old telephone service (POTS) 152 PMBOK phase 224 points of concern 193 Policy enforcement 160, 173 polling agents 129 portsharing 57 POWER Hypervisor 61 primary router (PRIRouter) 53 Private cloud adoption models 199 Project Management Body 224 Institute 224 Project Management Institute (PMI) 224 proprietary control plane architecture 208 protocol analysis module (PAM) 180 public Internet 13, 199
RIP (Routing Information Protocol) 55 RIRs Regional Internet Registries 200 risk/opportunity growth 1 Role of the network critical focus 195 Root Complex (RC) 189 Root Port (RP) 189 rootkit 183 Routing Information Protocol (RIP) 55
S
same IP 184 address 187 same subnet need 170 SAN Virtual Controller (SVC) 106 SAN-like model 210 scalability 27, 209 scalability and manageability 208 SEA (Shared Ethernet Adapter) 67 secondary router (SECRouter) 53 security 71 certifications 71 security framework 169, 173, 220 important implications 173 security layer 156 Security policy 27 security zones 27 security, definition of 27 Selecting standards, techniques, and technologies 196 semi-trusted region 27 server and storage specific functionalities 204 Server consolidation and virtualization initiatives 207 server network interface card (NIC) 24 server virtualization 33 requirement 203 technique 167 techologies 205 service delivery 24, 161, 197, 201 Service Drivers 201 Service Level Agreement (SLA) 28 Service Management 23 service provider 28 service-oriented approach 4 share physical appliances 210 share the physical network resources 208 Shared Ethernet Adapter (SEA) 67 shared partition 41 shared resources pools 210 Shortest Path Bridging SPB 209 significant saving 197 single distributed data center 203 Single Root I/O Virtualization (SR-IOV) 164, 188 single ToR 167 Site Protector (SP) 181 Software Development Kit (SDK) 91
Q
QDIO mode 50 QoS capability 186 QoS tag 191 Quality of Service (QOS) 26, 28, 207 Quality of Service (QoS) requirements 26
R
RACF 58 RACF (Resource Access Control Facility) 49 real-time access 5, 18, 152 real-time data 6 real-time information 3, 221 real-time integration 4 recentralization 157 Redbooks Web site 231 Contact us xi redundant data centers 24 Regional Internet Registries RIR 200 Regulatory and industry-specific regulations 209 regulatory compliance 12 related skills 212 requirement IPV6 only 200 Resource Access Control Facility (RACF) 49 resource efficiency 208, 211
spanning tree protocol STP 206 spanning tree protocol (STP) 166 SPB Shortest Path Bridging 209 Special Interest Group (SIG) 188 special-purpose software 8 spiraling number 199 SSL termination 162 stability of the control plane 206 Standards-based fabrics 209 static VIPA 56 storage and data converged networks 212 storage virtualization 33, 65 technology 199 trend 152 STP Spanning Tree Protocol 206 STP approach 172 STP instance 206 STP instances 206 suboptimal performance 213 Support new, network-demanding services 207 Supporting virtualization 195 SVC (SAN Virtual Controller) 106 Switch Fault Tolerance (SFT) 187 Sysplex Distributor (SD) 56 System management 28
use case 15
V
VCS Brocade Virtual Cluster Switching 208 vendor lock-in 11 vendor-specific alternative 210 Vendor-specific data center architectures 208 vendor-specific feature 216 vendor-specific implementation 210 vertical, vendor and model specific tools 208 VIPA (Virtual IP Addressing) 55 virtual appliances 209 virtual End of Row (VEOR) 163 virtual Ethernet 65 Virtual Ethernet Port Aggregator (VEPA) 231 Virtual I/O Server 6465 Shared Ethernet Adapters 65 Virtual IP Addressing (VIPA) 55 virtual MAC function 55 virtual machine network state 209 virtual machine (VM) 23, 47, 162, 196 virtual machine (VM) mobility 204 Virtual machine guest tagging (VGT) 89 Virtual Machine Observer main task 183 Virtual Machine Observer (VMO) 181 virtual networking 206 Virtual Private LAN Service (VPLS) 171 virtual resource pool level 204 Virtual Server Security (VSS) 180 virtual switch 153, 185 management boundary 212 network 180 port 188 support QoS 191 Virtual switch tagging (VST) 88 virtualization layer 13 virtualization technology 23, 149, 207, 220 virtualized data center layer 169 network 22 network environment 179 virtualized resource pools implications 204 virtualized system 164 virtualized, to meet (VM) 150, 183 VLAN tagging 190 VLAN translation 171 VM building block 206 VM (virtual machine) 47 VM behavior 180 VM entity 188 VM migration 161 VMotion 94 VMs 180
T
Technology 210 technology 210 technology alternatives 212 The New Voice of the CIO 197 thinking 193 throughput, definition of 26 Tivoli Service Automation Manager and ISM Architecture 116 TLS (Transport Layer Security) 58 Todays DCN Challenges 202 tools and methodologies 212 Top of Rack (TOR) 153 Top of Rack (ToR) 153 ToR model 163 ToR topology 164 tradeoff 211 Traffic Engineering (TE) 173 Transparent Interconnect of Lots of Links TRILL 209 Transport Layer Security (TLS) 58 TRILL Transparent Interconnect of Lots of Links 209 TSAM - High-Level Architecture 118 T-shaped skill 212 TTL 174 two-fold implication 204
W
WAN Acceleration 203 wide area network (WAN) 160161, 203 WOC 161 workforce 7, 152 www.internetworldstats.com/stats.htm 5
Z
z/VM guest 46 z/VM virtual switch 50
To determine the spine width of a book, you divide the paper PPI into the number of pages in the book. An example is a 250 page book using Plainfield opaque 50# smooth wh 250 by 526 which equals a spine width of .4752". In this case, you would use the .5 spine. Now select the Spine width for the book and hide the others: Special>Conditional Text>Show/Hide>SpineSize(-->Hide:)>Set . Move the changed Conditional text settings to all files in your book by opening the book file with the spine.fm still open and
To determine the spine width of a book, you divide the paper PPI into the number of pages in the book. An example is a 250 page book using Plainfield opaque 50# smooth wh 250 by 526 which equals a spine width of .4752". In this case, you would use the .5 spine. Now select the Spine width for the book and hide the others: Special>Conditional Text>Show/Hide>SpineSize(-->Hide:)>Set . Move the changed Conditional text settings to all files in your book by opening the book file with the spine.fm still open and
Back cover
IBM Data Center Networking: Planning for virtualization and cloud computing
Drivers for change in the data center IBM systems management networking capabilities The new data center design landscape
The enterprise data center has evolved dramatically in recent years. It has moved from a model that placed multiple data centers closer to end users to a more centralized dynamic model. The factors influencing this are varied but can mostly be attributed to regulatory, service level improvement, cost savings, and manageability. Multiple legal issues regarding the security of data housed in the data center have placed security requirements at the forefront of data center architecture. As the cost to operate data centers has increased, architectures have moved towards consolidation of servers and applications in order to better utilize assets and reduce "server sprawl." The more diverse and distributed the data center environment becomes, the more manageability increasingly becomes an issue. These factors have led to a trend of data center consolidation and resources on demand using technologies such as virtualization, higher WAN bandwidth technologies, and newer management technologies. In this book we discuss: The current state of the data center network The business drivers making the case for change The unique capabilities and network requirements of system platforms