POWER 智慧系统的完美虚拟化技术 云计算坚实可靠的源动力
议题 服务器虚拟化技术概览 IBM PowerVM 虚拟化技术介绍 CPU 虚拟化 I/O 虚拟化 内存虚拟化 分区在线迁移 负载分区 实施服务器虚拟化的考虑 服务器虚拟化的案例 虚拟化的管理 2
云计算 公共云 ( 服务提供商 互连网 ) 私有云 混合云 ( 数据中心 内部网 ) ( 公共和私有 ) Services as a service 软件即服务 (SaaS) ( 应用, 流程和信息 ) 软件平台即服务 (PaaS) ( 优化的中间件 ) 基础设施即服务 (IaaS) ( 计算, 存储, 网络 ) 3
虚拟化是云计算的技术基础 基础设施即服务 (IaaS) 基础设施 服务器服务 服务器服务器服务器服务器 + 磁盘阵列 磁盘阵列 存储服务 磁盘阵列 磁带库 虚拟服务器文件系统逻辑卷 服务器虚拟化 服务器服务器服务器服务器 + 磁盘阵列 存储虚拟化 磁盘阵列 磁盘阵列 磁带库 4
为什么在大型服务器上实施虚拟化 更高的服务器利用率 更有效的数据中心 CPU & memory I/O ports & adapters Large or small workloads Power and generator Cooling Floor space 提高生产力 强大的 RAS Order, track, retire Install, cable, fix Patch fewer hypervisors More RAS features Fewer parts Capacity on Demand 5
服务器虚拟化技术概览 硬件分区技术 Hardware Partitioning Apps ( 应用 1) OS... Apps ( 应用 N) OS 固件分区技术 Bare Metal Hypervisor Apps ( 应用 1) OS... Apps ( 应用 N) OS 软件分区技术 Hosted Hypervisor Apps ( 应用 1) OS... Apps ( 应用 N) OS Partition Controller ( 分区控制 ) Adjustable partitions SMP 服务器 Hypervisor ( 固件 ) SMP 服务器 Hypervisor ( 固件 ) Host OS ( 寄居平台 ) SMP 服务器 Server is subdivided into fractions each of which can run an OS Board-level partitioning S/370 SI->PP & PP->SI Sun Dynamic Domains HP npartitions (npar) Core/thread-level partitioning HP vpartitions Sun Logical Domains Original POWER4 LPAR 6 Hypervisor provides fine-grained timesharing of all resources Hypervisor software/firmware runs directly on server System z PR/SM and z/vm POWER Hypervisor VMware ESX Server Xen Hypervisor Kernel VM Microsoft Hyper-V 硬件分区受到限制很多, 发展有限 ; 因此采用硬件分区技术将越来越少 固件分区技术成为主流虚拟化实现的方式, 能提供丰富功能, 灵活性, 可用性 软件分区技术将主要被重视操作系统层整合的客户采用 Hypervisor uses OS services to do timesharing of all resources Hypervisor software runs on a host operating system AIX Workload Partition (WPAR) Solaris Container HP Integrity VM VMware GSX Microsoft Virtual Server
IBM 虚拟化技术的领导者 A 40-year track record in virtualization innovation continues with PowerVM 1967 1973 1987 1999 2004 2007 2008 IBM develops hypervisor that would become VM on the mainframe IBM announces first machines to do physical partitioning IBM announces LPAR on the mainframe IBM announces LPAR on POWER IBM intro s POWER Hypervisor for System p and System i IBM announces POWER6, the first UNIX servers with Live Partition Mobility IBM announces PowerVM 采用 IBM 虚拟化技术实现服务器集中的 TCO 分析 费用预计节省分析 维护成本 69% - 76% 服务器数量的减少降低了维护所需费用 软件 65% - 69% 更少的服务器数量以及更少的 CPU 数量, 显著降低软件的许可费用 升级和 服务费用 人工 31% - 45% 更少的服务器数量和简化的 IT 架构, 自动化的管理工具 运营成本 52% - 61% 节省机房空间 空调 用电等 7
PowerVM 虚拟化技术介绍 8
PowerVM 主要内容 Processor Shared or dedicated LPARs Capped or uncapped LPARs Multiple shared processor pools Dynamic LPAR operations (add/remove) Shared dedicated LPARs I/O Shared and/or dedicated I/O Virtual Ethernet, virtual SCSI Dynamic LPAR operations (add/remove) Integrated Virtual Ethernet Virtual FC (N_Port ID Virtualization) Virtual Tape Support Dedicated Processor LPARs OS LPAR OS LPAR WPAR LPAR Virtual I/O Server LPAR LPAR Shared Processor Pool Sub-Pool A OS LPAR OS LPAR Sub-Pool B OS LPAR OS LPAR PowerVM Hypervisor 9 Memory Dedicated memory Active Memory Sharing Dynamic LPAR operations (add/remove) Active Memory Expansion (POWER7/AIX 6.1) Other Integrated Virtualization Manager Live LPAR mobility Workload partitions (AIX 6.1) Workload partition mobility (AIX 6.1) Lx86 for Linux applications (Linux) IBM i Virtualization (IBM i)
PowerVM 主要内容 Processor 微分区或专有分区 处理能力分区内受限或不受限 多个共享处理器池 动态分区 ( 增加或移走 CPU) 共享专有分区的处理能力 I/O 虚拟 I/O 和真实 I/O 可以同时存在 虚拟以太网和虚拟 SCSI 设备 动态分区 ( 增加或移走 I/O 设备 ) 集成的虚拟以太网卡 (IVM) 虚拟 FC 卡 (NPIV) 虚拟磁带 Dedicated Processor LPARs OS LPAR OS LPAR WPAR LPAR Virtual I/O Server LPAR LPAR Shared Processor Pool Sub-Pool A OS LPAR OS LPAR Sub-Pool B OS LPAR OS LPAR PowerVM Hypervisor 10 Memory 专有内存 共享内存 (AMS) 动态分区 ( 增加或移走内存 ) 动态内存扩展 (AME) Other 集成的虚拟化管理 (IVM) 分区在线迁移 (LPM) 负载分区 (WPAR) 负载分区在线迁移 Lx86 for Linux applications (Linux) IBM i Virtualization (IBM i)
CPU 虚拟化 专有分区和微分区 专有分区 (Dedicated-processor partition): 颗粒度是 1 个 CPU 的分区 微分区 (Shared-processor partition): 颗粒度是 0.1 个 CPU 的分区 动态分区 (DLPAR): 调整分区内的资源 ( 例如 :CPU 或内存等 ) 时, 不需重启分区上的操作系统 丌使用分区或仅使用 DLPAR 技术 使用微分区和共享处理器池技术 QA2 DB #1 PR1 DB #2 Test APP APP QA 11
CPU 虚拟化 受限 (Capped) 和不受限 (Uncapped) 分区 受限分区 丌允许超过其分配的 CPU 容量 丌受限分区 允许超过其分配的 CPU 容量 容量加权 用于为丌受限分区分配优先级 范围值 :0-255 数值 0 称为 软限制 Capped Shared Processor LPAR Processor Capacity Utilization Maximum Processor Capacity Entitled Processor Capacity minimum processor capacity Pool Idle Capacity Available utilized capacity Time Uncapped Shared Processor LPAR LPAR Capacity ceded Utilization capacity Pool Idle Capacity Available Processor Capacity Utilization Maximum Processor Capacity Entitled Processor Capacity ceded capacity minimum processor capacity Utilized Capacity Time 12
CPU 虚拟化 处理能力共享 Dedicated LPARs Shared CPU Pool - 0 Pool 1 Pool 2 PowerVM Hypervisor POWER Server 微分区技术 CPU 划分的颗粒度是 0.1 颗 CPU 虚拟 CPU 的个数 最小 / 最大 受限和非受限分区 处理能力自动共享 多个 CPU 池 动态分区 Dedicated LPARs Shared Pool (Uncapped LPARs) 可共享专有分区的 CPU 处理能力 PowerVM Hypervisor 13
I/O 虚拟化 Virtual I/O Server No I/O devices in Hypervisor Multiple VIOS LPAR support High availability VIOS options VIOS for specific purposes VIOS sized to match workloads Performance Virtual FC/SCSI - logical redirected DMA avoids data copies PowerVM MTU sizes up to 65280 bytes TCP/IP checksum bypass option Benefits Better availability and security Better performance POWER Server Virtual I/O Server VIO Client LPAR VIO Client LPAR VIO Client LPAR VIO Client LPAR Virtual I/O Server Virtual Device Virtual Device Physical Adapter Virtual Adapter Virtual Adapter Virtual Adapter Virtual Adapter Physical Adapter PowerVM Hypervisor VIO 技术解放物理 I/O 对虚拟服务器的束缚 每个 LPAR 中, 真实 I/O 和虚拟 I/O 可以并存 14
Virtual I/O 概览 虚拟 FC 卡 虚拟 SCSI( 硬盘 ) 虚拟 SCSI( 逻辑卷 ) 虚拟以太网 Virtual FC (NPIV) Virtual SCSI Physical Volume Virtual SCSI Logical Volume Virtual Ethernet TCP/IP Application Multipath VIOC Virt Virt FC FC Virtual WWPNs Virt SCSI Virt SCSI Virt NIC Hypervisor Virtual Switch VIOS Ethernet Bridge LVM Multipath Multipath NPIV FC NPIV FC Disk Disk Disk Disk NIC NIC NPIV FC SAN Switch SAN or Direct Attached SAN or Direct Attached Ethernet Switch Note: Dual VIO server options are commonly used - these options are not shown above 15
I/O 虚拟化 虚拟以太网 PowerVM Ethernet switch Part of PowerVM Hypervisor Moves data between LPARs Shared Ethernet Adapter Part of the VIO server Logical device Bridges traffic to and from external networks Virtual I/O Server Ent0 (Phy) Shared Ethernet Adapter ent1 (Vir) Client 1 en0 (if) ent0 (Vir) VLAN-Aware Ethernet Switch PowerVM Hypervisor Client 2 en0 (if) ent0 (Vir) Additional capabilities VLAN aware Link aggregation for external networks Ethernet Switch 16
单 VIO 服务器,VIOS 内双网卡备份 VIOS 1 Client 1 Client 2 en4 (if) ent3 (LA) ent4 (SEA) en1 (if) en0 (if) ent0 (Phy) ent1 (Phy) ent2 (Vir) ent0 (Vir) ent0 (Vir) Hypervisor PVID 1 VLAN 1 PVID 1 PVID 1 Untagged Ethernet Switch 17
双 VIO 服务器, 分区内双网卡备份 VIOS 1 VIOS 2 AIX 1 AIX 2 en2 (if) en2 (if) en2 (if) en2 (if) ent2 (SEA) ent2 (SEA) ent2 (LA) NIB ent2 (LA) NIB ent0 (Phy) ent1 (Vir) ent1 (Vir) ent0 (Phy) ent0 (Vir) ent1 (Vir) ent0 (Vir) ent1 (Vir) Hypervisor PVID 1 PVID 2 VLAN 1 VLAN 2 PVID 1 PVID 2 PVID 2 PVID 1 Untagged Ethernet Switch Untagged Untagged Ethernet Switch Note: If you split the active client interfaces across VIOS s, those LPARs will talk to each other through the external switches no the Hypervisor. Active Passive 18
双 VIO 服务器, VIOS 内双网卡备份 分区内双网卡备份 VIOS 1 VIOS 2 AIX 1 AIX 2 en4 (if) en4 (if) en2 (if) en2 (if) ent3 (LA) ent4 (SEA) ent4 (SEA) ent3 (LA) ent2 (LA) NIB ent2 (LA) NIB ent0 (Phy) ent1 (Phy) ent2 (Vir) ent2 (Vir) ent0 (Phy) ent1 (Phy) ent0 (Vir) ent1 (Vir) ent0 (Vir) ent1 (Vir) Hypervisor PVID 1 PVID 2 VLAN 1 VLAN 2 PVID 1 PVID 2 PVID 2 PVID 1 Ethernet Switch Untagged Untagged Untagged Ethernet Switch Note: If you split the active client interfaces across VIOS s, those LPARs will talk to each other through the external switches no the Hypervisor. Active Passive 19
集成的虚拟网卡 (IVE Integrated Virtual Ethernet) Virtual I/O Server Linux AIX AIX Linux AIX AIX Packet Forwarder Virtual Ethernet Driver Virtual Ethernet Driver Virtual Ethernet Driver Ethernet Driver Ethernet Driver Ethernet Driver Virtual Ethernet Switch PowerVM Hypervisor P-VM Ethernet NIC IVE Virtual I/O Server Allows sharing of PCI Ethernet adapters Some small amount of CPU time spent by packet forwarder Integrated Virtual Ethernet Removes SW packet forwarding function of VIO server Provides equivalent performance to a dedicated Ethernet adapter Dual 1 Gb Quad 1 Gb Dual 10 Gb 20
I/O 虚拟化 虚拟 SCSI External Storage Micro-partitions A1 A2 A3 A4 A5 Shared Fiber Channel Adapter Shared SCSI Adapter v S C SI VIOS v L A N AIX V5.3 A1 Linux AIX V6.1 A2 B1 B2 B3 Linux A3 B1 B2 B3 B4 B5 Virtual SCSI POWER Hypervisor VIOS 拥有物理的磁盘资源 在 VIO Server 上的存储划分 LVM 存储设备可以通过光纤或者 SCSI 连接 分区把这些磁盘看成虚拟 SCSI 设备 (Virtual SCSI) 通过 HMC 把虚拟设备配置在分区上 分区上配置这些硬盘是虚拟硬盘 VIOS 需要一定的配置 多个分区可以使用相同和不同的物理硬盘 在 VIOS 配置成逻辑卷 在分区上显示为一个 hdisk 可以配置一整块物理硬盘给一个分区使用 Virtual I/O 通过共享磁盘减少硬件投资 21
虚拟 SCSI 基本架构 Virtual I/O Server Client Partition vscsi Target Device PV VSCSI LV VSCSI Optical VSCSI LVM Multi-Path or Disk Drivers Optical Driver Hdisk Adapter / Drivers vscsi Server Adapter vscsi Client Adapter POWER Hypervisor FC or SCSI Device No I/O devices in Hypervisor Virtual FC/SCSI - logical redirected DMA avoids data copies 22
虚拟 I/O 服务器 (VIOS) 多路径支持 IBM XIV Storage System 2105 Enterprise Server (800, 750, Fxx) IBM TotalStorage SAN Volume Controller DS83000 2107 Model 921 IBM System Storage DS8100 2107 Model 922 IBM System Storage DS8300 2107 Model 9A2 IBM System Storage DS8300 2107 Model 92E IBM System Storage DS8000 Expansion Unit 2107 Model 9AE IBM System Storage DS8000 Expansion Unit DS6000 1750 Model 511 IBM System Storage DS6800 1750 Model EX1 IBM System Storage DS6000 Expansion Unit NSeries IBM System Storage N3700, N5000 and N7000 (Fibre Channel and iscsi attach) DS5000 IBM System Storage DS5100 IBM System Storage DS5300 DS4000 IBM System Storage DS4800 IBM System Storage DS4700 IBM TotalStorage DS4500 (FAStT900) IBM TotalStorage DS4400 (FAStT700) IBM TotalStorage DS4300 (FAStT600) IBM System Storage DS4200 IBM TotalStorage DS4100 (FAStT100) FastT FAStT200 Storage Server FAStT500 Storage Server DS3000 IBM System Storage DS3400 Under-the-cover IBM parallel SCSI devices http://www14.software.ibm.com/webapp/set2/sas/f/vios/documentation/datasheet.html#multipath 23
虚拟 I/O 服务器 (VIOS) 多路径支持 The Virtual I/O Server has been tested on selected configurations, which include specific models from the following disk subsystems suppliers: EMC HDS HP NetApp Contact your storage vendor for model details. http://www14.software.ibm.com/webapp/set2/sas/f/vios/documentation/datasheet.html#multipath 24
单 VIO 服务器, 逻辑卷 (LV) 映射成虚拟 SCSI 盘 Client LPAR A A Client LPAR B B vscsi0 vscsi0 Hypervisor vhost0 vtscsi0 vhost1 vtscsi1 Complexity Simpler to setup and manage than dual VIOS No specialized setup on the client VIOS 1 LV scsi0 LV Resilience VIOS, SCSI adapter, SCSI disk are potential single points of failure The loss of a physical disk may impact more than one client A B Throughput / Scalability Performance limited by single SCSI adapter and internal SCSI disks. Notes Low cost disk alternative 25
单 VIO 服务器, 硬盘 (PV) 映射成虚拟 SCSI 盘 Client LPAR A A Client LPAR B B vscsi0 vscsi0 Hypervisor vhost0 vtscsi0 vhost1 vtscsi1 Complexity Simpler to setup and manage than dual VIOS No specialized setup on the client VIOS 1 scsi0 A B Resilience VIOS, SCSI adapter, SCSI disk are potential single points of failure The loss of a single physical client disk will affect only that client Throughput / Scalability Performance limited by single SCSI adapter and internal SCSI disks. 26
单 VIO 服务器, I/O 多路径 Client LPAR 1 A Client LPAR 2 B vscsi0 vscsi0 Hypervisor vhost0 vhost1 vtscsi0 vtscsi1 Multi-Path Driver Complexity Simpler to setup and manage than dual VIO servers Requires Multi-Path I/O setup on the VIOS No specialized setup on the client VIOS 1 fcs0 fcs1 Resilience VIOS is a single point of failure A B PV LUNs Throughput / Scalability Potential for increased bandwidth due to multi-path I/O. Could divide clients across independent VIOS allowing more VIOS adapter bandwidth. 27
分区内真实硬盘与虚拟硬盘的逻辑卷镜像 (LVM Mirroring) AIX Client LPAR 1 AIX Client LPAR 2 A A B B LVM Mirroring LVM Mirroring scsi0 vscsi0 vscsi0 scsi0 A disk PowerVM Hypervisor B disk vhost0 vhost1 vtscsi0 vtscsi1 Multi-Path Driver VIOS 1 fcs0 fcs1 A B PV LUNs Complexity Requires LVM mirroring to be setup on the VIOC Multi-Path I/O setup on the VIOS If a VIOS is rebooted, the mirrored disks will need to be resynchronized via a varyonvg on the VIOC Additional complexity due to multiple disk types, Multi- Path I/O setup, and client mirroring Resilience Protection against failure of single adapter failure (or path) or disk Potential protection against FC adapter failures within VIOS (if Multi-Path I/O is configured) 28
双 VIO 服务器, 逻辑卷 (LV) 映射成虚拟 SCSI 盘 分区内逻辑卷镜像 Hypervisor vhost0 AIX Client LPAR A AIX Client LPAR B A A B B LVM Mirroring LVM Mirroring vscsi0 vscsi1 vscsi0 vscsi1 vhost1 vhost0 vhost1 Complexity More complicated than single VIO server but does not require SAN ports or setup Requires LVM mirroring to be setup on the client If a VIOS is rebooted, the mirrored disks will need to be resynchronized via a varyonvg on the VIOC. vtscsi0 vtscsi1 vtscsi0 vtscsi1 LV LV LV LV VIOS 1 scsi0 VIOS 2 scsi0 A A B B Resilience Protection against failure of single VIOS / SCSI disk / SCSI controller. Throughput / Scalability VIOS performance limited by single SCSI adapter and internal SCSI disks. 29
双 VIO 服务器, 硬盘 (PV) 映射成虚拟 SCSI 盘 分区内逻辑卷镜像 Hypervisor AIX Client LPAR A A LVM Mirroring vscsi0 vscsi1 AIX Client LPAR A A B B LVM Mirroring vscsi0 vscsi1 Complexity More complicated than single VIO server but does not require SAN ports or setup Requires LVM mirroring to be setup on the client If a VIOS is rebooted, the mirrored disks will need to be resynchronized via a varyonvg on the VIOC vhost0 vtscsi0 VIOS 1 scsi0 A B vhost1 vtscsi1 VIOS 2 vhost0 vtscsi0 scsi0 A B vhost1 vtscsi1 Resilience Protection against single VIOS / SCSI disk / SCSI controller The loss of a single physical disk would affect only one client Throughput / Scalability VIOS performance limited by single SCSI adapter and internal SCSI disks. 30
双 VIO 服务器, I/O 单路径,LV 或 PV 映射成虚拟 SCSI 盘 分区内逻辑卷镜像 Hypervisor vhost0 vtscsi0 AIX Client LPAR A A LVM Mirroring vscsi0 vhost1 vtscsi1 vscsi1 AIX Client LPAR B A B B LVM Mirroring vscsi0 vscsi1 vhost0 vtscsi0 vhost1 vtscsi1 Complexity Requires SAN configuration Requires LVM mirroring to be setup on the client If a VIOS is rebooted, the mirrored disks will need to be resynchronized via a varyonvg on the VIOC Resilience Protection against failure of single VIOS / FC adapter (or path) No protection against FC adapter failures within VIOS VIOS 1 31 LV LV LV LV fcs0 Note: A multi-path code (such as RDAC) may be required in the VIOS even if there is a single FC adapter. A B LV LUNs A B LV LUNs VIOS 2 fcs0 Throughput / Scalability Performance limited by a single FC adapter. Notes Disks must only be seen by one VIO server. LV VSCSI LUNs could also be PV VSCSI LUNs.
双 VIO 服务器, I/O 多路径,LV 或 PV 映射成虚拟 SCSI 盘 分区内逻辑卷镜像 Hypervisor VIOS 1 32 vhost0 vtscsi0 AIX Client LPAR 1 LVM Mirroring vscsi0 vhost1 vtscsi1 Multi-Path Driver A vscsi1 fcs0 fcs1 VIOS 2 fcs0 fcs1 A B PV LUNs AIX Client LPAR 2 A B B A B PV LUNs LVM Mirroring vscsi0 vscsi1 vhost0 vtscsi0 vhost1 vtscsi1 Multi-Path Driver Complexity Requires LVM mirroring to be setup on the VIOC Requires Multi-Path I/O setup on the VIOS If a VIOS is rebooted, the mirrored disks will need to be resynchronized via a varyonvg on the VIOC Resilience Protection against failure of single VIOS / FC adapter failure (or path) Protection against FC adapter failures within VIOS Throughput / Scalability Potential for increased bandwidth due to multi-path I/O Notes LUNs used for this purpose can only be assigned to a single VIOS LV VSCSI LUNs could also be PV VSCSI LUNs.
双 VIO 服务器, I/O 单路径,LV 或 PV 映射成虚拟 SCSI 盘 分区内 MPIO 实现硬盘冗余 AIX Client LPAR 1 A MPIO Default PCM AIX Client LPAR 2 B MPIO Default PCM Complexity Simplest dual VIOS FC option Requires MPIO to be setup on the client Hypervisor vhost0 vtscsi0 vscsi0 vhost1 vtscsi1 vscsi1 vscsi0 vscsi1 vhost0 vtscsi0 vhost1 vtscsi1 Resilience Protection against failure of a single VIOS / FC adapter (or path) Throughput / Scalability Primary LUNs can be split across multiple VIOS to help balance the I/O load VIOS 1 33 fcs0 Active Passive A B PV LUNs VIOS 2 Note: A multi-path code (such as RDAC) may be required in the VIOS even if there is a single FC adapter. fcs1 Notes Lowest port costs of dual VIOS FC options Must be PV VSCSI disks.
双 VIO 服务器, I/O 多路径,LV 或 PV 映射成虚拟 SCSI 盘 分区内 MPIO 实现硬盘冗余 Hypervisor AIX Client LPAR 1 AIX Client LPAR 2 A B MPIO Default PCM MPIO Default PCM vscsi0 vscsi1 vscsi0 vscsi1 vhost0 vhost1 vhost0 vhost1 vtscsi0 vtscsi1 vtscsi0 vtscsi1 Multi-Path Driver Multi-Path Driver VIOS 1 fcs0 fcs1 VIOS 2 fcs0 fcs1 A B PV LUNs Complexity Requires MPIO to be setup on the client Requires Multi-Path I/O setup on the VIOS Resilience Protection against failure of a single VIOS, FC adapter, or path. Protection against FC adapter failures within VIOS Throughput / Scalability Potential for increased bandwidth due to Multi-Path I/O Primary LUNs can be split across multiple VIOS to help balance the I/O load. Notes Must be PV VSCSI disks. 34 Active Passive
双 VIO 服务器, I/O 多路径,LV 或 PV 映射成虚拟 SCSI 盘 分区内 MPIO 实现硬盘冗余双机 PowerHA (HACMP) AIX Client 1 (HACMP AIX Client 2 (HACMP) A A MPIO Default PCM MPIO Default PCM vscsi0 vscsi1 vscsi0 vscsi1 PowerVM Hypervisor PowerVM Hypervisor VIOS1 vhost0 VIOS2 vhost0 VIOS1 vhost0 VIOS2 vhost0 vtscsi0 vtscsi0 vtscsi0 vtscsi0 Multi-Path Driver Multi-Path Driver Multi-Path Driver Multi-Path Driver fcs0 fcs1 fcs0 fcs1 fcs0 fcs1 fcs0 fcs1 A Active Passive 35
NPIV 简化 SAN 管理 (N_Port ID Virtualization) N_Port ID Virtualization Multiple Virtual World Wide Port Names per FC port PCIe 8 Gb adapter LPARs have direct visibility on SAN (Zoning/Masking) I/O Virtualization configuration effort is reduced Virtual SCSI Model N_Port ID Virtualization AIX Generic SCSI Disks DS8000 HDS AIX VIOS VIOS FC Adapters FC Adapters SAN SAN DS8000 HDS DS8000 HDS 36
NPIV 虚拟化分区中直接使用虚拟的 FC 卡 Client LPAR 1 A Multi-Path Software Client LPAR 2 B Multi-Path Software fcs0 fcs1 fcs2 fcs3 fcs0 fcs1 fcs2 fcs3 Hypervisor vfchost0 vfchost1 vfchost2 vfchost3 vfchost0 vfchost1 vfchost2 vfchost3 VIOS 1 fcs0 fcs1 VIOS 2 fcs0 fcs1 A B PV LUNs 37
VIOS Virtual Tape Support Enables client partitions to directly access selected SAS tape devices, sharing resources and simplifying backup & restore operations SAS adapter is owned by VIOS partition Included with PowerVM Express, Standard, or Enterprise Edition Supports AIX 5.3 & 6.1 partitions POWER6 processor-based systems VIOS SAS Adapter Virtual SCSI Adapter Virtual SCSI Adapter Power Hypervisor Tape drives supported DAT72: Feature Code 5907 DAT160: Feature Code 5619 HH LTO4: Feature Code 5746 Statement of Direction IBM intends to support VIOS virtual tape capabilities with IBM i and Linux environments in 2009. 38
内存虚拟化 动态内存共享 (Active Memory Sharing) AMS 可以动态地把内存从一个分区移动到另一个分区, 从而提高内存的使用率和灵活性 Power 系统上的内存虚拟化技术 POWER Server 内存根据分区工作负载自动分配 分区使用 VIOS 的 paging 设备 LPAR #1 LPAR #2 LPAR #3 LPAR #4 LPAR #5 提高内存的使用率 适合三种丌同的内存使用场景 PowerVM Hypervisor AMS 平均内存使用量很少 激活 / 非激活分区 分区负载的高峰时间在丌同的时段 LPAR #1 Mem LPAR #2 Mem Shared Memory Pool Virtual I/O Server Paging Devices 39
Memory Usage (GB) memory usage Memory Uage (GB) IBM Power Systems 8 5 6 3 5 6 2 1 3 4 5 6 8 5 6 1.2 0.8 0.6 AMS 技术可减少使用成本 2 25 0 根据分区工作负载自动分配内存 20 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Series1 减少内存的需求总量 1 0.4 0.2 0 1 6 4 30 15 time 10 5 0 Time Series2 Workload4 Workload3 Active / Inactive Partitions Workload2 Workload1 Dedicated Memory Shared Memory Required Physical Memory 36 GB 14 GB 4 Partitions with Dedicated Memory 4 Partitions with Shared Memory d4 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 40 35 30 25 20 15 10 5 0 Time Workload4 Workload3 Workload2 Workload1 40 35 30 25 20 15 10 5 0 time Workload4 Workload3 Workload2 Workload1 40
内存虚拟化 动态内存扩展 (Active Memory Expansion) Innovative POWER7 technology For AIX 6.1 or later For POWER7 servers Uses compression/decompression to effectively expand the true physical memory available for client workloads Often a small amount of processor resource provides a significant increase in the effective memory maximum Processor resource part of AIX partition s resource and licensing Actual expansion results dependent upon how compressible the data being used in the application A SAP ERP sample workload shows up to 100% expansion, Your results will vary Estimator tool and free trial available Will not compress AIX kernel Will not compress pinned memory pages (pinned pages to see) Will not compress File memory pages incl. code, cache & mmap (numperm to see) Already compressed data (e.g. compress database objects) will not compress well Computational memory good target (data:heap/stack, but not the code) Localized memory access pattern works well with AME 41
分区在线迁移 (Live LPAR mobility) 可以将一个分区从一台服务器上在线迁移到另外一台服务器上, 不丢失一笔交易! 降低计划的停机时间 在丌同的服务器之间, 重新部署负载 在迁移期间, 业务继续运行 Virtualized SAN and Network Infrastructure 在线分区迁移功能需要购买 PowerVM 企业版许可证, 支持 POWER6 服务器和 AIX 5.3, AIX 6.1 以及 Linux 操作系统 42
Live Partition Mobility Partition Mobility supported on POWER6 AIX 5.3, AIX 6.1 and Linux Power6 System #1 Power6 System #2 Suspended AIX Client Partition 1 Shell AIX Client Partition 1 Once Finish enough the migration memory M M M M M M M Validate pages Start environment Create and remove shell virtual migrating have partition been the SCSI M M M M M M M for moved, on memory appropriate original target devices suspend system pages LPAR the resources source definitions system A A en0 (if) en0 (if) vscsi0 ent1 ent1 vscsi0 VLAN HMC VLAN Hypervisor Hypervisor VASI vhost0 ent1 ent1 vhost0 VASI Mover Service vtscsi0 ent2 SEA en2 (if) en2 (if) ent2 SEA vtscsi0 Mover Service fcs0 ent0 VIOS VIOS ent0 fcs0 Storage Subsystem A 43
工作负载分区 (Workload Partitions WPARs) 在单一 AIX 操作系统下, 为客户应用提供独立的安全空间 是一种软件虚拟化技术 运行在 AIX V6.1 环境下 支持 POWER4,POWER5,POWER6,POWER7 服务器 两种类型的应用负载分区 System WPAR: 拥有独立 安全 隔离的环境, 类似一个完全独立的操作系统 Application WPAR: 为单一应用程序提供的隔离的运行环境 可以对 WPAR 的资源进行控制 包括 CPU, memory, paging space, threads 和 processes 的数量 Workload Partition A Workload Partition C AIX Image Workload Partition D Workload Partition B Workload Partition E 共享系统资源 操作系统 / 共享库和文本等 处理器, 内存,I/O 设备 44
AIX V6 实时应用迁移 (Live Application Mobility) NFS/SAN AIX # 1 AIX # 2 Workload Partition Database Workload Partition ERP Workload Partition QA Workload Partition Web Application Partition Dev Workload Partition Billing Workload Partition Data Mining Workload Partition Manager Policy Move a running Workload Partition from one server to another for outage avoidance and multi-system workload balancing Works on any hardware supported by AIX 6 45
PowerVM Lx86 Accelerates Linux Workload Migration PowerVM Lx86 cross-platform virtualization runs unmodified Linux/x86 applications within VMs using Linux on Power Copy x86 application binaries and run them no rewriting necessary Run Linux/x86 workloads with AIX, IBM i and Linux on Power workloads Simplifies migration and virtualization of workloads from x86-based platforms to higher-performance Power Systems servers running Linux Provides a convenient bridge to native compilation of Linux workloads Included with all PowerVM Editions Linux/x86 Application Linux x86 platform Linux/x86 Application PowerVM Lx86 Linux PowerVM Power Systems platform Linux on Power Application Linux PowerVM Power Systems platform 46
Integrated Virtualization Manager A virtualization solution for small and mid-size companies Simplifies Management Brower-based tool for creating and managing partitions Reduces Costs Eliminates the need to purchase a dedicated hardware console Included with all PowerVM Editions Runs in the Virtual I/O Server partition 47
PowerVM 虚拟化功能汇总 AIX 5.3 AIX 6.1 IBM i 6.1.1 RHEL 5.4 SLES V10 SP3 DLPAR Processor add/remove Yes Yes Yes Yes Yes Yes DLPAR I/O adapter add/remove Yes Yes Yes Yes Yes Yes DLPAR Memory add Yes Yes Yes Yes Yes Yes DLPAR Memory remove Yes Yes Yes No No Yes Micro Partitioning Yes Yes Yes Yes Yes Yes Capacity Upgrade on Demand Yes Yes Yes Yes Yes Yes Virtual I/O Server Yes Yes Yes Yes Yes Yes Integrated Virtualization Manager Yes Yes Yes Yes Yes Yes Virtual SCSI (VIO Server) Yes Yes Yes Yes Yes Yes Virtual Ethernet (VIO Server) Yes Yes Yes Yes Yes Yes Workload Partitions (AIX only) No Yes No No No No Application Mobility (AIX only) No Yes No No No No Lx86 (Statement of direction for POWER7) No No No Yes Yes Yes Shared Dedicated Capacity (POWER6) Yes Yes Yes Yes Yes Yes Multiple Processor Pools (POWER6) Yes Yes Yes Yes Yes Yes Live Partition Mobility (POWER6) Yes Yes No Yes Yes Yes N_Port ID Virtualization (POWER6) Yes Yes Yes Yes No Yes Active Memory Expansion (POWER7) No Yes No No No No Active Memory Sharing (POWER6) Yes Yes Yes No No Yes Integrated Virtual Ethernet (POWER6) Yes Yes Yes Yes Yes Yes Active Memory Expansion (POWER7 / AIX 6.1) No Yes No No No No Note: Some features are supported with specific fix levels or only on specific hardware models. Check the appropriate documentation. SLES V11 48
PowerVM 的三种版本 PowerVM Express Edition Evaluations, pilots, PoCs Single-server projects PowerVM Standard Edition Production deployments Server consolidation PowerVM Enterprise Edition Multi-server deployments Cloud infrastructure PowerVM Edition Express Standard Enterprise 服务器 最大分区数 管理平台 Power Blade, 520, 550 2 DLPARS +1 VIOS per Server VMControl IVM Power Blade, Power System* 10 / Core (up to 1000) VMControl IVM, HMC Power Blade Power System* 10 / Core (up to 1000) VMControl IVM, HMC 虚拟 IO 服务器 分区在线迁移 动态内存共享 PowerVM Editions offer a unified virtualization solution for any Power workloads 多 CPU 共享池 共享专有分区 客户分区支持 AIX / Linux / i AIX / Linux / i AIX / Linux / i Lx86 49
实施服务器虚拟化的考虑 50
实施虚拟化的步骤 1. 明确需求 组织架构 服务器或应用部署方式等 2. CPU 虚拟化 Its easy and is appropriate for almost all workloads. Understand multiple shared pools 3. 操作系统引导盘 (rootvg) 虚拟化 Internal disks can be the first limiting factor in number of LPARs Need to understand multi-path requirements 4. 以太网虚拟化 Many NICs run at very low utilization Understand VLAN, link aggregation, and connectivity requirements 5. 非 rootvg 虚拟化 - virtual SCSI, virtual FC or both 6. 内存虚拟化 requires virtual CPU and I/O 7. 迁移性 requires virtual CPU and I/O 51
PowerVM 虚拟化技术的选择 Dedicated CPU, Memory, I/O Shared CPU Shared CPU and I/O I/O I/O I/O I/O VIOS V-NET Memory Memory Memory Memory Memory Memory V-SCSI V-FC CPU CPU Shared CPU Shared CPU PowerVM Hypervisor PowerVM Hypervisor PowerVM Hypervisor Dedicated LPAR Benefits More flexible & granular than stand alone servers Good for high peak to average CPU utilization Considerations Considerable unused CPU I/O, and memory capacity Potentially higher software costs Shared CPU Benefits Much better use of CPU resources Lowers software costs Considerations Boot drives can limit number of configurable LPARs Significant low use I/O cards Memory dedicated to low use environments Virtual I/O Benefits No limitation on # of boot devices more LPARs Sufficient I/O for capacity availability, & connectivity Good use of CPU resource Considerations Memory dedicated to low use environments may still limit consolidation 52
PowerVM 虚拟化技术的选择 ( 续 ) Shared CPU, I/O, Memory Multiple CPU Pools LPAR Mobility LPAR Mobility Pool of Shared Memory Shared CPU VIOS V-NET V-SCSI V-FC Paging Shared or Dedicated Memory CPU Pool CPU Pool VIOS V-NET V-SCSI V-FC Paging All Virtual I/O Memory CPU Pools Memory CPU Pools All Virtual I/O PowerVM Hypervisor AMS PowerVM Hypervisor AMS PowerVM Hypervisor PowerVM Hypervisor Shared Memory Benefits Increase memory utilization especially for low utilization environments Can mix shared and dedicated memory LPARs on the same frame May allow for increased CPU utilization where previously limited by memory Requires virtual I/O & CPU Multiple CPU Pools Allows CPU resources to be grouped by middleware, department, etc May lower core based software licensing Can guarantee a group of LPARs a given amount of CPU capacity LPAR Mobility Move running LPARs between servers to improve availability Move dedicated memory to dedicated memory and shared to shared Can move LPARs to specific CPU pools for licensing purposes Requires virtual I/O 53
服务器虚拟化的案例 54
某行数据中心 ( 北京 ) 测试环境 PowerVM 方案 数据中心 ( 北京 ) 骨干网 4 x 4 x Ethernet Switch 以太网络连接 4 x 4 x FCx2 FCx2 光纤网络连接 FCx2 FCx2 Power570-1 SAN 光纤交换机 SAN 光纤交换机 Power570-2 HDS 存储阵列 解决方案描述 采用 PowerVM 虚拟化技术, 利用有限的硬件资源创建了 20 个分区满足测试需要 ; 采用 PowerVM 微分区技术, 使用较少的 CPU 满足大量测试分区需要 ; 采用 PowerVM 高可用技术, 整个系统无单点故障 ; 解决方案优势 采用 PowerVM 虚拟化技术, 在节约成本的同时能够快捷 轻松地满足以后测试要求 ; 采用 PowerVM 微分区技术, 能够提高系统整机利用率 ; 整合服务器和虚拟化资源, 减少硬件, 软件, 管理和能源成本 ; 55
IBM POWER 服务器虚拟化在交行 2003 年 -2004 年 5 月, 大集中一期生产环境, 使用 LPAR 技术 2006~2007, 在数据中心生产环境, 全面部署 DLPAR 方案 2009 年 6 月后, 新上线应用系统启用微分区, 实现资源自劢调整功能 2009 年 10 月, 活劢分区迁移 LPM 在数据中心的测试环境和准生产环境投产应用 2003 2004 2005 2006 2007 2008 2009 2005 年 3 月, 大集中二期 UAT 环境, 尝试 P5 570+ 微分区 + VIO 技术 2008 年, 微分区和 VIOS 技术在开发测试环境全面推广 2008 年 5 月起, 生产系统 Web 层部署微分区和 VIOS 56
SAP 客户虚拟化案例 Customer environment exploiting virtualization capabilities on POWER5 systems Total number of SAP systems: 77 (PROD + DEV + QAS + idle HA partitions) Using SAP Quicksizer Total number of CPU = 189 (from 0.25 core up to 8 cores) Total amount of memory = 1.5 TB (assuming 8 GB/p5 core) Implemented on 4x p570 12-way (16-way o/w 4 CUoD) Total number of active CPU = 48 (= 4 x 12) Total amount of memory = 672 GB (Total = 1024 GB o/w 352 for CUoD) SAP productive partitions can expand up to 12 CPUs All LPARs are shared and uncapped configured Main memory Less in total, but more per CPU 57
SAP 客户虚拟化案例 ( 续 ) Customer environment exploiting virtualization capabilities on Power5 systems (4x p570, each with 12 active CPU + 4 CPU CUoD) 58
Example: Distributing 21 SAP instances on 23 LPARs within one single p570 system (12 active CPUs, 4 CoD) SAP-System LPAR Memory (in G) CPU-Entitlement vcpu Weighting ID Minimum Desired Maximum Desired Minimum Maximum desired min max uncapped VIO-Server1 2 1 2 4 0,4 0,2 2 2 1 4 128 VIO-Server2 3 1 2 4 0,4 0,2 2 2 1 4 128 P33 4 4 6 12 0,4 0,2 1,2 2 1 2 64 P23 5 4 9 18 0,5 0,3 2 3 1 8 64 CS-Test 6 1 2 4 0,1 0,1 0,4 1 1 4 32 C03 7 2 6 12 0,3 0,2 0,8 2 1 2 32 C17 8 1 2 4 0,2 0,2 0,5 2 1 2 32 C21 9 2 3 6 0,2 0,2 0,5 2 1 2 32 C07 10 2 6 12 0,4 0,2 0,9 2 1 4 32 P07_idle 11 1/4 16 16 0,4 0,1 4 4 1 8 64 P31_idle 12 4 6 12 0,4 0,2 0,8 2 1 8 64 I03 13 1 2 4 0,1 0,1 0,3 1 1 3 32 C04 14 3 8 16 0,1 0,1 0,5 1 1 5 32 T98 15 2 8 16 0,1 0,1 0,5 1 1 5 16 C91 16 2 6 12 0,1 0,1 0,5 1 1 5 32 IDMS-X 17 1 4 8 0,1 0,1 0,5 1 1 5 32 A41 18 1 4 8 0,1 0,1 0,5 1 1 5 32 K07 19 1 4 8 0,1 0,1 0,5 1 1 5 16 IDMSXP2 20 1 4 8 0,2 0,1 0,5 1 1 5 64 sideprod 21 4 5 6 0,1 0,1 0,5 1 1 5 64 Q04 22 4 8 16 0,1 0,1 0,5 1 1 5 32 P04_idle 23 1/4 8 16 0,2 0,1 0,5 2 1 5 64 Q72 24 4 20 20 1,6 0,4 2 8 2 8 32 MME RZ2 141 242 4,7 3,0 19,9 44 Production Qualityassurance Development K-Backup (idle LPAR) Virtual-I/O Server 59
PowerHA 与 PowerVM 的结合 备份机实现 HA 或 DR 功能 处理能力共享 提高服务器整体使用率 生产系统 App 3 App 2 App 1 备份系统 & 其它应用系统 其它应用 3 其它应用 1 App 3 备份 App HA 3 备份 1 其它应用 2 其它应用 2 App 2 备份 App HA 2 备份 2 其它应用 1 其它应用 1 App 1 备份 App HA 1 备份 3 多台服务器 or 服务器上的多个分区 一台服务器 60
虚拟化的管理 61
平台的演化 Systems Director VMControl 像管理单一系统一样来管理系统池 更多的虚拟, 更少的管理 迁移 SW OS SW OS SW OS 应用软件 操作系统 Virtual Server Virtual Server Virtual Server Virtualization 处理器 内存 存储 网络 处理区 内存 存储 网络 优化了. 可用性 性能 能源 服务器 Servers 虚拟服务器 Virtual Servers 系统池 System Pools 62 62
对虚拟资源进行最有效的控制和管理 Systems Director VMControl 云计算 - 就绪 易捷版 - 虚拟机 基于业务需求选择最佳实践匹配 创建像构建块一样的标准化的虚拟机 选择默认的属性 标准版 - 管理虚拟映像 映像库 Image Image App Image App Image OS App OS App OS OS 捕获并收编数据中心中使用的虚拟映像 标准化虚拟映像构建块 定制虚拟环境运行时需求 通过虚拟映像简化部署 企业版 - 优化系统池 像管理单一系统一样管理系统池 把多系统一体化管理 自动安排新的工作负载 统一为固件更新管理配置 跨硬件代的持久的 即插即用的能力 工作负载基于服务水平目标集中管理 确保达到服务水平承诺 完整的 IT 过程的虚拟化管理 持续可用 弹性扩充 自动供给 63
Power 虚拟化不损失性能 所有发布指标均在虚拟环境下测试 8,000,000 6,000,000 4,000,000 2,000,000 0 tpmc 4-core 8-core 16-core 64-core IBM TPC-C 测试结果 IBM 是目前公布带 PowerVM 等虚拟技术 TPC-C 测试结果的唯一供应商, 测试结果表明 Power 具有更强的线性扩展能力 信息来源 :http://www.tpc.org 表中所示为 10/16/09 测试结果 用户数 / 内核 用户数 / 内核 600 230 500 400 300 200 100 220 210 200 190 0 全部资源分区 部分资源分区 ( 虚拟 cpu) 180 本机模式下单系统 OS 映像 VMware 虚拟化 64 IBM SAP SD 2-Tie 全部资源分区 与 部分资源分区 ( 虚拟 cpu) 对比 IBM 全部资源分区 与采用 部分资源分区 ( 虚拟 CPU) 的 SAP SD 2-tier 测试结果对比, 两种测试结果每个内核用户数大致相同, 说明 IBM Power 虚拟化没有损失任何性能 信息来源 :http://www.sap.com/benchmark SAP SD 2-tier 本机模式下单系统 OS 映像 与 VMware 虚拟化 对比 本机模式下单系统 OS 映像 与采用 VMware 虚拟化 的测试结果对比, 两种结果之间存在很大差别 虚拟化 测试结果每个内核用户数更少, 说明 Vmware 虚拟化对性能带来了非常大的损失 信息来源 :http://www.sap.com/benchmark
智慧的地球需要智慧的系统 65