More Related Content Similar to IBM I and blade center update 2009 (20) IBM I and blade center update 20091. IBM i and BladeCenter 2Q 2009 Update
Vess Natchev and Kyle Wurgler
vess@us.ibm.com, wurgler@us.ibm.com,
IBM Systems Lab Services and Training
8 Copyright IBM Corporation, 2009. All Rights Reserved.
This publication may refer to products that are not currently
available in your country. IBM makes no commitment to make
available any products referred to herein.
2. Agenda
• Where to start with IBM i on blade
• Hardware overview:
– Power blade servers technical overview
– New expansion adapters
– BladeCenter S components and I/O connections
– BladeCenter H components and I/O connections
– Switch module portfolio
– Expansion adapter portfolio for IBM i
– Feature codes and ordering
• Virtualization overview
– VIOS-based virtualization overview
– I/O options for BladeCenter H and BladeCenter S
– Configuring storage for IBM i on blade
– Configuring storage with the SAS RAID Controller Module
– Virtual tape
– Multiple Virtual SCSI adapters
– Active Memory Sharing on blade
2
© 2009 IBM Corporation
3. IBM i on Blade: Where Do I Start?
• New versions by May 22 at: http://www.ibm.com/systems/power/hardware/blades/ibmi.html
3
© 2009 IBM Corporation
4. IBM BladeCenter JS23 Express
• 2 sockets, 4 POWER6 cores @ 4.2 GHz
• Enhanced 65-nm lithography
• 32 MB L3 cache per socket
• 4 MB L2 cache per core
• 8 VLP DIMM slots, up to 64 GB memory
• FSP-1 service processor
• 2 x 1Gb embedded Ethernet ports (HEA)
• 2 PCIe connectors (CIOv and CFFh)
• 1 x onboard SAS controller
• Up to 1 SSD or SAS onboard disk
• EnergyScale™ power management
• PowerVM Hypervisor virtualization
4
© 2009 IBM Corporation
6. IBM BladeCenter JS43 Express
• 4 sockets, 8 POWER6 cores @ 4.2 GHz
• Enhanced 65-nm lithography
• 32 MB L3 cache per socket
• 4 MB L2 cache per core
• 16 VLP DIMM slots, up to 128 GB memory
• FSP-1 service processor
+ •
•
4 x 1Gb embedded Ethernet ports (HEA)
4 PCIe connectors (CIOv and CFFh)
• 1 x onboard SAS controller
• Up to 2 SSD or SAS onboard disks
• EnergyScale™ power management
• PowerVM Hypervisor virtualization
6
© 2009 IBM Corporation
8. CFFv and CFFh I/O Expansion Adapters
Combination Form Factor (CFF) allows for 2
different expansion adapters on the same HSSM1
blade HSSM3
CFFv (Combo Form Factor – Vertical)
Connects to PCI-X bus to provide access to switch CFFX
CFFv SM3
modules in bays 3 & 4 SerDes SM4
Vertical switch form factor PCI-X
Supported for IBM i: SAS (#8250)
CFFh (Combo Form Factor – Horizontal) HSSM2
HSSM4
Connects to PCIe bus to provide access to the switch CFFE
CFFh
modules in bays 7 – 10
Horizontal switch form factor, unless MSIM used
PCI-Express
Supported for IBM i: Fibre Channel and Ethernet
(#8252)
Note: See IBM i on Power Blade Supported Environments for
hardware supported by IBM i:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html 8
© 2009 IBM Corporation
9. CIOv and CFFh I/O Expansion Adapters
Combination I/O Form Factor – Vertical
(CIOv) is available only on JS23 and JS43
CFFv adapters not supported on JS23 and
JS43
CIOv
Connects to new PCIe bus to provide access to switch
modules in bays 3 & 4
Vertical switch form factor
Supported for IBM i: SAS passthrough (#8246),
Fibre Channel (#8240, #8241, #8242)
Can provide redundant FC adapters with CFFh
CFFh
Connects to PCIe bus to provide access to the switch
modules in bays 7 – 10
Horizontal switch form factor, unless MSIM used
Supported for IBM i: Fibre Channel and Ethernet
(#8252) Note: See IBM i on Power Blade Supported Environments for hardware
supported by IBM i:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html
9
© 2009 IBM Corporation
10. Meet the BladeCenter S – Front View
Service label cards slot enable quick and easy
reference to BladeCenter S SAS and SATA disks can be
mixed
SAS disks recommended for
IBM i production
RAID 0, 1, 5, 0+1 supported
with RAID SAS Switch
Module (RSSM)
7U Separate RAID arrays for
IBM i recommended
Shared USB ports Supports up to 6 BladeServers
and CD-RW / DVD-
ROM Combo Battery Backup Units for use only with RAID
SAS Switch Module
10
© 2009 IBM Corporation
11. Meet the BladeCenter S – Rear View
Hot-swap Power Supplies 3 & 4 are Hot-swap Power Supplies 1 &
optional, Auto-sensing b/w 950W / 2 are standard, Auto-sensing
1450W b/w 950W / 1450W
Power supplies 3 and 4
required if using > 1 blade
7U
Top: AMM standard
Four Blower Bottom: Serial Pass-thru Module optional
modules standard Top(SW1) & Bottom(SW2) left: Ethernet
Top(SW3) & Bottom(SW4) right: SAS
Both CIOv (#8246) and CFFv (#8250) adapters
supported 11
© 2009 IBM Corporation
12. BladeCenter S Midplane - Blade to I/O Bay Mapping
AMM Bay
Blade “A”
I/O Bay 1
#1
Ethernet Bay
“B”
Blade #2
Blade #3
Blade #4
Blade #5 I/O Bay 3
Blade #6 ENet Switch
Fibre SAS Switch Bay
SAS
“A”
“B” RAID Battery Bay
PCI-X (CFFv) or PCIe (CIOv) D.C. Blade
D.C. Blade
#1
Blade Daughter Card D.C. Blade
#2
eNet, Fibre, SAS, SAS RAID D.C. Blade
#3 I/O Bay 4
#4
D.C. Blade ENet Switch
#5
D.C. Blade Fibre
#6 SAS
SAS Switch Bay
“A”
RAID Battery Bay
“B”
C.C. Blade I/O Bay 2
PCI-E (CFFh) C.C. Blade
#1
Blade Daughter Card C.C. Blade
#2 Option Bay
C.C. Blade
#3
C.C. Blade
#4
C.C. Blade
#5
#6
BC-S Mid-Plane
12
© 2009 IBM Corporation
13. BladeCenter H - front view
Power
Module 3
Power Filler
Module 1
and Fan
pack Front
System
HS20 Panel
Blade # 1
9U CD DVD- drive
Blade Filler
Front USB
Power Power
Module 2 Module 4
Filler and Fan
pack
13
© 2009 IBM Corporation
14. IBM BladeCenter H - Rear View
• Multi-Switch Interconnect Module
• Ethernet switch (left side bay 9)
I/O module bay
7 and 8 • Fibre Channel switch (right side bay 10)
Power Power
Connector 2 Connector 1
Ethernet I/O Module SAS or
switch I/O Module bay 1
bay 3 Fibre
I/O Module bay 5 Advanced Channel
Management module
Blower Module 1 Module 1
and 2 Advanced
Ethernet Management
switch I/O Module bay 2 Module 2 slot
I/O Module bay 6 I/O Module
bay 4
Rear LED panel
and Serial
connector Left Shuttle Right Shuttle release lever
release lever I/O module bay 9 and 10
• Multi-Switch Interconnect Module
• Ethernet switch (left side bay 9)
• Fibre Channel switch (right side bay 10)
14
© 2009 IBM Corporation
15. BCH: CFFv and CFFh I/O Connections
Blade #N Switch #1
On-Board Dual
Gbit Ethernet Ethernet
On-Board Dual
Gbit Ethernet M
I Switch #2
POWER SAS CFFv Ethernet
Expansion Card D
Blade Server #1 P
L Switch #3
QLogic CFFh A
Expansion Card
N Switch #4
E
QLogic CFFh Expansion Card: Switch #7
• Provides 2 x 4Gb Fibre Channel connections to SAN
• 2 Fibre Channel ports externalized via Switch 8 & 10
• Provides 2 x 1 Gb Ethernet ports for additional Switch #8
networking
• 2 Ethernet ports externalized via Switch 7 & 9
Switch #9
SAS CFFv Expansion Card:
• Provides 2 SAS ports for connection to SAS tape drive Switch #10
• 2 SAS ports externalized via Switch 3 & 4 15
© 2009 IBM Corporation
16. BCH: CIOv and CFFh I/O Connections
Blade #N Switch #1
On-Board Dual
Gbit Ethernet Ethernet
On-Board Dual
Gbit Ethernet M
I Switch #2
POWER CIOv Expansion Ethernet
Card D
Blade Server #1 P
L Switch #3
QLogic CFFh A
Expansion Card
N Switch #4
CIOv Expansion Card:
E
Switch #7
• 2 x 8Gb or 2 x 4Gb Fibre Channel
• OR, 2 x 3Gb SAS passthrough
• Uses 4Gb or 8Gb FC vertical switches in bays 3 & 4 Switch #8
• OR, 3Gb SAS vertical switches in bays 3 & 4
• Redundant FC storage connection option for IBM i
CFFh Expansion Card: Switch #9
• 2 x 4Gb and 2 x 1Gb Ethernet
Switch #10
16
© 2009 IBM Corporation
17. BladeCenter Ethernet I/O Modules
Nortel Layer 2/3 Gb Cisco Systems Nortel L2-7 GbE Switch Nortel L2/3 10GbE
Ethernet Switch Intelligent Gb Ethernet Module Uplink Switch Module
Modules Switch Module
Copper Pass-Through Nortel 10Gb Ethernet Intelligent Copper
Module Switch Module Pass-Through Module
Note: See IBM i on Power Blade Supported Environments for
hardware supported by IBM i:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html 17
© 2009 IBM Corporation
18. BladeCenter Fibre Channel I/O Modules
Cisco 4Gb 10 and 20 Brocade 4Gb 10 and QLogic 8Gb 20 port QLogic 4Gb 10 and 20
port Fibre Channel 20 port Fibre Channel Fibre Channel Switch port Fibre Channel
Switch Modules Switch Modules Module Switch Module
Brocade Intelligent 8Gb Brocade Intelligent 4Gb
Pass-Thru Fibre Channel Pass-Thru Fibre Channel
Switch Module Switch Module
Note: See IBM i on Power Blade Supported Environments for
hardware supported by IBM i:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html 18
© 2009 IBM Corporation
19. BladeCenter SAS I/O Modules
BladeCenter S SAS
RAID Controller
Module (FC #3734)
• Supported only in BladeCenter S
• RAID support for SAS drives in chassis
• Supports TS2240 attachment
• No support for attaching DS3200
• 2 are always required
BladeCenter SAS
Controller Module
(FC #3267)
• Supported in BladeCenter S and BladeCenter H
• No RAID support
• Supports TS2240 attachment
• Supports DS3200 attachment
• 1 is required, 2 recommended
Note: See IBM i on Power Blade Supported Environments for
hardware supported by IBM i:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html 19
© 2009 IBM Corporation
20. SAS RAID Controller Switch Module
• RAID controller support provides additional
protection options for BladeCenter S storage
• SAS RAID Controller Switch Module
– High-performance, fully duplex, 3Gbps speeds
– Support for RAID 0, 1, 5, & 10
– Supports 2 disk storage modules with up to 12 SAS drives
– Supports external SAS tape drive
– Supports existing #8250 CFFv SAS adapter on blade
– Supports new #8246 CIOv SAS passthrough adapter
– 1GB of battery-backed write cache between the 2 modules
– Two SAS RAID Controller Switch Modules (#3734) required
• Supports Power and x86 Blades
– Recommend separate RAID sets
• For each IBM i partition
• For IBM i and Windows storage
– Requirements
• Firmware update for SAS RAID Controller Switch Modules
• VIOS 2.1.1, eFW 3.4.2
Note: Does not support connection to DS3200
IBM i is not pre-installed with RSSM configurations
20
© 2009 IBM Corporation
21. Multi-switch Interconnect Module for BCH
• Installed in high-speed bays 7 & 8
and/or 9 & 10
• Allows a “vertical” switch to be
MSIM
installed and use the “horizontal” high-
speed fabric (bays 7 – 10)
• High-speed fabric is used by CFFh
expansion adapters
• Fibre Channel switch module must be
installed in right I/O module bay (switch
bay 8 or 10)
• If additional Ethernet networking
required additional Ethernet switch
module can be installed in left I/O
module bay (switch bay 7 or 9)
21
© 2009 IBM Corporation
22. I/O Expansion Adapters for IBM i
#8252 QLogic Ethernet and 4Gb Fibre #8250 LSI 3Gb SAS Dual
Channel Expansion Card (CFFh) Port Expansion Card (CFFv)
#8246 3Gb SAS #8240 Emulex 8Gb #8242 QLogic 8Gb #8241 QLogic 4Gb
Passthrough Expansion Fibre Channel Fibre Channel Fibre Channel
Card (CIOv) Expansion Card (CIOv) Expansion Card (CIOv) Expansion Card (CIOv)
Note: See IBM i on Power Blade Supported Environments for
hardware supported by IBM i:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html
22
© 2009 IBM Corporation
23. IBM BladeCenter S Configuration for IBM i on Power Blade
Description Part# Feature Notes
Chassis IBM BladeCenter S 8886-xxx 7779-BCS
AMM Advanced Management Module 25R5778 3201 One Standard
Power AC Power Module 43W3582 4548 Two Standard, Two optional
SAS SAS Connectivity Module 39Y9195 3267 One Required, 2nd optional
Module SAS RAID Controller Connectivity Module 43W3584 3734 Two always required
Disk IBM BladeCenter S 6-Disk Storage Module 43W3581 4545 One Required for each 6 disk
Storage drives, max of two
Modules 73 GB 15K RPM SAS Disk Drive 43W7523 3748
and SAS One Required, max of 12 disks
Disk Drives 146 GB 15K RPM SAS Disk Drive 43W7524 3749
300 GB 15K RPM SAS Disk Drive 43X0802 3747
450 GB 15K RPM SAS Disk Drive 42D0519 3762
Ethernet Nortel Networks L2/L3 Copper Gb Ethernet Switch Mod 32R1860 3212 One Required
Switch Nortel Networks L2/L3 Fibre Gb Ethernet Switch Mod 32R1861 3213
Nortel Networks L2-7 Gb Ethernet Switch Module 32R1859 3211
Cisco Catalyst Ethernet Switch Module - 3012 43W4395 3174
IBM BladeCenter Copper Passthru Module 39Y9320 3219
IBM BladeCenter Optical Passthru Module 39Y9316 3218
Server Connectivity Module 39Y9324 3220
Nortel 10Gb Uplink Ethernet Switch Module 32R1783 3210
Intelligent Copper Pass-Thru Module for IBM BladeCenter 44W4483 5452
http://www.ibm.com/systems/power/hardware/blades/supported_environments.pdf
23
© 2009 IBM Corporation
24. IBM BladeCenter H Configuration for IBM i on Power Blade
Description Part# Feature Notes
Chassis IBM BladeCenter H 8852-xxx 7989-BCH
AMM Advanced Management Module 25R5778 3201 1 Standard, 2nd optional
Power AC Power Module 31R3335 3200 1 Standard, 2nd optional
SAN Fibre Switch Brocade 10-port 4 Gb SAN Switch Module 32R1813 3207 One required, 2nd optional
Brocade 20-port 4 Gb SAN Switch Module 32R1812 3206
QLogic 10-port 4Gb SAN Switch Module 43W6724 3243
QLogic 20-port 4Gb SAN Switch Module 43W6723 3244
QLogic® 20-port 8Gb SAN Switch Module 44X1905 3284 other SAN Fibre switches supported
Cisco Systems 4Gb 10-port Fibre Channel Module 39Y9284 3241 see Supported Environments PDF
Cisco Systems 4Gb 20-port Fibre Channel Module 39Y9280 3242
Ethernet Switch Cisco Systems Intelligent GbE Ethernet Switch Module 32R1892 3215 One required, 2nd optional
Nortel Networks L2/L3 Copper Gb Ethernet Switch Mod 32R1860 3212
Nortel Networks L2/L3 Copper Gb Ethernet Switch Mod 32R1859 3211
IBM BladeCenter Copper Passthru Module 39Y9320 3219 other Ethernet switches supported
Server Connectivity Module 39Y9324 3220 see Supported Environments PDF
SAS Switch SAS Connectively Module 39Y9195 3267 Optional for Tape attachment
MSIM Multi-Switch Interconnect Module 39Y9314 3239 One Required per SAN Fibre switch
SFP IBM Short Wave SFP Module 22R4902 3238 1 per active port on SAN switch
IBM Long Wave SFP Module 19K1272 3237
Cisco Systems Short Wave SFP Module 41Y8598 3261
Cisco Systems Long Wave SFP Module 42Y8600 3262
Other Power Cords, Cables, Publications
http://www.ibm.com/systems/power/hardware/blades/supported_environments.pdf
24
© 2009 IBM Corporation
25. IBM BladeCenter JS12 Configuration
Description Feature Notes
Blade IBM BladeCenter JS12 2-core, 3.8 GHz 8442 7998-60X
Processor Processor Entitlement (Qty 2) 8444 Two processor entitlements required
or with Express Configuration
Processor Entitlement (Qty 1) 8444
Zero-priced Processor Entitlement (Qty 1) 8443
Memory 4 GB (2 x 2 GB) DDR2 667 MHz DIMMs 8229 One required, max of four
8 GB (2 x 4 GB) DDR2 667 MHz DIMMs 8239
16 GB (2 x 8 GB) DDR2 533 MHz DIMMs 8245
Disk IBM 73 GB SAS 10K SFF HDD 8237 One required, max of two
IBM 146 GB SAS 10K SFF HDD 8236
SAS Adapter SAS Expansion Card (CFFv) 8250 • Required for SAS Disk and Tape in BCS
• Optional for tape connection in BCH*
Fibre Adapter QLogic Ethernet and 4 GB Fibre Channel 8252 • Not supported in BCS
Expansion Card (CFFh) • Required for SAN connection in BCH*
PowerVM PowerVM Standard Edition (Qty 2) with 5409 Required
VIOS 1.5 with latest service pack
Software Preinstall 5005, 8146 Optional preinstall of VIOS
http://www.ibm.com/systems/power/hardware/blades/supported_environments.pdf
25
© 2009 IBM Corporation
26. IBM BladeCenter JS22 Configuration
Description Feature Notes
Blade IBM BladeCenter JS22 4-core, 4.0 GHz 8400 7998-61X
Processor Processor Entitlement (Qty 4) 8401 Four processor entitlements
or with Express Configuration required
Processor Entitlement (Qty 2) 8401
Zero-priced Processor Entitlement (Qty 2) 8399
Memory 4 GB (2 x 2 GB) DDR2 667 MHz DIMMs 8233 One required
8 GB (2 x 4 GB) DDR2 667 MHz DIMMs 8234 Option second pair
16 GB (2 x 8 GB) DDR2 533 MHz DIMMs 8235
Disk IBM 73 GB SAS 10K SFF HDD 8237 One required
IBM 146 GB SAS 10K SFF HDD 8236
Fibre Adapter QLogic Ethernet and 4 GB Fibre Channel Expansion 8252 Required for connection to SAN
Card (CFFh)
PowerVM PowerVM Standard Edition (Qty 4) with VIOS 1.5 with 5409 Required
latest service pack
Software Preinstall 5005, 8146 Optional preinstall of VIOS
SAS Adapter SAS Expansion Card (CFFv) 8250 Optional for connection to SAS
Tape
• Plus
– IBM i Processor and User Entitlements
– SAN – DS3200, DS3400, DS4700, DS4800, DS8100, DS8300
– SAS Tape – TS2230 or TS2240 (Optional; virtual tape supported only with TS2240)
– IBM i LAN Console
Note: A minimum of one copy of the Service Warranty Publications (#8259) and one copy of the JS22 Installation and User's Guide (#8260-8263, #8266-8269, or #8278-8281) is required at each customer installation.
26
© 2009 IBM Corporation
27. IBM BladeCenter JS23 and JS43 Configuration
Description Feature Notes
Blade IBM BladeCenter JS23 4-core, 4.2 GHz with L3 cache 7778-23X
IBM BladeCenter JS43 8-core, 4.2 GHz with L3 cache 7778-23X with FC #8446
Processor Processor Entitlement (Qty 4) 8395 Four processor entitlements
or with Express Configuration 8395 required
Processor Entitlement (Qty 2) 8393
Zero-priced Processor Entitlement (Qty 2)
Memory 4 GB (2 x 2 GB) DDR2 667 MHz DIMMs 8233 One required
8 GB (2 x 4 GB) DDR2 667 MHz DIMMs 8234 Option second pair
16 GB (2 x 8 GB) DDR2 533 MHz DIMMs 8235
Disk IBM 73 GB SAS 10K SFF HDD 8237 One optional
IBM 146 GB SAS 10K SFF HDD 8236
IBM 300GB SAS 10K SFF HDD 8274
IBM 69GB SFF SAS Solid State Drive-Blade 8273
Fibre Adapter QLogic Ethernet and 4 GB Fibre Channel Expansion Card 8252 One required for connection to SAN
(CFFh) unless DS3200 used
QLogic 8 Gb Fibre Channel Expansion Card (CIOv) 8242
Emulex 8 Gb Fibre Channel Expansion Card (CIOv) 8240
QLogic 4 Gb Fibre Channel Expansion Card (CIOv) 8241
PowerVM PowerVM Standard Edition (Qty 4) with VIOS 1.5 with latest 5409 Required
service pack
Software Preinstall 5005, 8146 Optional preinstall of VIOS
SAS Adapter SAS Passthrough Expansion Card (CIOv) 8246 Optional for connection to DS3200
or SAS tape
• Plus
– IBM i Processor and User Entitlements
– SAN – DS3200, DS3400, DS4700, DS4800, DS8100, DS8300
– SAS Tape – TS2230 or TS2240 (Optional; virtual tape supported only with TS2240)
– IBM i LAN Console
27
© 2009 IBM Corporation
29. VIOS, IVM and i on Power Blade
Linux AIX VIOS = Virtual I/O Server =
Client Client virtualization software in a partition
HEA HEA HEA HEA Does not run other applications
First LPAR installed on blade
VIOS owns physical hardware (Fibre
CFFh FC USB
and/or
CFFv SAS exp card Channel, Ethernet, DVD, SAS)
exp card SAS HEA
or CIOv FC
exp card or VIOS virtualizes disk, DVD,
CIOv SAS
exp card SSD
VIOS / IVM networking, tape to i partitions
SAS Switch FC Switch IVM = Integrated Virtualization
Manager = browser interface to manage
DS3400 LAN
DVD
DS4700 partitions, virtualization
DS3200* DS4800
IVM / Virtual Op Panel
SAS-attached DS8100 IVM installed with VIOS
LTO4 tape drive DS8300
(virtual tape)
SVC i uses LAN console through Virtual
AMM / LAN Console
Ethernet bridge in VIOS
* Not supported with RSSM
29
© 2009 IBM Corporation
30. Storage, Tape and DVD for i on JS12/JS22 in BCH
MSIM with Fibre Channel
I/O module inside VIOS Host i Client
Fibre Channel
Storage hdiskX LUNs DDxx
CFFh
Fibre Channel
BladeCenter midplane
I/O module
Virtual SCSI
connection
SAS Storage and/or tape
DS3200
CFFv Virtual SCSI
SAS I/O module connection
TS2240
USB OPTxx
/dev/cd0 DVD
DVD
Media tray
Power Blade
With BCH and JS12/JS22, IBM i can use:
Fibre Channel storage (MSIM, FC module and CFFh adapter required)
SAS storage (SAS module and CFFv adapter required)
SAS tape (SAS module and CFFv adapter required)
USB DVD in BladeCenter
Physical I/O resources are attached to VIOS, assigned to IBM i in IVM
Storage LUNs (physical volumes) assigned directly to IBM i; storage pools in VIOS not used
30
© 2009 IBM Corporation
31. Storage, Tape and DVD for i on JS23/JS43 in BCH
MSIM with Fibre Channel
I/O module inside VIOS Host i Client
Fibre Channel
Storage hdiskX LUNs DDxx
CFFh
Fibre Channel
BladeCenter midplane
I/O module
Virtual SCSI
connection
CIOv
SAS Storage and/or tape
OR
DS3200 CIOv Virtual SCSI
SAS I/O module connection
TS2240
USB OPTxx
/dev/cd0 DVD
DVD
Media tray
Power Blade
With BCH and JS23/JS43, IBM i can use:
Fibre Channel storage (MSIM, FC module and CFFh adapter required; or FC module and CIOv adapter required)
Redundant FC adapters can be configured (CFFh and CIOv)
SAS storage (SAS module and CIOv adapter required)
SAS tape (SAS module and CIOv adapter required)
USB DVD in BladeCenter
Physical I/O resources are attached to VIOS, assigned to IBM i in IVM
Storage LUNs (physical volumes) assigned directly to IBM i; storage pools in VIOS not used
31
© 2009 IBM Corporation
32. Storage, Tape and DVD for i on JS12/JS22 in BCS
SAS drives in BCS Non-RAID SAS VIOS Host IBM i Client
module in I/O Bay 3/4
hdiskX LUNs DDxx
BladeCenter midplane
Virtual SCSI
connection
RAID SAS module SAS
TS2240 in I/O Bay 3 & 4 CFFv
DS3200 Virtual SCSI
connection
USB
OPTxx
/dev/cd0 DVD
DVD
Media tray
Power Blade
With BCS and JS12/JS22, IBM i can use:
SAS storage (SAS module and CFFv adapter required)
SAS tape (SAS module and CFFv adapter required)
USB DVD
Drives in BCS, TS2240, DS3200 supported with Non-RAID SAS Switch Module (NSSM)
Only drives in BCS and TS2240 supported with RAID SAS Switch Module (RSSM)
Physical I/O resources are attached to VIOS, assigned to IBM i in IVM
Storage LUNs (physical volumes) assigned directly to IBM i; storage pools in VIOS not used
32
© 2009 IBM Corporation
33. Storage, Tape and DVD for i on JS23/JS43 in BCS
SAS drives in BCS Non-RAID SAS VIOS Host IBM i Client
module in I/O Bay 3/4
hdiskX LUNs DDxx
BladeCenter midplane
Virtual SCSI
connection
RAID SAS module SAS
TS2240 in I/O Bay 3 & 4 CIOv
DS3200
Virtual SCSI
connection
USB
OPTxx
/dev/cd0 DVD
DVD
Media tray
Power Blade
With BCS and JS23/JS43, IBM i can use:
SAS storage (SAS module and CIOv adapter required)
SAS tape (SAS module and CIOv adapter required)
USB DVD
Drives in BCS, TS2240, DS3200 supported with Non-RAID SAS Switch Module (NSSM)
Only drives in BCS and TS2240 supported with RAID SAS Switch Module (RSSM)
Physical I/O resources are attached to VIOS, assigned to IBM i in IVM
Storage LUNs (physical volumes) assigned directly to IBM i; storage pools in VIOS not used
33
© 2009 IBM Corporation
34. Storage and Tape Support 2Q 2009
• Storage support
– BladeCenter H and JS12/JS22/JS23/JS43:
• SAS – DS3200
• Fibre Channel – DS3400, DS4700, DS4800, DS8100, DS8300, SVC
– Multiple storage subsystems supported with SVC
• IBM is investigating DS5100, DS5300 and XIV support for Power blades
– BladeCenter S and JS12/JS22/JS23/JS43:
• SAS – BCS drives with NSSM and RSSM; DS3200 only with NSSM
• Tape support
– BladeCenter H and BladeCenter S:
• TS2240 LTO-4 SAS – supported for virtual tape and for VIOS backups
• TS2230 LTO-3 SAS – not supported for virtual tape, only for VIOS backups
– IBM is investigating Fibre Channel tape library support for 4Q 2009
34
© 2009 IBM Corporation
35. Configuring Storage for IBM i on Blade
• Step 1: Perform sizing
– Use Disk Magic, where applicable
– Use the PCRM, Ch. 14.5 – http://www.ibm.com/systems/i/advantages/perfmgmt/resource.html
– Number of physical drives is still most important
– VIOS itself does not add significant disk I/O overhead
– For production workloads, keep each i partition on a separate RAID array
• Step 2: Use appropriate storage UI and Redbook for your environment to create
LUNs for IBM i and attach to VIOS (or use TPC or SSPC where applicable)
Storage Configuration DS Storage Manager for DS8000 Storage Manager SVC Console for
Manager for NSSM and DS3200, DS3400, DS4700, for DS8100 and DS8300 SVC
RSSM DS4800
35
© 2009 IBM Corporation
36. Configuring Storage for IBM i on Blade, Cont.
• Step 3: Assign LUNs or physical drives in BCS to IBM i
– ‘cfgdev’ in VIOS CLI necessary to detect new physical volumes if VIOS is running
– Virtualize whole LUNs/drives (“physical volumes”) to IBM i
– Do not use storage pools in VIOS
36
© 2009 IBM Corporation
37. Configuring Storage with the RSSM
• Step 1: download SCM – http://www-947.ibm.com/systems/support/supportsite.wss/docdisplay?lndocid=MIGR-5078617&brandind=5000016
• Step 2: install SCM and add RSSM in Bay 3 (get IP address from AMM)
• Step 3: use SCM to create RAID arrays (storage pools) and volumes,
and to assign volumes to blades
• See Readme for details
37
© 2009 IBM Corporation
38. IBM i Support for Virtual Tape
• Virtual tape support enables IBM i partitions to directly
backup to PowerVM VIOS attached tape drive saving
hardware costs and management time
• Simplifies backup and restore processing with
BladeCenter implementations
– IBM i 6.1 partitions on BladeCenter JS12, JS22, JS23, JS43
– Supports IBM i save/restore commands & BRMS
– Supports BladeCenter S and H implementations
• Simplifies migration to blades from tower/rack servers
– LTO-4 drive can read backup tapes from LTO-2, 3, 4 drives
• Supports IBM Systems Storage SAS LTO-4 Drive
– TS2240 SAS ONLY for BladeCenter
– IBM is investigating Fibre Channel tape library support for 4Q 2009
• Requirements
– VIOS 2.1.1, eFW 3.4.2, IBM i 6.1 PTFs
38
© 2009 IBM Corporation
39. Virtual Tape Hardware and Virtualization
VIOS Host IBM i Client
SAS-attached LTO4 SAS I/O module
BladeCenter midplane
tape drive (TS2240)
CFFv
SAS
OR
OR
/dev/rmt0 TAP01
RAID SAS I/O CIOv Separate
SAS 3580 004
module Virtual SCSI
connection
Power Blade
• TS2240 LTO4 SAS tape drive attached to SAS switch in BladeCenter:
– NSSM or RSSM in BCS
– NSSM in BCH
• VIOS virtualizes tape drive to IBM i directly
• Tape drive assigned to IBM i in IVM
• Tape drive available in IBM i as TAPxx, type 3580 model 004
39
© 2009 IBM Corporation
40. Assigning Virtual Tape to IBM i
• No action required in IBM i to make tape drive available
– If QAUTOCFG is on (default)
40
© 2009 IBM Corporation
41. Migrating IBM i to Blade
• Virtual tape makes migration to blade similar to migration to tower/rack
server:
– On existing system, go save option 21 on LTO-2, LTO-3 or LTO-4 media
– On blade, use virtual tape to perform D-mode IPL and complete restore
– Existing system does not have to be at IBM i 6.1
• Previous-to-current migration also possible
• IBM i partition saved on blade can be restored on tower/rack server
– IBM i can save to LTO-3 and LTO-4 media on blade
• For existing servers that do not have access to LTO tape drive, there are
two options:
– Save on different media, convert to LTO as a service, restore from LTO
– Use Migration Assistant method
41
© 2009 IBM Corporation
42. Multiple Virtual SCSI Adapters for IBM i
• Since VIOS 2.1 in November 2008, IBM i is no longer limited to 1 VSCSI
connection to VIOS and 16 disk + 16 optical devices
• What IVM will do:
– Create 1 VSCSI server adapter in VIOS for each IBM i partition created
– Create 1 VSCSI client adapter in IBM i and correctly map to Server adapter
– Map any disk and optical devices you assign to IBM i to the first VSCSI server
adapter in VIOS
– Create a new VSCSI server-client adapter pair only when you assign a tape
device to IBM i
– Create another VSCSI server-client adapter pair when you assign another tape
device
• What IVM will not do:
– Create a new VSCSI server-client adapter pair if you assign more than 16 disk
devices to IBM i
42
© 2009 IBM Corporation
43. Multiple Virtual SCSI Adapters for IBM i, Cont.
• Scenario I: you have <=16 disk devices and you want to add virtual tape
– Action required in VIOS:
• In IVM, click on tape drive, assign to IBM i partition
– Separate VSCSI server-client adapter pair created automatically
• Scenario II: you have 16 disk devices and you want to add more disk and
virtual tape
– Actions required in VIOS:
• In VIOS CLI, create new VSCSI client adapter in IBM i
– VSCSI server adapter in VIOS created automatically
• In VIOS CLI, map new disk devices to new VSCSI server adapter using ‘mkvdev’
• In IVM, click on tape drive, assign to IBM i partition
• For details and instructions, see IBM i on Blade Read-me First:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html
43
© 2009 IBM Corporation
44. Networking on Power Blade
VIOS Host i Client
Ethernet
I/O module
Embedded Ethernet
BladeCenter midplane
ports on blade
10.10.10.35
IVE
Local PC for: (HEA) CMN01
10.10.10.20
AMM browser Virtual LAN LAN console
IVM browser connection
LAN console
LAN IVE
Virtual 10.10.10.37
Ethernet CMN02
(HEA)
bridge
IVE
Production
interface
10.10.10.38
10.10.10.5 Power Blade
VIOS is accessed from local PC via embedded Ethernet ports on blade
(IVE/HEA)
For both IVM browser and VIOS command line
Same PC can be used to connect to AMM and for LAN console for i5/OS
For i connectivity, IVE/HEA port is bridged to Virtual LAN
44
© 2009 IBM Corporation
45. LAN Console for i on Power Blade
Required for i on Power
blade
Uses System i Access
software on PC (can use
same PC for IVM connection)
Full console functionality
Uses existing LAN console
capability
45
© 2009 IBM Corporation
46. PowerVM Active Memory Sharing
• PowerVM Active Memory Sharing is an advanced memory
virtualization technology which intelligently flows memory from Around the World
15
one partition to another for increased utilization and flexibility
of memory usage
Memory Usage (GB)
10
Asia
Americas
• Memory virtualization enhancement for Power 5
Europe
Systems
0
– Partitions share a pool of memory
Time
– Memory dynamically allocated based on partition’s workload demands
Day and Night
15
• Extends Power Systems Virtualization Leadership
Memory Usage (GB)
– Capabilities not provided by Sun and HP virtualization offerings 10
Night
Day
5
• Designed for partitions with variable memory
requirements 0
Time
– Workloads that peak at different times across the partitions
– Active/inactive environments
– Test and Development environments Infrequent Use
15
#10
– Low average memory requirements
#9
Memory Usage (GB)
#8
10
#7
• Available with PowerVM Enterprise Edition #6
#5
– Supports AIX 6.1, i 6.1, and SUSE Linux Enterprise Server 11 5
#4
#3
– Partitions must use VIOS and shared processors #2
0
Time #1
– POWER6 processor-based systems
46
© 2009 IBM Corporation
48. Service Voucher for IBM i on Power Blade
• Let IBM Systems Lab Services and Training help you install i on blade!
• 1 service voucher for each Power blade AND IBM i license purchased
• http://www.ibm.com/systems/i/hardware/editions/services.html
48
© 2009 IBM Corporation
49. Further Reading
• IBM i on Blade Read-me First:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html
• IBM i on Blade Supported Environments:
http://www.ibm.com/systems/power/hardware/blades/ibmi.html
• IBM i on Blade Performance Information:
http://www.ibm.com/systems/i/advantages/perfmgmt/resource.html
• Service vouchers:
http://www.ibm.com/systems/i/hardware/editions/services.html
• IBM i on Blade Training:
http://www.ibm.com/systems/i/support/itc/educ.html
49
© 2009 IBM Corporation
50. Trademarks and Disclaimers
8 IBM Corporation 1994-2007. All rights reserved.
References in this document to IBM products or services do not imply that IBM intends to make them available in every country.
Trademarks of International Business Machines Corporation in the United States, other countries, or both can be found on the World Wide Web at
http://www.ibm.com/legal/copytrade.shtml.
Intel, Intel logo, Intel Inside, Intel Inside logo, Intel Centrino, Intel Centrino logo, Celeron, Intel Xeon, Intel SpeedStep, Itanium, and Pentium are trademarks or registered
trademarks of Intel Corporation or its subsidiaries in the United States and other countries.
Linux is a registered trademark of Linus Torvalds in the United States, other countries, or both.
Microsoft, Windows, Windows NT, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both.
IT Infrastructure Library is a registered trademark of the Central Computer and Telecommunications Agency which is now part of the Office of Government Commerce.
ITIL is a registered trademark, and a registered community trademark of the Office of Government Commerce, and is registered in the U.S. Patent and Trademark Office.
UNIX is a registered trademark of The Open Group in the United States and other countries.
Java and all Java-based trademarks are trademarks of Sun Microsystems, Inc. in the United States, other countries, or both.
Other company, product, or service names may be trademarks or service marks of others.
Information is provided "AS IS" without warranty of any kind.
The customer examples described are presented as illustrations of how those customers have used IBM products and the results they may have achieved. Actual
environmental costs and performance characteristics may vary by customer.
Information concerning non-IBM products was obtained from a supplier of these products, published announcement material, or other publicly available sources and does
not constitute an endorsement of such products by IBM. Sources for non-IBM list prices and performance numbers are taken from publicly available information,
including vendor announcements and vendor worldwide homepages. IBM has not tested these products and cannot confirm the accuracy of performance, capability, or
any other claims related to non-IBM products. Questions on the capability of non-IBM products should be addressed to the supplier of those products.
All statements regarding IBM future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only.
Some information addresses anticipated future capabilities. Such information is not intended as a definitive statement of a commitment to specific levels of performance,
function or delivery schedules with respect to any future products. Such commitments are only made in IBM product announcements. The information is presented here
to communicate IBM's current investment and development activities as a good faith effort to help with our customers' future planning.
Performance is based on measurements and projections using standard IBM benchmarks in a controlled environment. The actual throughput or performance that any
user will experience will vary depending upon considerations such as the amount of multiprogramming in the user's job stream, the I/O configuration, the storage
configuration, and the workload processed. Therefore, no assurance can be given that an individual user will achieve throughput or performance improvements
equivalent to the ratios stated here.
Prices are suggested U.S. list prices and are subject to change without notice. Starting price may not include a hard drive, operating system or other features. Contact
your IBM representative or Business Partner for the most current pricing in your geography.
Photographs shown may be engineering prototypes. Changes may be incorporated in production models.
50
© 2009 IBM Corporation
51. Special notices
This document was developed for IBM offerings in the United States as of the date of publication. IBM may not make these offerings available in
other countries, and the information is subject to change without notice. Consult your local IBM business contact for information on the IBM
offerings available in your area.
Information in this document concerning non-IBM products was obtained from the suppliers of these products or other public sources. Questions
on the capabilities of non-IBM products should be addressed to the suppliers of those products.
IBM may have patents or pending patent applications covering subject matter in this document. The furnishing of this document does not give
you any license to these patents. Send license inquires, in writing, to IBM Director of Licensing, IBM Corporation, New Castle Drive, Armonk, NY
10504-1785 USA.
All statements regarding IBM future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives
only.
The information contained in this document has not been submitted to any formal IBM test and is provided "AS IS" with no warranties or
guarantees either expressed or implied.
All examples cited or described in this document are presented as illustrations of the manner in which some IBM products can be used and the
results that may be achieved. Actual environmental costs and performance characteristics will vary depending on individual client configurations
and conditions.
IBM Global Financing offerings are provided through IBM Credit Corporation in the United States and other IBM subsidiaries and divisions
worldwide to qualified commercial and government clients. Rates are based on a client's credit rating, financing terms, offering type, equipment
type and options, and may vary by country. Other restrictions may apply. Rates and offerings are subject to change, extension or withdrawal
without notice.
IBM is not responsible for printing errors in this document that result in pricing or information inaccuracies.
All prices shown are IBM's United States suggested list prices and are subject to change without notice; reseller prices may vary.
IBM hardware products are manufactured from new parts, or new and serviceable used parts. Regardless, our warranty terms apply.
Any performance data contained in this document was determined in a controlled environment. Actual results may vary significantly and are
dependent on many factors including system hardware configuration and software design and configuration. Some measurements quoted in this
document may have been made on development-level systems. There is no guarantee these measurements will be the same on generally-
available systems. Some measurements quoted in this document may have been estimated through extrapolation. Users of this document
should verify the applicable data for their specific environment.
Revised September 26, 2006
51
© 2009 IBM Corporation
52. Special notices (cont.)
IBM, the IBM logo, ibm.com AIX, AIX (logo), AIX 6 (logo), AS/400, BladeCenter, Blue Gene, ClusterProven, DB2, ESCON, IBM i, IBM i (logo), IBM Business Partner
(logo), IntelliStation, LoadLeveler, Lotus, Lotus Notes, Notes, Operating System/400, OS/400, PartnerLink, PartnerWorld, PowerPC, pSeries, Rational, RISC
System/6000, RS/6000, THINK, Tivoli, Tivoli (logo), Tivoli Management Environment, WebSphere, xSeries, z/OS, zSeries, AIX 5L, Chiphopper, Chipkill, Cloudscape, DB2
Universal Database, DS4000, DS6000, DS8000, EnergyScale, Enterprise Workload Manager, General Purpose File System, , GPFS, HACMP, HACMP/6000, HASM, IBM
Systems Director Active Energy Manager, iSeries, Micro-Partitioning, POWER, PowerExecutive, PowerVM, PowerVM (logo), PowerHA, Power Architecture, Power
Everywhere, Power Family, POWER Hypervisor, Power Systems, Power Systems (logo), Power Systems Software, Power Systems Software (logo), POWER2,
POWER3, POWER4, POWER4+, POWER5, POWER5+, POWER6, System i, System p, System p5, System Storage, System z, Tivoli Enterprise, TME 10, Workload
Partitions Manager and X-Architecture are trademarks or registered trademarks of International Business Machines Corporation in the United States, other countries, or
both. If these and other IBM trademarked terms are marked on their first occurrence in this information with a trademark symbol (® or ™), these symbols indicate U.S.
registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law trademarks in
other countries. A current list of IBM trademarks is available on the Web at "Copyright and trademark information" at www.ibm.com/legal/copytrade.shtml
The Power Architecture and Power.org wordmarks and the Power and Power.org logos and related marks are trademarks and service marks licensed by Power.org.
UNIX is a registered trademark of The Open Group in the United States, other countries or both.
Linux is a registered trademark of Linus Torvalds in the United States, other countries or both.
Microsoft, Windows and the Windows logo are registered trademarks of Microsoft Corporation in the United States, other countries or both.
Intel, Itanium, Pentium are registered trademarks and Xeon is a trademark of Intel Corporation or its subsidiaries in the United States, other countries or both.
AMD Opteron is a trademark of Advanced Micro Devices, Inc.
Java and all Java-based trademarks and logos are trademarks of Sun Microsystems, Inc. in the United States, other countries or both.
TPC-C and TPC-H are trademarks of the Transaction Performance Processing Council (TPPC).
SPECint, SPECfp, SPECjbb, SPECweb, SPECjAppServer, SPEC OMP, SPECviewperf, SPECapc, SPEChpc, SPECjvm, SPECmail, SPECimap and SPECsfs are
trademarks of the Standard Performance Evaluation Corp (SPEC).
NetBench is a registered trademark of Ziff Davis Media in the United States, other countries or both.
AltiVec is a trademark of Freescale Semiconductor, Inc.
Cell Broadband Engine is a trademark of Sony Computer Entertainment Inc.
InfiniBand, InfiniBand Trade Association and the InfiniBand design marks are trademarks and/or service marks of the InfiniBand Trade Association.
Other company, product and service names may be trademarks or service marks of others.
Revised April 24, 2008
52
© 2009 IBM Corporation
53. Notes on performance estimates
rPerf for AIX
rPerf (Relative Performance) is an estimate of commercial processing performance relative to other IBM UNIX systems. It is derived
from an IBM analytical model which uses characteristics from IBM internal workloads, TPC and SPEC benchmarks. The rPerf
model is not intended to represent any specific public benchmark results and should not be reasonably used in that way. The
model simulates some of the system operations such as CPU, cache and memory. However, the model does not simulate disk or
network I/O operations.
• rPerf estimates are calculated based on systems with the latest levels of AIX and other pertinent software at the time of system
announcement. Actual performance will vary based on application and configuration specifics. The IBM eServer pSeries 640 is
the baseline reference system and has a value of 1.0. Although rPerf may be used to approximate relative IBM UNIX commercial
processing performance, actual system performance may vary and is dependent upon many factors including system hardware
configuration and software design and configuration. Note that the rPerf methodology used for the POWER6 systems is
identical to that used for the POWER5 systems. Variations in incremental system performance may be observed in commercial
workloads due to changes in the underlying system architecture.
All performance estimates are provided "AS IS" and no warranties or guarantees are expressed or implied by IBM. Buyers should
consult other sources of information, including system benchmarks, and application sizing guides to evaluate the performance
of a system they are considering buying. For additional information about rPerf, contact your local IBM office or IBM authorized
reseller.
========================================================================
CPW for IBM i
Commercial Processing Workload (CPW) is a relative measure of performance of processors running the IBM i operating system.
Performance in customer environments may vary. The value is based on maximum configurations. More performance
information is available in the Performance Capabilities Reference at:
www.ibm.com/systems/i/solutions/perfmgmt/resource.html
Revised April 2, 2007
53
© 2009 IBM Corporation