Professional Documents
Culture Documents
^Eserver pSeries
Section Objectives
On completion of this unit you should be able to:
Describe the relationship between technology and
solutions.
List key IBM technologies that are part of the POWER5
products.
Be able to describe the functional benefits that these
technologies provide.
Be able to discuss the appropriate use of these
technologies.
^Eserver pSeries
^Eserver pSeries
^Eserver pSeries
Eserver pSeries
POWER5 Technology
^Eserver pSeries
POWER5 Core
^Eserver pSeries
POWER5
SMT Core
1.9 MB L2 Cache
Mem Ctrl
SMT Core
L3 Dir
servers
POWER4 systems
GX+
^Eserver pSeries
Larger L2 cache
1.9 MB, 10-way set associative
SMT Core
1.9 MB L2 Cache
Mem Ctrl
SMT Core
L3 Dir
^Eserver pSeries
Processor
Processor
Processor
Processor
L2
L2
Cache
Fabric
Fabric
controller
controller
L3
Cache
Processor
Processor
Processor
Processor
Processor
Processor
L2
L2
Cache
Cache
L2
L2
Cache
Cache
Fabric
Fabric
controller
controller
Fabric
Fabric
controller
controller
Fabric
Fabric
controller
controller
L3
Cache
L3
Cache
Memory
Memory
controller
controller
Memory
Memory
controller
controller
Memory
controller
Memory
controller
Memory
Memory
Memory
Memory
Faster
access to
memory
L3 Dir
Dir
L3
L2
L2
Cache
Cache
Processor
Processor
L3 Dir
Dir
L3
Processor
Processor
L3
Cache
Larger
SMPs
64-way
Number of
chips cut
in half
^Eserver pSeries
^Eserver pSeries
POWER4 pipeline
Out-of-order processing
Branch redirects
Instruction Fetch
IF
IC
BP
D0
D1
D2
D3
Xfer
GD
Branch
pipeline
Load/store
pipeline
MP
ISS
RF
EX
MP
ISS
RF
EA
MP
ISS
RF
EX
MP
ISS
RF
F6
F6
F6
F6
F6
F6
DC
Fmt
Fixed-point
pipeline
WB Xfer
WB Xfer
CP
WB Xfer
WB Xfer
Floatingpoint pipeline
POWER4 instruction pipeline (IF = instruction fetch, IC = instruction cache, BP = branch predict, D0
= decode stage 0, Xfer = transfer, GD = group dispatch, MP = mapping, ISS = instruction issue, RF =
register file read, EX = execute, EA = compute address, DC = data caches, F6 = six-cycle floatingpoint execution pipe, Fmt = data format, WB = write back, and CP = group commit)
POWER5 pipeline
^Eserver pSeries
Multi-threading evolution
Memory
Instruction streams
i-Cache
FX0
FX1
LS0
LS1
FP0
FP1
BFX
CRL
Processor Cycles
^Eserver pSeries
Coarse-grained multi-threading
Two instruction streams, one thread at any instance
Swap
Swap
FX0
FX1
LS0
LS1
FP0
FP1
BFX
CRL
i-Cache
Instruction streams
Swap
Memory
Processor Cycles
^Eserver pSeries
^Eserver pSeries
Fine-grained multi-threading
Memory
Instruction streams
i-Cache
FX0
FX1
LS0
LS1
FP0
FP1
BFX
CRL
Processor Cycles
^Eserver pSeries
POWER5 pipeline
Out-of-order processing
Branch redirects
Instruction Fetch
IF
IF
IC
BP
D0
D1
D2
D3
Xfer
GD
Branch
pipeline
Load/store
pipeline
MP
ISS
RF
EX
MP
ISS
RF
EA
MP
ISS
RF
EX
MP
ISS
RF
F6
F6
F6
F6
F6
F6
DC
Fmt
Fixed-point
pipeline
WB Xfer
WB Xfer
CP
CP
WB Xfer
WB Xfer
Floatingpoint pipeline
POWER5 instruction pipeline (IF = instruction fetch, IC = instruction cache, BP = branch predict, D0
= decode stage 0, Xfer = transfer, GD = group dispatch, MP = mapping, ISS = instruction issue, RF =
register file read, EX = execute, EA = compute address, DC = data caches, F6 = six-cycle floatingpoint execution pipe, Fmt = data format, WB = write back, and CP = group commit)
POWER4 pipeline
^Eserver pSeries
i-Cache
Instruction streams
Processor Cycles
^Eserver pSeries
(Cont.)
^Eserver pSeries
^Eserver pSeries
2
2
Single-threaded operation
1
1
1
1
Power
Save
Mode
1
0
0
0
0,7 2,7 4,7 6,7 7,7 7,6 7,4 7,2 7,0 1,1
Thread 1 IPC
Hardware thread priorities
^Eserver pSeries
Single-threaded operation
Advantageous for execution unit
limited applications
Floating or fixed point intensive
workloads
Dormant
Software
Hardware
or Software
Active
Software
Null
Eserver pSeries
Micro-Partitioning
^Eserver pSeries
Micro-Partitioning overview
Mainframe inspired technology
Virtualized resources shared by multiple partitions
Benefits
Finer grained resource allocation
More partitions (Up to 254)
Higher resource utilization
POWER Hypervisor
Virtual processors
Fractional processor capacity partitions
Operating system optimized for Micro-Partitioning exploitation
Virtual I/O
^Eserver pSeries
Processor terminology
Shared processor
partition
SMT Off
Shared processor
partition
SMT On
Dedicated
processor partition
SMT Off
Logical (SMT)
Virtual
Shared
Dedicated
Entitled capacity
Inactive (CUoD)
Deconfigured
Installed physical
processors
^Eserver pSeries
CPU 0
CPU 1
CPU 3
CPU 4
LPAR 1
LPAR 2
LPAR 3
LPAR 4
LPAR 5
LPAR 6
Dedicated memory
Minimum of 128 MB and 16 MB increments
^Eserver pSeries
^Eserver pSeries
Minimum requirement
0.1 processing units
Processing capacity
1 physical processor
1.0 processing units
^Eserver pSeries
Uncapped partition
Is allowed to exceed its entitlement
Capacity weight
Used for prioritizing uncapped partitions
Value 0-255
Value of 0 referred to as a soft cap
^Eserver pSeries
Partition 2 activated
Min = 1.0, max = 2.0, desired = 1.0
Does not start
Partition 3 activated
Min = 0.1, max = 1.0, desired = 0.8
Starts with 0.5 allocated processing units
^Eserver pSeries
^Eserver pSeries
5
0
1
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Elapsed time
16 virtual processors.
Uncapped.
Can use all available resource.
The workload requires 26 minutes to complete.
^Eserver pSeries
10
0
1
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Elapsed time
12 virtual processors.
Even though the partition is uncapped, it can only use 12
processing units.
^Eserver pSeries
Capped
Capped (16PPs/12VPs/9.5E)
15
10
5
0
1
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Elapses time
^Eserver pSeries
^Eserver pSeries
Dynamic LPAR
Standard on all new systems
Part#1
Production
AIX
5L
Part#2
Part#3
Part#4
Legacy
Test/
File/
Print
Move Apps
resourcesDev
between live
partitions
AIX
AIX
5L
Linux
5L
Hypervisor
HMC
Eserver pSeries
Firmware
POWER Hypervisor
^Eserver pSeries
^Eserver pSeries
H-Call Interface
iSeries
Location codes
Nucleus (SLIC)
Bus recovery
I/O configuration
FSP
SCSI IOA
Virtual Ethernet
LAN IOA
HSC
HMC
Dump
255 partitions
Partition on demand
Virtual I/O
VLAN IOA
VLAN
^Eserver pSeries
CPU 3
SMT
SMTCore
Core
MemCtrl
Ctrl
Mem
SMT
SMTCore
Core
1.9
1.9MB
MBL2
L2Cache
Cache
SMT
SMTCore
Core
MemCtrl
Ctrl
Mem
L3Dir
Dir
L3
SMT
SMTCore
Core
L3Dir
Dir
L3
SMT
SMTCore
Core
MemCtrl
Ctrl
Mem
Dynamic LPAR
SMT
SMTCore
Core
SMT
SMTCore
Core
1.9
1.9MB
MBL2
L2Cache
Cache
SMT
SMTCore
Core
MemCtrl
Ctrl
Mem
CPU 1
CPU 2
L3Dir
Dir
L3
L3Dir
Dir
Enhanced distributed
switch
L3
Virtual I/O
Planned
Actual
Disk
LAN
^Eserver pSeries
^Eserver pSeries
CPU 0
SMT
SMTCore
Core
1.9
1.9MB
MBL2
L2Cache
Cache
1.9
1.9MB
MBL2
L2Cache
Cache
MemCtrl
Ctrl
Mem
SMT
SMTCore
Core
SMT
SMTCore
Core
MemCtrl
Ctrl
Mem
1.9
1.9MB
MBL2
L2Cache
Cache
MemCtrl
Ctrl
Mem
MemCtrl
Ctrl
Mem
1.9
1.9MB
MBL2
L2Cache
Cache
SMT
SMTCore
Core
L3Dir
Dir
L3
SMT
SMTCore
Core
L3Dir
Dir
L3
SMT
SMTCore
Core
CPU 1
L3Dir
Dir
L3
SMT
SMTCore
Core
SMT
SMTCore
Core
L3Dir
Dir
L3
POWER
Hypervisors
processor
dispatch
CPU 2
CPU 3
^Eserver pSeries
^Eserver pSeries
POWER
Hypervisors
processor
dispatch
Affinity scheduling
SMT
SMTCore
Core
1.9
1.9MB
MBL2
L2Cache
Cache
MemCtrl
Ctrl
Mem
MemCtrl
Ctrl
Mem
1.9
1.9MB
MBL2
L2Cache
Cache
SMT
SMTCore
Core
L3Dir
Dir
L3
SMT
SMTCore
Core
L3Dir
Dir
L3
MemCtrl
Ctrl
Mem
1.9
1.9MB
MBL2
L2Cache
Cache
SMT
SMTCore
Core
SMT
SMTCore
Core
L3Dir
Dir
L3
MemCtrl
Ctrl
Mem
1.9
1.9MB
MBL2
L2Cache
Cache
SMT
SMTCore
Core
SMT
SMTCore
Core
L3Dir
Dir
L3
SMT
SMTCore
Core
Home node
CPU 0
CPU 1
CPU 2
CPU 3
^Eserver pSeries
Affinity scheduling
When dispatching a VP, the POWER Hypervisor attempts to
preserve affinity by using:
^Eserver pSeries
^Eserver pSeries
Example
Physical LPAR 1 LPAR 3 LPAR 1
processor 0 VP 1
VP 2
VP 1
Physical LPAR 2
processor 1 VP 0
0
LPAR 1
VP 0
2
LPAR 3
VP 0
IDLE
LPAR 1
VP 0
LPAR 1
VP 1
IDLE
LPAR 3 LPAR 2
VP 1
VP 0
10 11 12 13 14 15 16 17 18 19 20
LPAR1
Capacity entitlement = 0.8 processing units; virtual processors = 2 (capped)
LPAR2
Capacity entitlement = 0.2 processing units; virtual processors = 1 (capped)
LPAR3
Capacity entitlement = 0.6 processing units; virtual processors = 3 (capped)
^Eserver pSeries
LAN
^Eserver pSeries
SMT
Thread priorities compound the variable speed rate.
Twice as many logical CPUs.
Eserver pSeries
^Eserver pSeries
Virtual I/O Server Version 1.1 is addressed for selected configurations, which include specific models of
EMC, HDS, and STK disk subsystems, attached using Fiber Channel
^Eserver pSeries
^Eserver pSeries
Virtual SCSI
Allows sharing of storage devices
Vital for shared processor partitions
Overcomes potential limit of adapter slots due to MicroPartitioning
Allows the creation of logical partitions without the need for
additional physical resources
^Eserver pSeries
Virtual I/O
Client
Client
Server partition partition partition
AIX
Linux
LVM
Logical
volume 1
Logical
volume 2
hdisk
hdisk
VSCSI server
adapter
VSCSI server
adapter
VSCI client
adapter
VSCI client
adapter
POWER Hypervisor
Physical adapter
Physical disk
(SCSI, FC)
^Eserver pSeries
Virtual devices
Are defined as LVs in the I/O server
partition
Client partition
Virtual
disk
hdisk
LVM
VSCI client
adapter
POWER Hypervisor
VSCSI server
adapter
LVM
LV
hdisk
Virtual I/O Server partition
^Eserver pSeries
Protocol (SRP).
SCSI initiators and targets have the
ability to directly transfer information
between their respective address
spaces.
Physical
adapter device
driver
VSCI device
driver (target)
Device
Mapping
VSCI device
driver (initiator)
Data Buffer
POWER Hypervisor
Physical adapter
^Eserver pSeries
^Eserver pSeries
Performance considerations
Twice as many processor cycles to do VSCSI as a locally attached
disk I/O (evenly distributed on the client partition and virtual I/O
server)
The path of each virtual I/O request involves several sources of
overhead that are not present in a non-virtual I/O request.
For a virtual disk backed by the LVM, there is also the performance
impact of going through the LVM and disk device drivers twice.
^Eserver pSeries
Limitations
Hosting partition must be available before hosted
partition boot.
^Eserver pSeries
Implementation guideline
Partitions with high performance and disk I/O
requirements are not recommended for
implementing VSCSI.
^Eserver pSeries
LVM mirroring
This configuration
Virtual I/O
Server
partition
Client
partition
Virtual I/O
Server
partition
LVM
LVM
LVM
VSCSI server
adapter
VSCSI
client
adapter
VSCSI
client
adapter
VSCSI server
adapter
POWER Hypervisor
Physical SCSI
adapter
Physical SCSI
adapter
Physical disk
(SCSI)
Physical disk
(SCSI)
^Eserver pSeries
Multipath I/O
This configuration protects
Virtual I/O
Server
partition
Client
partition
Virtual I/O
Server
partition
LVM
(hdisk)
LVM
LVM
(hdisk)
VSCSI server
adapter
VSCSI
client
adapter
VSCSI
client
adapter
VSCSI server
adapter
POWER Hypervisor
Physical FC adapter
Physical FC adapter
SAN Switch
^Eserver pSeries
Node A-1
VLANs provides:
Node A-2
Switch A
Switch B
Switch C
^Eserver pSeries
Virtual Ethernet
Enables inter-partition communication.
In-memory point to point connections
^Eserver pSeries
AIX
AIX
Linux
partition partition partition
Virtual
Ethernet
adapter
Virtual
Ethernet
adapter
Virtual
Ethernet
adapter
POWER Hypervisor
^Eserver pSeries
^Eserver pSeries
How it works
Virtual Ethernet adapter
IEEE VLAN
header?
N
Insert VLAN header
Y
Port allowed?
N
Dest. MAC in
table?
Configured associated switch
port
Y
N
Trunk adapter
defined?
Match for
VLAN Nr. in
table?
N
Deliver
Pass to Trunk
adapter
Drop packet
^Eserver pSeries
Performance considerations
Throughput per 0.1 entitlement
Throughput/0.1
entitlement
[Mb/s]
1000
800
600
400
200
0
0.1
0.3
0.5
0.8
65394
9000
1500 MTU
size
CPU entitlements
Throughput, TCP_STREAM
Throughput
[Mb/s]
10000
8000
6000
VLAN
4000
Gb Ethernet
2000
0
MTU
Simpl./Dupl.
1
1500
S
1500 9000
D
S
^Eserver pSeries
Limitations
Virtual Ethernet can be used in both shared and
dedicated processor partitions provided with the
appropriate OS levels.
^Eserver pSeries
Implementation guideline
Know your environment and the network traffic.
Choose a high MTU size, as it makes sense for the
network traffic in the Virtual LAN.
^Eserver pSeries
Routing
The partition that routes the traffic to the external work does not necessarily have to be
the virtual I/O server.
AIX partition
1.1.1.100
3.1.1.1
AIX
Linux
partition partition
3.1.1.10
AIX partition
3.1.1.10
2.1.1.100
4.1.1.1
AIX
Linux
partition partition
4.1.1.10
4.1.1.11
POWER Hypervisor
POWER Hypervisor
Physical adapter
Physical adapter
IP subnet 1.1.1.X
IP Router
1.1.1.1
2.1.1.1
IP subnet 2.1.1.X
AIX
Server
Linux
Server
1.1.1.10
2.1.1.10
^Eserver pSeries
^Eserver pSeries
One SEA can be shared by multiple VLANs and multiple subnets can
connect using a single adapter on the Virtual I/O Server.
^Eserver pSeries
^Eserver pSeries
Performance considerations
Virtual I/O-Server
performance
Adapters stream data at
media speed if the Virtual
I/O server has enough
capacity entitlement.
CPU utilization per Gigabit
of throughput is higher with
a Shared Ethernet adapter.
2000
1500
1000
500
0
MTU
Simplex/Duplex
CPU
Utilisation
[%cpu/Gb]
1500
simplex
1500
duplex
9000
simplex
9000
duplex
100
80
60
40
20
0
MTU
Simplex/Duplex
1500
simplex
1500
duplex
9000
simplex
9000
duplex
^Eserver pSeries
Limitations
System processors are used for all communication functions,
leading to a significant amount of system processor load.
^Eserver pSeries
Implementation guideline
Know your environment and the network traffic.
Use a dedicated network adapter if you expect heavy
network traffic between Virtual Ethernet and local
networks.
Choose 9000 for MTU size, if this makes sense for your
network traffic.
With MTU size 1500, you need about 1 CPU per gigabit
Ethernet adapter streaming at media speed.
^Eserver pSeries
AIX
Linux
partition partition
ent0
VLAN 2 VLAN 1
VLAN 1
10.1.1.11
VLAN 2
10.1.2.11
POWER Hypervisor
Physical adapter
VLAN 1
VLAN 2
AIX
Server
Linux
Server
10.1.1.14
10.1.2.15
^Eserver pSeries
Maximizing throughput
Using several Shared Ethernet
Adapters
ent0 ent1
VLAN VLAN
2
1
VLAN 1
10.1.1.11
VLAN 2
10.1.2.11
More queues
More performance
AIX
Linux
partition partition
POWER Hypervisor
Physical adapter Physical adapter
VLAN 1
VLAN 2
AIX
Server
Linux
Server
10.1.1.14
10.1.2.15
^Eserver pSeries
Virtual I/O
Server 2
AIX partition
Multipath routing
with
dead gateway
detection
ent0
Virtual I/O
Server 2
VLAN 1
9.3.5.11
VLAN 1 VLAN 2
9.3.5.12 9.3.5.22
VLAN 2
9.3.5.21
ent0
POWER Hypervisor
Physical adapter
Physical adapter
Gateway
9.3.5.10
Gateway
9.3.5.20
External
network
^Eserver pSeries
^Eserver pSeries
^Eserver pSeries
a. True
b. False
a. True
b. False
^Eserver pSeries
^Eserver pSeries
^Eserver pSeries
^Eserver pSeries
^Eserver pSeries
^Eserver pSeries
Unit Summary
You should now be able to:
Describe the relationship between technology and
solutions.
List key IBM technologies that are part of the POWER5
products.
Be able to describe the functional benefits that these
technologies provide.
Be able to discuss the appropriate use of these
technologies.
^Eserver pSeries
Reference
You may find more information here:
IBM eServer pSeries AIX 5L Support for Micro-Partitioning
and Simultaneous Multi-threading White Paper
Introduction to Advanced POWER Virtualization on IBM
eServer p5 Servers SG24-7940
IBM eServer p5 Virtualization Performance
Considerations SG24-5768