IBM 88554RU Installation Guide - Page 25

Explicitly Parallel Instruction Computing EPIC, Level 2 cache, 256 KB

Page 25 highlights

This is new and the "closest" to the processor, and is used to store micro-operations. These are decoded executable machine instructions. It serves these to the processor at rated speed. This additional level of cache saves decode time on cache hits. Level 2 cache, 256 KB This is equivalent to L1 cache on the Pentium® III Xeon. Level 3 cache 3-6 MB This is equivalent to L2 cache on the Pentium III Xeon or the L3 cache on the Pentium Xeon MP processor. Unlike the design of the original Itanium processor, this L3 cache is now on the processor die, greatly improving performance, up to 2 times greater than that of the original Itanium. The x455 also implements a Level 4 cache as described in "IBM XceL4 Accelerator Cache" on page 12. Intel has also introduced a number of features associated with its Itanium micro-architecture. These are available in the x455, including: 400 MHz frontside bus The Pentium III Xeon processor had a 100 MHz frontside bus that equated to a burst throughput of 800 MBps. With protocols such as TCP/IP, this had been shown to be a bottleneck in high-throughput situations. The Itanium 2 processor improves on this by using a single 200 MHz clock but using both edges of each clock cycle to transmit data. This is shown in Figure 1-6. 200 MHz clock Figure 1-6 Dual-pumped frontside bus This increases the performance of the frontside bus. The end result is an effective burst throughput of 6.4 GBps (128-bit wide data path running at 400 MHz), which can have a substantial impact, especially on TCP/IP-based LAN traffic. This is opposed to the Itanium processor, which had a burst throughput of only 2.1 GBps (64-bit wide data path running at 266 MHz). Explicitly Parallel Instruction Computing (EPIC) EPIC technology, developed by Intel and HP, leads to more efficient, faster processors because it eliminates numerous processing inefficiencies in current processors and attacks the perennial data bottleneck problems by increasing parallelism, rather than simply boosting the raw "clock" speed of the processor. Chapter 1. Technical description 11

  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16
  • 17
  • 18
  • 19
  • 20
  • 21
  • 22
  • 23
  • 24
  • 25
  • 26
  • 27
  • 28
  • 29
  • 30
  • 31
  • 32
  • 33
  • 34
  • 35
  • 36
  • 37
  • 38
  • 39
  • 40
  • 41
  • 42
  • 43
  • 44
  • 45
  • 46
  • 47
  • 48
  • 49
  • 50
  • 51
  • 52
  • 53
  • 54
  • 55
  • 56
  • 57
  • 58
  • 59
  • 60
  • 61
  • 62
  • 63
  • 64
  • 65
  • 66
  • 67
  • 68
  • 69
  • 70
  • 71
  • 72
  • 73
  • 74
  • 75
  • 76
  • 77
  • 78
  • 79
  • 80
  • 81
  • 82
  • 83
  • 84
  • 85
  • 86
  • 87
  • 88
  • 89
  • 90
  • 91
  • 92
  • 93
  • 94
  • 95
  • 96
  • 97
  • 98
  • 99
  • 100
  • 101
  • 102
  • 103
  • 104
  • 105
  • 106
  • 107
  • 108
  • 109
  • 110
  • 111
  • 112
  • 113
  • 114
  • 115
  • 116
  • 117
  • 118
  • 119
  • 120
  • 121
  • 122
  • 123
  • 124
  • 125
  • 126
  • 127
  • 128
  • 129
  • 130
  • 131
  • 132
  • 133
  • 134
  • 135
  • 136
  • 137
  • 138
  • 139
  • 140
  • 141
  • 142
  • 143
  • 144
  • 145
  • 146
  • 147
  • 148
  • 149
  • 150
  • 151
  • 152
  • 153
  • 154
  • 155
  • 156
  • 157
  • 158
  • 159
  • 160
  • 161
  • 162
  • 163
  • 164
  • 165
  • 166
  • 167
  • 168
  • 169
  • 170
  • 171
  • 172
  • 173
  • 174
  • 175
  • 176
  • 177
  • 178
  • 179
  • 180
  • 181
  • 182
  • 183
  • 184
  • 185
  • 186
  • 187
  • 188
  • 189
  • 190
  • 191
  • 192
  • 193
  • 194
  • 195
  • 196
  • 197
  • 198
  • 199
  • 200
  • 201
  • 202
  • 203
  • 204
  • 205
  • 206
  • 207
  • 208
  • 209
  • 210
  • 211
  • 212
  • 213
  • 214
  • 215
  • 216
  • 217
  • 218
  • 219
  • 220
  • 221
  • 222
  • 223
  • 224
  • 225
  • 226
  • 227
  • 228
  • 229
  • 230
  • 231
  • 232

Chapter 1. Technical description
11
This is new and the “closest” to the processor, and is used to store
micro-operations. These are decoded executable machine instructions. It
serves these to the processor at rated speed. This additional level of cache
saves decode time on cache hits.
±
Level 2 cache, 256 KB
This is equivalent to L1 cache on the Pentium® III Xeon.
±
Level 3 cache 3–6 MB
This is equivalent to L2 cache on the Pentium III Xeon or the L3 cache on the
Pentium Xeon MP processor. Unlike the design of the original Itanium
processor, this L3 cache is now on the processor die, greatly improving
performance, up to 2 times greater than that of the original Itanium.
The x455 also implements a Level 4 cache as described in “IBM XceL4
Accelerator Cache” on page 12.
Intel has also introduced a number of features associated with its Itanium
micro-architecture. These are available in the x455, including:
±
400 MHz frontside bus
The Pentium III Xeon processor had a 100 MHz frontside bus that equated to
a burst throughput of 800 MBps. With protocols such as TCP/IP, this had been
shown to be a bottleneck in high-throughput situations. The Itanium 2
processor improves on this by using a single 200 MHz clock but using both
edges of each clock cycle to transmit data. This is shown in Figure 1-6.
Figure 1-6
Dual-pumped frontside bus
This increases the performance of the frontside bus. The end result is an
effective burst throughput of 6.4 GBps (128-bit wide data path running at 400
MHz), which can have a substantial impact, especially on TCP/IP-based LAN
traffic. This is opposed to the Itanium processor, which had a burst throughput
of only 2.1 GBps (64-bit wide data path running at 266 MHz).
±
Explicitly Parallel Instruction Computing (EPIC)
EPIC technology, developed by Intel and HP, leads to more efficient, faster
processors because it eliminates numerous processing inefficiencies in
current processors and attacks the perennial data bottleneck problems by
increasing parallelism, rather than simply boosting the raw “clock” speed of
the processor.
200 MHz clock