(→Documents) |
|||
(20 intermediate revisions by the same user not shown) | |||
Line 14: | Line 14: | ||
|l1d=2 KiB | |l1d=2 KiB | ||
|l1d per=core | |l1d per=core | ||
− | |successor=Larrabee | + | |successor=Rock Creek |
− | |successor link=intel/microarchitectures/larrabee | + | |successor link=intel/microarchitectures/rock creek |
+ | |successor 2=Larrabee | ||
+ | |successor 2 link=intel/microarchitectures/larrabee | ||
}} | }} | ||
− | '''Polaris''' was a research [[microarchitecture]] designed by [[Intel]] [[Intel Labs|Labs]] demonstarting the theoretical capabilities of a [[many-core]] chip performing 1 [[trillion floating point operations]]. | + | '''Polaris''' was a research [[microarchitecture]] designed by [[Intel]] [[Intel Labs|Labs]] demonstarting the theoretical capabilities of a [[many-core]] chip performing 1 [[trillion floating point operations]]. Polaris built the foundation that eventually resulted in Intel's {{intel|MIC architecture|Many Integrated Cores (MIC) Architecture}}. |
== History == | == History == | ||
Line 34: | Line 36: | ||
*** 20 MiB, 256 KiB/core | *** 20 MiB, 256 KiB/core | ||
* High efficiency | * High efficiency | ||
− | ** Up to 27 GFLOPS/W with 310 GFLOPS | + | ** Up to 27 GFLOPS/W with 310 [[GFLOPS]] |
* [[network on a chip]] (NoC) | * [[network on a chip]] (NoC) | ||
** mesh architecture | ** mesh architecture | ||
Line 57: | Line 59: | ||
Each tile contains the core and the crossbar router. | Each tile contains the core and the crossbar router. | ||
+ | === Core === | ||
+ | The core, also called a processing engine (PE), executes 96-bit [[VLIW]]s encoding up to eight operations per cycle. The core contains two independent [[fused multiply-accumulate]]s. Each FMA can sustain a multiply-add result (2 FLOPs) every cycle. Operating at 4 GHz, each core is capable of 16 GFLOPS, for a total chip peak performance of 1.28 TFLOPS. | ||
+ | |||
+ | There is a 3 KiB [[single-cycle]] [[instruction cache]] and a 2 KiB [[data cache]]. Each core has a single 32-entry [[register file]] with 10 ports (6 reads and 4 writes). A large number of ports allows for the simultaneous execution of a large set of operations such as scheduling to both FMAs, performing a load and store from the L1D$, program control, dynamic sleep instruction, and perform package sending/receiving from the network. Two 32-bit data words can be transferred each cycle between the register file and the data cache. Additionally, up to four 32-bit data words can be issued to the two FMA units each cycle while retaining two 32-bit results on each cycle. | ||
+ | |||
+ | The two FMAs are implemented as fully-pipelined nine stages. This is implemented in nine stages using a single-cycle accumulate with base 32 and internal carry-save arithmetic and delay addition. To perform the single-cycle accumulate, the multiplier output carry-save format is kept and a [[Wallace tree]] of 4-2 carry-save adder accumulates the result in a temporary format. Intel claims this eliminated the need for a carry-propagate adder in the [[critical path]]. Additionally, since the accumulation is performed in base 32, the variable shifters were replaced by cheap constant shifters. | ||
+ | |||
+ | ==== FMA Unit ==== | ||
+ | The FMA unit, called single-precision floating-point multiply-accumulators (FPMAC), is a nine-stage base-32 single-cycle accumulate algorithm using an intermediary carry-save with delayed addition. Three cycles are dedicated to the multiplier unit (S<sub>1</sub> - S<sub>3</sub> for four Wallace steps), a single cycle is for the sign inversion (S<sub>4</sub>) followed by a single cycle for the accumulation loop (S<sub>5</sub>). Finally, three stages are dedicated to the addition, normalization, and conversion back to base-2 steps (S<sub>6</sub>-S<sub>8</sub>). | ||
+ | |||
+ | The unit takes in two {{arch|32}} [[IEEE-754]] [[single-precision]] operands and is designed to sustain a single FMA instruction every 250ps (4 GHz). The multiplier itself is a [[Wallace tree]] of 4-2 carry-save adders. In total, four Wallace tree stages are dedicated to the compression of partial product bits to a sum and carry and pair. Instead of the traditional carry propagate adder on the final stage, Intel retains the output in carry-save format and then converts the result to a base 32 at stage S<sub>3</sub> just before the accumulation. | ||
+ | |||
+ | == Network-on-chip (NoC) == | ||
=== Router === | === Router === | ||
[[File:intel polaris tiling.png|right|400px]] | [[File:intel polaris tiling.png|right|400px]] | ||
− | The underlying concept was to create a highly modular design which could scale from a low number of tiles to a large number of tiles without necessitating a large set of changes. Each tile is connected to a 5-port [[wormhole-switched router]] with [[mesochronous]] interfaces using two dead-lock free routing and a fully non-blocking crossbar switch with a total bandwidth of 80 GB/s at 4 GHz with a 1.25ns latency. Taking up a total area of just 0.34 mm², the router is connected to its four nearing neighbors - on the east, west, north, and south. The fifth port is designed to interface with the stacked SRAM which was located directly below the core itself. Links were designed with FIFO-based synchronization using a 4-deep circular FIFO which captures data using a delayed link strobe at the receiving end. Each lane has a 16 | + | The underlying concept was to create a highly modular design which could scale from a low number of tiles to a large number of tiles without necessitating a large set of changes. Each tile is connected to a 5-port [[wormhole-switched router]] with [[mesochronous]] interfaces using two lanes (lane 0/1) for dead-lock free routing and a fully non-blocking crossbar switch with a total bandwidth of 80 GB/s at 4 GHz (32-bit * 4 GHz * 5 ports) with a 1.25ns latency. Taking up a total area of just 0.34 mm², the router is connected to its four nearing neighbors - on the east, west, north, and south. The fifth port is designed to interface with the stacked SRAM which was located directly below the core itself. Links were designed with FIFO-based synchronization using a 4-deep circular FIFO which captures data using a delayed link strobe at the receiving end. Each lane has a 16 [[FLIT]] queue, arbiter, and flow control logic. The router is implemented using a 5-stage pipeline with a two-stage round robin arbitration. On the first stage, an input port is bound to an output port in each lane and on the second stage, a spending FLIT is then selected from one of the two lanes. This means the same lanes are reused per-FLIT basis each time. The crossbar data bus is 36 bits wide and is [[double-pumped]] at the 4th pipe stage using a [[dual edge-trigged]] [[flip-flop]], interleaving alternate data bits. This reduced the crossbar area by roughly 50%. |
+ | |||
+ | The entire on-chip network features a bisectional bandwidth of 256 GB/s. The router interface block (RIB) interfaces between the core and the router and performs the packet encapsulations. The architecture allows any core to send or receive instructions and data packets from and to any other core. | ||
− | The | + | === Packet === |
+ | Packets consist of two or more [[Flow control unITs]] (FLITs). A FLIT is 38 bits of information made up of a 6 control and 32 data bits. The header FLIT (FLIT0) is always sent. The header has a 3-bit destination ID field (DID) which indicates the router exit port. There are 30 bits for a total of 10 maximum hops in the header FLIT. This field is updated at each hop. Flow control management between routers is debit-based using almost-full bits - queue signals are sent by the receiver via the two flow control bits (FC<sub>0</sub> and FC<sub>1</sub>) when its buffers reach a specified threshold. For packets that need to exceed the 10 maximum hops in the header FLIT, the chained header (CH) bit is asserted to extend support for a second FLIT. | ||
− | === | + | The second FLIT (FLIT1) includes various control information such as sleep and wakeup control. While a minimum of two FLITs are always sent, there is no limit as to maximum number. |
− | + | ||
+ | |||
+ | : '''FLIT0''' <div><div style="text-align: center; width: 125px; float: left;">Control <sub>6 bit</sub><div style="border: 1px solid black;">FC<sub>0</sub> FC<sub>1</sub> L V T H</div></div><div style="float: left;"> </div><div style="text-align: center; width: 325px; float: left;">Data <sub>32 bit</sub><div style="border: 1px solid black;">CH DID DID DID DID DID DID DID DID DID DID</div></div></div> | ||
+ | {{clear}} | ||
+ | : '''FLIT1''' <div><div style="text-align: center; width: 125px; float: left;">Control <sub>6 bit</sub><div style="border: 1px solid black;">FC<sub>0</sub> FC<sub>1</sub> L V T H</div></div><div style="float: left;"> </div><div style="text-align: center; width: 325px; float: left;">Data <sub>32 bit</sub><div style="border: 1px solid black;">NPC PCA SLP REN 0 0 ADDR</div></div></div> | ||
+ | {{clear}} | ||
+ | : '''FLIT2''' <div><div style="text-align: center; width: 125px; float: left;">Control <sub>6 bit</sub><div style="border: 1px solid black;">FC<sub>0</sub> FC<sub>1</sub> L V T H</div></div><div style="float: left;"> </div><div style="text-align: center; width: 325px; float: left;">Data <sub>32 bit</sub><div style="border: 1px solid black;">d<sub>31</sub> ... d<sub>23</sub> ... d<sub>15</sub> ... d<sub>7</sub> ... d<sub>0</sub></div></div></div> | ||
+ | |||
+ | {{clear}} | ||
+ | :{| class="wikitable" | ||
+ | |- | ||
+ | | L || Lane ID || || T || Packet tail | ||
+ | |- | ||
+ | | V || Valid FLIT || || H || Packet Header | ||
+ | |- | ||
+ | | FC || Flow control (Lane 1:0) || || CH || Chained Header | ||
+ | |- | ||
+ | | NPC || New PC address enable || || PCA || PC address | ||
+ | |- | ||
+ | | SLP || PE sleep/wake || || REN || PE execution enable | ||
+ | |- | ||
+ | | ADDR || I$/D$ write address | ||
+ | |} | ||
− | + | === Router Interface Block (RIB) === | |
+ | The RIB serves as the interface between the core and the router and performances the necessary synchronizations of data transfers and the power management (e.g., since the core may be stalling on data). 38-bit FLITs get buffered in a 16-entry queue where they get multiplexed and routed to either the D$ (64-bit packets) or I$ (96-bit packet) based on the lane ID field of the FLIT. Note that buffering is done because register file stores to the data cache have priority over the RIB. FLIT1 gets decoded and several control signals are generated, allowing the program execution (REN) to start at the specific address (PCA) through the new program counter bit (NPC). | ||
− | + | Note that the RIB is also in charge of the power management during WAKE and SLEEP instructions and can put the entire core into sleep or wake it up depending on the incoming or outgoing packets. | |
== Freya (SRAM) == | == Freya (SRAM) == | ||
Line 78: | Line 119: | ||
The full chip uses 3,200 [[through-silicon vias]]. Using wafer-level stacking high-density interconnect could be realized, albeit at reduced flexibility and increased complexity. Overall a bandwidth of greater than 1 TB/s was demonstrated. | The full chip uses 3,200 [[through-silicon vias]]. Using wafer-level stacking high-density interconnect could be realized, albeit at reduced flexibility and increased complexity. Overall a bandwidth of greater than 1 TB/s was demonstrated. | ||
− | == Clocking == | + | == ISA == |
+ | Each core operation on a 96-bit [[very long instruction word]] which can pack up to eight operations to be issued in a single cycle. There are six classes of instructions with a large set of them possible to be performed simultaneously. For the most parts, instructions are 1-2 cycles in latency with the exception of the 9-cycle FMA. Instructions can issue to both FMA units and perform simultaneous data load and store from memory. There are also instructions for sending and receiving packets from the mesh network as well as program control jumps and branch instructions. There are a number of synchronization primitives for data transfer between cores and a number of sleep and wake instructions for stalling on data. Stalling can be done when the core is waiting for data to arrive from another tile in order to reduce the power consumption. | ||
+ | |||
+ | {| class="wikitable" | ||
+ | |- | ||
+ | ! Instruction Class !! Latency (cycles) | ||
+ | |- | ||
+ | | FPU || 9 (See [[#FMA Unit|see § FMA Unit]]) | ||
+ | |- | ||
+ | | Load/Store || 2 | ||
+ | |- | ||
+ | | Send/Receive || 2 | ||
+ | |- | ||
+ | | Jump/Branch || 1 | ||
+ | |- | ||
+ | | Stall/WaitForData (WFD) || N/A | ||
+ | |- | ||
+ | | Sleep/Wake || 1-6 | ||
+ | |} | ||
+ | |||
+ | == Clocking & Power == | ||
The on-die [[PLL]] uses horizontal M8 and vertical M7 layers for clock distribution. Each layer consists of differential clocks for ow due-cycle variations along the worst-case clock route which is 26 mm. The [[differential clocks]] are converted to a single-ended clock at 50% due cycle using the [[opamp]] located at each of the tiles. At 4 GHz, 1.2 V supply is 2.2 W, helped by the simultaneous distribution of clock switching across the entire die. | The on-die [[PLL]] uses horizontal M8 and vertical M7 layers for clock distribution. Each layer consists of differential clocks for ow due-cycle variations along the worst-case clock route which is 26 mm. The [[differential clocks]] are converted to a single-ended clock at 50% due cycle using the [[opamp]] located at each of the tiles. At 4 GHz, 1.2 V supply is 2.2 W, helped by the simultaneous distribution of clock switching across the entire die. | ||
+ | |||
+ | Polaris features a large number of power regions with 21 sleep regions per tile. The ISA offers the ability to sleep and wake any core on demand as desired. Additionally, it's possible to finely control the regions. For example, instructions can independently sleep and wake up each of the FMA units as needed. For example, the core or an FMA can be put to sleep until it recieves more work to operate on from another tile, allowing for an efficient producer-consumer type of workload. | ||
== Performance == | == Performance == | ||
Line 104: | Line 167: | ||
==== Chip ==== | ==== Chip ==== | ||
* Package LGA-1248 | * Package LGA-1248 | ||
+ | ** 66mm x 66mm | ||
** 8,390 [[C4 solder bumps]] | ** 8,390 [[C4 solder bumps]] | ||
** 14 layers organic substrate | ** 14 layers organic substrate | ||
Line 122: | Line 186: | ||
* 3 mm² die size | * 3 mm² die size | ||
:[[File:intel polaris core.png|400px]] | :[[File:intel polaris core.png|400px]] | ||
+ | |||
+ | |||
+ | :[[File:intel polaris core 3.png|600px]] | ||
=== Variant 2 === | === Variant 2 === | ||
Line 156: | Line 223: | ||
* IEEE ISSCC 2007 | * IEEE ISSCC 2007 | ||
* IEEE Hot Chips 20 Symposium (HCS) 2008. | * IEEE Hot Chips 20 Symposium (HCS) 2008. | ||
+ | * Vangal, Sriram R., et al. "An 80-tile sub-100-w teraflops processor in 65-nm cmos." IEEE Journal of Solid-State Circuits 43.1 (2008): 29-41. |
Latest revision as of 22:22, 17 May 2019
Edit Values | |
Polaris µarch | |
General Info | |
Arch Type | CPU |
Designer | Intel |
Manufacturer | Intel |
Introduction | February 2007 |
Process | 65 nm |
Core Configs | 80 |
Pipeline | |
Type | VLIW |
Stages | 9 |
Cache | |
L1I Cache | 3 KiB/core |
L1D Cache | 2 KiB/core |
Succession | |
Polaris was a research microarchitecture designed by Intel Labs demonstarting the theoretical capabilities of a many-core chip performing 1 trillion floating point operations. Polaris built the foundation that eventually resulted in Intel's Many Integrated Cores (MIC) Architecture.
Contents
History[edit]
Intel originally announced and presented a Polaris wafer at the September 2006 Intel Developer Forum. Technical details were finally presented at IEEE ISSCC 2007. It was a concept project designed to demonstrate the potential of integrating many processing elements on a single silicon chip enabled by Moore's Law in order to achieve a high trillion floating point operations throughput. Polaris was Intel's first public chip as a direct consequence of their Tera-scale Computing Research Program and is the basis of Intel's later research projects which paved the way for Intel's Many Integrated Cores (MIC) architecture and the Xeon Phi many-core processor family.
Architecture[edit]
- ≥ 1 teraFLOPS (single-precision)
- ≥ 5 GHz clock
- 3D die stacking
- ≥ 1 TB/s bandwidth
- stacked SRAM
- 20 MiB, 256 KiB/core
- High efficiency
- Up to 27 GFLOPS/W with 310 GFLOPS
- network on a chip (NoC)
- mesh architecture
- 80 tiles
Block Diagram[edit]
Entire Chip[edit]
Individual Tile[edit]
System Overview[edit]
Polaris is a monolithic silicon consisting of 80 tiles arranged in a two-dimensional array of ten rows by eight columns. Each tile is effectively a core (Intel terms a processing engine (PE)) hooked to a router with mesochronous interfaces which is used to forward the packets between the tiles, allowing for clock-phase-insensitive tile-to-tile communication and synchronous operations within each tile. Tiles are linked to all four of their nearest neighbors as well as directly to the stacked SRAM memory.
Tile[edit]
Each tile contains the core and the crossbar router.
Core[edit]
The core, also called a processing engine (PE), executes 96-bit VLIWs encoding up to eight operations per cycle. The core contains two independent fused multiply-accumulates. Each FMA can sustain a multiply-add result (2 FLOPs) every cycle. Operating at 4 GHz, each core is capable of 16 GFLOPS, for a total chip peak performance of 1.28 TFLOPS.
There is a 3 KiB single-cycle instruction cache and a 2 KiB data cache. Each core has a single 32-entry register file with 10 ports (6 reads and 4 writes). A large number of ports allows for the simultaneous execution of a large set of operations such as scheduling to both FMAs, performing a load and store from the L1D$, program control, dynamic sleep instruction, and perform package sending/receiving from the network. Two 32-bit data words can be transferred each cycle between the register file and the data cache. Additionally, up to four 32-bit data words can be issued to the two FMA units each cycle while retaining two 32-bit results on each cycle.
The two FMAs are implemented as fully-pipelined nine stages. This is implemented in nine stages using a single-cycle accumulate with base 32 and internal carry-save arithmetic and delay addition. To perform the single-cycle accumulate, the multiplier output carry-save format is kept and a Wallace tree of 4-2 carry-save adder accumulates the result in a temporary format. Intel claims this eliminated the need for a carry-propagate adder in the critical path. Additionally, since the accumulation is performed in base 32, the variable shifters were replaced by cheap constant shifters.
FMA Unit[edit]
The FMA unit, called single-precision floating-point multiply-accumulators (FPMAC), is a nine-stage base-32 single-cycle accumulate algorithm using an intermediary carry-save with delayed addition. Three cycles are dedicated to the multiplier unit (S1 - S3 for four Wallace steps), a single cycle is for the sign inversion (S4) followed by a single cycle for the accumulation loop (S5). Finally, three stages are dedicated to the addition, normalization, and conversion back to base-2 steps (S6-S8).
The unit takes in two 32-bit IEEE-754 single-precision operands and is designed to sustain a single FMA instruction every 250ps (4 GHz). The multiplier itself is a Wallace tree of 4-2 carry-save adders. In total, four Wallace tree stages are dedicated to the compression of partial product bits to a sum and carry and pair. Instead of the traditional carry propagate adder on the final stage, Intel retains the output in carry-save format and then converts the result to a base 32 at stage S3 just before the accumulation.
Network-on-chip (NoC)[edit]
Router[edit]
The underlying concept was to create a highly modular design which could scale from a low number of tiles to a large number of tiles without necessitating a large set of changes. Each tile is connected to a 5-port wormhole-switched router with mesochronous interfaces using two lanes (lane 0/1) for dead-lock free routing and a fully non-blocking crossbar switch with a total bandwidth of 80 GB/s at 4 GHz (32-bit * 4 GHz * 5 ports) with a 1.25ns latency. Taking up a total area of just 0.34 mm², the router is connected to its four nearing neighbors - on the east, west, north, and south. The fifth port is designed to interface with the stacked SRAM which was located directly below the core itself. Links were designed with FIFO-based synchronization using a 4-deep circular FIFO which captures data using a delayed link strobe at the receiving end. Each lane has a 16 FLIT queue, arbiter, and flow control logic. The router is implemented using a 5-stage pipeline with a two-stage round robin arbitration. On the first stage, an input port is bound to an output port in each lane and on the second stage, a spending FLIT is then selected from one of the two lanes. This means the same lanes are reused per-FLIT basis each time. The crossbar data bus is 36 bits wide and is double-pumped at the 4th pipe stage using a dual edge-trigged flip-flop, interleaving alternate data bits. This reduced the crossbar area by roughly 50%.
The entire on-chip network features a bisectional bandwidth of 256 GB/s. The router interface block (RIB) interfaces between the core and the router and performs the packet encapsulations. The architecture allows any core to send or receive instructions and data packets from and to any other core.
Packet[edit]
Packets consist of two or more Flow control unITs (FLITs). A FLIT is 38 bits of information made up of a 6 control and 32 data bits. The header FLIT (FLIT0) is always sent. The header has a 3-bit destination ID field (DID) which indicates the router exit port. There are 30 bits for a total of 10 maximum hops in the header FLIT. This field is updated at each hop. Flow control management between routers is debit-based using almost-full bits - queue signals are sent by the receiver via the two flow control bits (FC0 and FC1) when its buffers reach a specified threshold. For packets that need to exceed the 10 maximum hops in the header FLIT, the chained header (CH) bit is asserted to extend support for a second FLIT.
The second FLIT (FLIT1) includes various control information such as sleep and wakeup control. While a minimum of two FLITs are always sent, there is no limit as to maximum number.
- FLIT0 Control 6 bitFC0 FC1 L V T HData 32 bitCH DID DID DID DID DID DID DID DID DID DID
- FLIT1 Control 6 bitFC0 FC1 L V T HData 32 bitNPC PCA SLP REN 0 0 ADDR
- FLIT2 Control 6 bitFC0 FC1 L V T HData 32 bitd31 ... d23 ... d15 ... d7 ... d0
L Lane ID T Packet tail V Valid FLIT H Packet Header FC Flow control (Lane 1:0) CH Chained Header NPC New PC address enable PCA PC address SLP PE sleep/wake REN PE execution enable ADDR I$/D$ write address
Router Interface Block (RIB)[edit]
The RIB serves as the interface between the core and the router and performances the necessary synchronizations of data transfers and the power management (e.g., since the core may be stalling on data). 38-bit FLITs get buffered in a 16-entry queue where they get multiplexed and routed to either the D$ (64-bit packets) or I$ (96-bit packet) based on the lane ID field of the FLIT. Note that buffering is done because register file stores to the data cache have priority over the RIB. FLIT1 gets decoded and several control signals are generated, allowing the program execution (REN) to start at the specific address (PCA) through the new program counter bit (NPC).
Note that the RIB is also in charge of the power management during WAKE and SLEEP instructions and can put the entire core into sleep or wake it up depending on the incoming or outgoing packets.
Freya (SRAM)[edit]
In order to provide a very high amount of bandwidth Intel had to move the memory very close to the CPU. Even placed next to the CPU die in the same package was not sufficient enough to provide the necessary bandwidth they were seeing. Stacked under the Polaris chip is a large SRAM (codename Freya) consisting of 256 KiB of memory per core for a total of 20 MiB. The connection between the dies uses a highly dense bumps. Intel reported the density to be roughly 4x that of the their 4C. It's worth noting that the reason the SRAM is placed under the CPU die is to allow proper heat dissipation for the CPU die through an active cooling solution on top.
The full chip uses 3,200 through-silicon vias. Using wafer-level stacking high-density interconnect could be realized, albeit at reduced flexibility and increased complexity. Overall a bandwidth of greater than 1 TB/s was demonstrated.
ISA[edit]
Each core operation on a 96-bit very long instruction word which can pack up to eight operations to be issued in a single cycle. There are six classes of instructions with a large set of them possible to be performed simultaneously. For the most parts, instructions are 1-2 cycles in latency with the exception of the 9-cycle FMA. Instructions can issue to both FMA units and perform simultaneous data load and store from memory. There are also instructions for sending and receiving packets from the mesh network as well as program control jumps and branch instructions. There are a number of synchronization primitives for data transfer between cores and a number of sleep and wake instructions for stalling on data. Stalling can be done when the core is waiting for data to arrive from another tile in order to reduce the power consumption.
Instruction Class | Latency (cycles) |
---|---|
FPU | 9 (See see § FMA Unit) |
Load/Store | 2 |
Send/Receive | 2 |
Jump/Branch | 1 |
Stall/WaitForData (WFD) | N/A |
Sleep/Wake | 1-6 |
Clocking & Power[edit]
The on-die PLL uses horizontal M8 and vertical M7 layers for clock distribution. Each layer consists of differential clocks for ow due-cycle variations along the worst-case clock route which is 26 mm. The differential clocks are converted to a single-ended clock at 50% due cycle using the opamp located at each of the tiles. At 4 GHz, 1.2 V supply is 2.2 W, helped by the simultaneous distribution of clock switching across the entire die.
Polaris features a large number of power regions with 21 sleep regions per tile. The ISA offers the ability to sleep and wake any core on demand as desired. Additionally, it's possible to finely control the regions. For example, instructions can independently sleep and wake up each of the FMA units as needed. For example, the core or an FMA can be put to sleep until it recieves more work to operate on from another tile, allowing for an efficient producer-consumer type of workload.
Performance[edit]
Intel implemented key numerical algorithms from LAPACK on Polaris. Polaris, 110 °C, can achieve a maximum frequency of 3.125 GHz at 1 V for a peak performance of 1 teraFLOPS with a power consumption of 98 W. Alternatively, the chip can also do 4 GHz at 1.2 V for a performance of 1.28 TFLOPS and a power consumption of 181 W.
From a purely power-efficiency perspective, up to 27 GFLOPS/W can be achieved at 968.75 MHz at 0.6 V for a total performance of 310 GFLOPS. At that rating the chip power consumption is a mere 11 Watts.
Power | Voltage | Frequency | Bisectional Bandwidth | Performance | Efficiency |
---|---|---|---|---|---|
11 W | 0.6 V | 968.75 MHz | 496.64 Gb/s | 310 GFLOPS | 28.18 GFLOPS/W |
62 W | 0.95 V | 3.16 GHz | 1.62 Tb/s | 1.01 TFLOPS | 16.31 GFLOPS/W |
175 W | 1.2 V | 5.1 GHz | 2.61 Tb/s | 1.632 TFLOPS | 9.33 GFLOPS/W |
265 W | 1.35 V | 5.7 GHz | 2.92 Tb/s | 1,824 TFLOPS | 6.88 GFLOPS/W |
Die[edit]
Two different versions of the die were presented at a few different events. The exact difference is not known.
Variant 1[edit]
Chip[edit]
- Package LGA-1248
- 66mm x 66mm
- 8,390 C4 solder bumps
- 14 layers organic substrate
- 343 signal pins
- 65 nm process
- 1 poly, 8 metal (Cu) layers
- 21.72 mm x 12.64 mm
- 274.54 mm² die size
- 100,000,000 transistors
- 96,000,000 tiles
- 1,200,000 transistors/tile
- 4,000,000 other
- 96,000,000 tiles
Tile[edit]
- 1.5 mm x 2.0 mm
- 3 mm² die size
Variant 2[edit]
Chip[edit]
- 65 nm process
- 1 poly, 8 metal (Cu) layers
- 22 mm x 13.75
- 302.5 mm² die size
Tile[edit]
Additional Shots[edit]
Additional die and wafer shots provided by Intel:
Documents[edit]
- Teraflops Research Chip
- Tera Tera Tera, 2006
- Polaris, MPSoC 2007
- Aim High, Intel Technical Update, Teratec ’07 Symposium, June 20, 2007
- Teraflops Research Chip Overview, Feb 2007
References[edit]
- Intel Developer Forum, September 2006
- IEEE ISSCC 2007
- IEEE Hot Chips 20 Symposium (HCS) 2008.
- Vangal, Sriram R., et al. "An 80-tile sub-100-w teraflops processor in 65-nm cmos." IEEE Journal of Solid-State Circuits 43.1 (2008): 29-41.
codename | Polaris + |
core count | 80 + |
designer | Intel + |
first launched | February 2007 + |
full page name | intel/microarchitectures/polaris + |
instance of | microarchitecture + |
manufacturer | Intel + |
microarchitecture type | CPU + |
name | Polaris + |
pipeline stages | 9 + |
process | 65 nm (0.065 μm, 6.5e-5 mm) + |