From WikiChip
Difference between revisions of "nec/microarchitectures/sx-aurora"
< nec

(Memory subsystem)
(Vector engine (VE) card)
Line 145: Line 145:
  
 
== Vector engine (VE) card ==
 
== Vector engine (VE) card ==
{{empty section}}
+
The SX-Aurora chip is part of a complete [[accelerator card]] known as the Vector Engine v1. NEC makes those in three varients - passive cooling, active air cooling, and water-cooled.
 +
 
 +
<gallery heights=225px widths=325px>
 +
File:sx-aurora ve 1 active air cooling.jpg|Active air cooling card
 +
File:sx-aurora ve 1 passive air cooling.jpg|Passive air cooling card
 +
File:sx-aurora ve 1 air cooling no cover.jpg|Air cooling without cover
 +
File:sx-aurora ve 1 naked card.jpg|Naked VE card
 +
File:sx-aurora ve 1 naked card 2.jpg|Naked VE card
 +
</gallery>
  
 
== Die ==
 
== Die ==

Revision as of 13:48, 9 December 2018

Edit Values
SX-Aurora µarch
General Info
Arch TypeVPU
DesignerNEC
ManufacturerTSMC
Introduction2018
Core Configs8
Pipeline
TypeSuperscalar, Pipelined
OoOEYes
SpeculativeYes
Reg RenamingYes
Stages8
Decode4-way
Cache
L1I Cache32 KiB/core
L1D Cache32 KiB/core
L2 Cache256 KiB/core
L3 Cache16 MiB/chip
Succession

SX-Aurora is NEC's successor to the SX-ACE, a 16 nm microarchitecture for vector processors first introduced in 2018.

History

The SX-Aurora is the successor to the SX-ACE, a series of vector processors that NEC has been developing since the early 1980s. NEC first announced the SX-Aurora TSUBASA product line on October 26 2017 with a complete lineup that ranges from a workstation featuring one vector engine (VE) card to a full supercomputer with 64 VEs. NEC disclosed the architecture of the chip at Hot Chips 30 in August 2018.

Architecture

Key changes from SX-ACE

  • 16 nm process (from 28 nm)
  • 1.6x frequency (1.6 GHz, up from 1 GHz)
  • 2x vector cores (8, up from 4)
  • Vector core
    • 1.5x FMAs EUs (3, up from 2)
    • 2x VPPs (32, up from 16)
    • 3x FLOPs/cycle (192 FLOPs/cycle, up from 64 FLOPs/cycle)
  • Memory
    • 16 MiB L3 LLC
    • 6x HBM2 (from 12x DDR3)
      • 4.7x memory bandwidth (1.2 TB/s, up from 256 GB/s)

This list is incomplete; you can help by expanding it.

Block Diagram

Entire SoC

sx-aurora block diagram.svg

Vector core

sx-aurora vector core block diagram.svg

Memory Hierarchy

  • Vector core
  • L3 Cache/LLC:
    • 16 MiB
      • 8 x 2 MiB
      • write-back
      • inclusive of L1 & L2
      • 128 banks
      • 3 TiB/s bandwidth
  • System DRAM:
    • 4Hi / 8Hi HBM2
    • 6 SDRAM KGD stacks
    • 1.2 TiB/s

Overview

Overview of the SX-Aurora

The SX-Aurora is NEC's successor to the SX-ACE, a vector processor designed for high-performance scientific/research applications and supercomputers. The SX-Aurora deviates from all prior chips in the kind of markets it's designed to address. Therefore, NEC made slightly different design choice compared to prior generations of vector processors. In an attempt to broaden their market, NEC extended beyond supercomputers to the conventional server and workstation market. This is done through the use of PCIe-based accelerator cards.

Moving to an accelerator card is not without its challenges. To keep the high memory bandwidth, and thus high bytes per FLOP, while moving a smaller form factor, it was necessary to drop the large amount of DDR memory channels. Instead, NEC opted to utilizing high-bandwidth memory on-chip instead. The card itself is designed to communicate with other cards on the system in order to scale up from just a single card for workstation use to a supercomputer with 64 cards per rack.

The chip itself consists of eight very big cores along with 16 MiB of last level cache on a 2-dimensional mesh. Attached to the LLC are the two memory controllers which interface with the six high-bandwidth memory sitting on an interposer. Fabricated on TSMC's 16 nm process, the SX-Aurora operates at up to 1.6 GHz delivering up to 307.2 gigaFLOPS (double-precision) per core for a total of up to 2.45 teraFLOPS.

Vector core

Vector core

Like its predecessor, the SX-Aurora vector core itself has three main components - the scalar processing unit (SPU), the vector processing unit (VPU), and the memory subsystem. Though all three were enhanced over the SX-ACE, the emphasis was on the VPU and the new memory subsystem. The majority of the compute capability is in the VPU which can deliver up to 192 DP FLOPS per cycle for up to 307.2 GFLOPS at 1.6 GHz. The VPU is fed by a high-bandwidth mesh capable of 256 B/cycle for a total of up to 410 GB/s for load and store each. The SPU is designed to provide all basic functionalities a typical CPU. Because the SX-Aurora is not a typical offload engine but a self-hosted accelerator, the SPU is designed to deliver high performance to keep pace with the VPU and other operating system-related tasks.

The memory subsystem on the SX-Aurora has been greatly enhanced from the SX-ACE. The primary focus of here is to support contiguous vector memory accesses. To that end, in order to feed the vector pipeline, the address generation unit can receive a vector instruction from the SPU in advance in order to calculate the address and disperse the request to the sixteen ports on the mesh network. 16 elements/cycle vector address generation and translation, as well as 17 requests issued/cycle, can be performed. The bandwidth of the crossbars is matched with the bandwidth of the loads and stores at 256 B/cycle for up to 410 GB/s for load and stores each.

Scalar processing unit

The scalar processing unit (SPU) is 4-wide.

Although the majority of the processing is handled by the VPU, the scalar processing unit (SPU) has to have a reasonably high performance in order to act as a host CPU for all the serial workloads including handling all the required operating system tasks. Additionally, the SPU has to provide the VPU with a sufficient stream of operations in order to maintain a high sustained performance as well as to calculate the base memory address for vector accesses.

Front-end

Scheduler

The SPU has 32 KiB of level 1 cache from which it fetches instructions from. The SPU on the SX-Aurora is four wide - capable of fetching and decoding four instructions per cycle. NEC stated that the SPU features a sophisticated branch predictor for hardware prefetching, however, they did not delve into any details. The SPU supports out-of-order execution with 8 stages of speculative execution.

Back end

Execution Ports

On the back end, the SPU includes a unified scheduler capable of issuing up to five instructions per cycle. Instructions are issued out-of-order once dependencies are resolved. The SX-Aurora has five separate execution units - two general ALUs for both integer and floating-point pipelines for general scalar arithmetic, a dedicated branch ALU, a single LSU execution unit, and a vector EU. Beyond the typical scalar operations, the SPU is required to calculate the base memory address for the vector accesses. Typically this is done prior to the vector instructions being sent to the VPU in order to ensure that the data is ready by that time. The SPU features a vector coherence control logic which supports high-speed vector memory accesses with up to 16 elements/cycle for vector stores. The SPU can send up to a single instruction to the VPU each cycle which goes to the instruction buffer of the VPU for reodering and scheduling.

Vector processing unit

vector processing unit (VPU) and 32 VPPs

The bulk of the compute work is done on the vector processing unit (VPU). The VPU has a fairly simple pipeline, though it does employes out-of-order scheduling. Instructions issued by the SPU are sent to the instruction buffer where they await renaming, reordering, and scheduling. NEC renames the 64 vector registers (VRs) into 256 physical registers. There is support for enhanced preloading and avoids WAR/WAW dependencies. Scheduling is relatively simple. There is a dedicated pipeline for complex operations. Things such as vector summation, division, mask population count, are sent to this execution unit. The dedicate execution unit for complex operations is there to prevent stalls due to the high latency involved in those operations.

The majority of the operations are handled by the vector parallel pipeline (VPP). The SX-Aurora doubles the number of VPPs per VPU from the SX-ACE. Each VPU now has 32 VPPs - all identical. Note that all of the control logic described before are outside of the VPP which is relatively a simple block of vector execution. The VPP has an eight-port vector register, 16 mask registers, and six execution pipes, and a set of forwarding logic between them.

The six execution pipes include three floating-point pipes, two integer ALUs, and a complex and store pipe for data output. Note that ALU1 and the Store pipe share the same read ports. Likewise, FMA2 and ALU0 share a read port. All in all, the effective number of pipelines executing each cycle is actually four. Compared to the SX-Ace, the SX-Aurora now has one extra FMA unit per VPP.

The peak theoretical performance that can be achieved is 3 FMAs per VPP per cycle. With 32 VPPs per VPU, there are a total of 96 FMAs/cycle for a total of 192 DP FLOPs/cycle. With a peak frequency of 1.6 GHz for the SX-Aurora Tsubasa vector processor, each VPU has a peak performance of 307.2 gigaFLOPS. Each FMA can perform operations on packed data types. That is, the single-precision floating-point is doubled through the packing of 2 32-bit elements for a peak performance of 614.4 gigaFLOPS.

Memory subsystem

Maintaining a high bytes Per FLOP is important vector operations that rely on large data sets. With over five times the FLOPS per core, the SX-Aurora had to significantly improve the memory subsystem to prevent workloads from memory bottlenecking, preventing them from reaching the peak compute power of the chip. The SX-Ace reached 256 GB/s of memory bandwidth using a whopping 16 channels of DDR3 memory. It becomes impossible to increase this further to bring a sufficiently large improvement in bandwidth. For this reason, NEC opted to use HBM2 memory instead. The SX-Aurora has six HBM2 modules delivering 1.22 TB/s of bandwidth, nearly 5-fold improvement over the SX-Ace. This provides the new last level cache with around 200 GB/s of bandwidth (1.6 TB/s of aggrageted bandwidth).

The SX-Aurora features a 16-layer 2D mesh network. With the SX-Ace, replys to the cores were slightly outbalanced. In the SX-Aurora, the requests crossbars and the reply crossbars, each, have the same bandwidth. With each request being 16B, this works out to around 410 GB/s for each crossbar or 820 GB/s per core. Coming from the cache size, there is a total of 3.05 GB/s of avilable bandwidth distributed accross all eight LLC chunks.

SX-Aurora Memory Subsystem.

Mesh interconnect

New text document.svg This section is empty; you can help add the missing info by editing this page.
SX-Aurora utilizes a 16-layer 2D mesh.

Package

The SX-Aurora chip uses six HBM2 stacks. Those are either 4 Hi or 8 Hi stacks. The chip utilizes TSMC's second-generation chip on wafer on substrate (CoWoS) technology with NEC's implementation developed in collaboration with TSMC and Broadcom. This chip became the world's first to utilize six HBM2s.


The package itself is very big at 60 mm x 60 mm. The VE processor die itself is 15 mm x 33 mm with a very large interposer with a total Si area of 1,235 mm² (32.5 mm x 38 mm).


nec sx-aurora tsubasa package.svg


Though other chips have reached very large interposer sizes before, the SX-Aurora is the first 6 HBM2 implementation. It uses the second-generation CoWoS packaging technology (CoWoS-XL2) to exceed the reticle size through the use of mask stitching.


sx-aurora-package-xsection.svg

Vector engine (VE) card

The SX-Aurora chip is part of a complete accelerator card known as the Vector Engine v1. NEC makes those in three varients - passive cooling, active air cooling, and water-cooled.

Die

  • 16 nm process
  • 4,800,000,000 transistors
  • 14.96 mm x 33.00 mm
    • 493.68 mm² die size

Bibliography

  • Template:hcbib
  • Supercomputing 2018, NEC Aurora Forum
  • Some information was obtained directly from NEC