From WikiChip
Editing nvidia/tegra/xavier

Warning: You are not logged in. Your IP address will be publicly visible if you make any edits. If you log in or create an account, your edits will be attributed to your username, along with other benefits.

The edit can be undone. Please check the comparison below to verify that this is what you want to do, and then save the changes below to finish undoing the edit.

This page supports semantic in-text annotations (e.g. "[[Is specified as::World Heritage Site]]") to build structured and queryable content provided by Semantic MediaWiki. For a comprehensive description on how to use annotations or the #ask parser function, please have a look at the getting started, in-text annotation, or inline queries help pages.

Latest revision Your text
Line 5: Line 5:
 
|designer=Nvidia
 
|designer=Nvidia
 
|manufacturer=TSMC
 
|manufacturer=TSMC
|model number=Tegra194
 
 
|market=Artificial Intelligence
 
|market=Artificial Intelligence
|market 2=Embedded
 
 
|first announced=January 8, 2018
 
|first announced=January 8, 2018
|first launched=June, 2018
 
 
|family=Tegra
 
|family=Tegra
 
|isa=ARMv8
 
|isa=ARMv8
 
|isa family=ARM
 
|isa family=ARM
|microarch=Carmel
 
|microarch 2=Volta
 
|core name=Carmel
 
 
|process=12 nm
 
|process=12 nm
 
|transistors=9,000,000,000
 
|transistors=9,000,000,000
Line 23: Line 17:
 
|core count=8
 
|core count=8
 
|thread count=8
 
|thread count=8
|max cpus=4
 
 
|tdp=30 W
 
|tdp=30 W
 
|tdp typical=20 W
 
|tdp typical=20 W
 
}}
 
}}
'''Tegra Xavier''' is a {{arch|64}} [[ARM]] high-performance system on a chip for autonomous machines designed by [[Nvidia]] and introduced in [[2018]]. Xavier is incorporated into a number of [[Nvidia]]'s computers including the Jetson Xavier, Drive Xavier, and the Drive Pegasus.
+
'''Tegra Xavier''' is a {{arch|64}} [[ARM]] high-performance autonomous machine system on a chip designed by [[Nvidia]] and introduced in [[2018]]. Xavier is incorporated into Nvidia's Drive Pegasus, Jetson, and a number of other autonomous computer.  
  
 
== Overview ==
 
== Overview ==
 
[[File:xavier overview.png|right|thumb|Overview (HC 30)]]
 
[[File:xavier overview.png|right|thumb|Overview (HC 30)]]
Xavier is an autonomous machine processor designed by [[Nvidia]] and introduced at CES 2018. Silicon came back in the last week of December 2017 with sampling started in the first quarter of 2018. NVIDIA plans on mass production by the end of the year. NVIDIA reported that the product is a result of $2 billion R&D and 8,000 engineering years. The chip is said to have full redundancy and diversity in its functional blocks.
+
Xavier is an autonomous machine process designed by [[Nvidia]] and introduced at CES 2018. Silicon came back in the last week of December 2017 with sampling started in the first quarter of 2018. NVIDIA plans on mass production by the end of the year. NVIDIA reported that the product is a result of $2 billion R&D and 8,000 engineering years. The chip is said to have full redundancy and diversity in its functional blocks.
  
:[[File:xavier block.svg|500px]]
+
The design targets and architecture started back in [[2014]]. The chip itself comprises an eight-core CPU cluster, GPU, [[neural processor|deep learning accelerator]], vision accelerator, ISP, and a set of multimedia accelerators. Xavier features a large set of I/O and has been designed for safety and reliability supporting various standards such as Functional safety [[ISO-26262]] and [[ASIL]] level C. The CPU cluster is fully [[cache coherent]] and the coherency is extended to all the other [[accelerators]] on-chip.
 
 
The design targets and architecture started back in [[2014]]. Fabricated on [[TSMC]] [[12 nm process]], the chip itself comprises an eight-core CPU cluster, GPU with additional inference optimizations, [[neural processor|deep learning accelerator]], vision accelerator, and a set of multimedia accelerators providing additional support for machine learning (stereo, LDC, optical flow). The ISP has been enhanced to provide native HDR support, higher precision math without offloading work to the GPU. Xavier features a large set of I/O and has been designed for safety and reliability supporting various standards such as Functional safety [[ISO-26262]] and [[ASIL]] level C. The CPU cluster is fully [[cache coherent]] and the coherency is extended to all the other [[accelerators]] on-chip.
 
 
 
At the platform level, one of the bigger changes took place at the I/O subsystem. Xavier features [[NVLink]] 1.0 supporting 20 GB/s in each direction for connecting a [[discrete graphics processor]] to Xavier in a [[cache coherent]] manner. Xavier has PCIe Gen 4.0 support (16 GT/s). It's worth noting that Xavier added support for an end-point mode in addition to the standard root complex support. This support meant they can connect two Xaviers directly one to another (2-way multiprocessing) without going through a PCIe switch or alike.
 
  
 
== Architecture ==
 
== Architecture ==
 
=== CPU ===
 
=== CPU ===
[[File:xavier cpu complex.svg|right|thumb|300px|CPU Block Diagram]]
 
 
{{main|nvidia/microarchitectures/carmel|l1=Carmel core}}
 
{{main|nvidia/microarchitectures/carmel|l1=Carmel core}}
The chip features eight control/management {{nvidia|Carmel|l=arch}} cores, Nvidia's own custom {{arch|64}} [[ARM]] cores. Those cores implement [[ARMv8.2]] with [[RAS]] support and safety built-in, including dual-execution mode. The cluster consists of 4 duplexes, each sharing 2 MiB of L2 cache. All cores are fully [[cache coherent]] which is extended to {{nvidia|Volta|the GPU|l=arch}} and all the other accelerators in the chip. Compared to {{\\|Parker}} which was based on {{nvidia|Denver 2|l=arch}}, Nividia reports around 2x the multithreaded performance.
+
The chip features eight {{nvidia|Carmel|l=arch}} cores, Nvidia's own custom {{arch|64}} [[ARM]] cores.  
 
 
 
=== GPU ===
 
=== GPU ===
[[File:xavier gpu.svg|right|thumb|300px|GPU Block Diagram]]
 
 
{{main|nvidia/microarchitectures/volta|l1=Volta}}
 
{{main|nvidia/microarchitectures/volta|l1=Volta}}
Xavier implements a derivative of their {{nvidia|Volta|l=arch}} GPU with a set of finer changes to address the machine learning market, particularly improving inference performance over training. It has eight Volta stream multiprocessors along with their standard 128 KiB of L1 cache and a 512 KiB of shared L2. Compared to Parker, Nvidia claims this GPU has 2.1x the graphics performance. Whereas their desktop parts (e.g., GV100) are a very powerful GPU that is used for training, the GPU here is optimized for inference. The most obvious change is that each Volta multiprocessor contains eight tensor cores, each of which can perform 64x FP16 MACs or 128x INT8 MACs per cycle. All of this yields a maximum 22.6 tera-operations (int8) per second.
+
The chip incorporates a {{nvidia|Volta|l=arch}} GPU with 512 {{nvidia|CUDA Cores}} capable of operating in 64-bit and 32-bit floating point as well as 8-bit integer. This allows the various [[deep learning]] [[artificial neural networks]] types to run efficiently in the format most suitable for them. This translates to 1.3 CUDA TOPS (32-bit FP) and another 20 Tensor Core TOPS (16-bit FP).
 
 
{| class="wikitable"
 
|-
 
! Throughput
 
|-
 
| 22.6 DL TOPS 8-bit
 
|-
 
| 2.8 CUDA TFLOPS FP16
 
|-
 
| 1.4 CUDA TFLOPS FP32
 
|}
 
  
 
=== Accelerators ===
 
=== Accelerators ===
Xavier incorporates a set of accelerators designed to augment the functionality offered by the GPU and CPU in order to provide added flexibility and perhaps offer a way to implement some of the more common set of algorithms slightly more efficiently.
+
Xavier incorporates a Programmable Vision Accelerator (PVA) for processing computer vision. It is capable of 1.6 [[TOPS]] and the ability to do [[stereo disparity]] (e.g., processing parallax between two cameras to obtain useful information such as depth), optical flow (e.g., direction and speed of vectors), and image processing. Additionally, since the chip is expected to be connected to a network of cameras (e.g., side, front, inside), the chip is capable of doing real-time [[encoding]] for all camera in [[high dynamic range]].  
==== Programmable Vision Accelerator ====
 
[[File:xavier pva block.svg|right|thumb|300px|PVA Block Diagram]]
 
Xavier incorporates a Programmable Vision Accelerator (PVA) for processing computer vision. There are actually two exact instances of the PVA on-chip, each can be used in lock-step or independently and are capable of implementing some of the common filter loop and other detection algorithms (e.g. [[Harris corner]], [[fast Fourier transform|FFTs]]). For each of the PVAs, there is a {{armh|Cortex-R5|l=arch}} core along with two dedicated vector processing units, each with its own memory and DMA. The DMA on the PVA is designed to operate on tiles of memory. To that end, the DMA performs the address calculation and can perform prefetching while the processing pipes operate. This is 7-slot VLIW architecture made of 2 scalar slots, 2 vector slots, and 3 memory operations. The pipe is 256 bit wide (slightly wider because of the [[guard bits]] keeping the precision for the operation) and all types can operate at full throughput (32x8b, 16x16b, and 8x32b vector math). The pipe supports additional operations beyond vector such as custom logic for table lookup and hardware looping.
 
 
 
{| class="wikitable"
 
|-
 
! Throughput
 
|-
 
| 1.7 CV TOPS
 
|}
 
 
 
It's worth noting that since the chip is expected to be connected to a network of cameras (e.g., side, front, inside), the PVA is capable of doing real-time [[encoding]] for all camera in [[high dynamic range]].  
 
  
 
{| class="wikitable"
 
{| class="wikitable"
Line 88: Line 51:
  
 
==== Deep Learning Accelerator ====
 
==== Deep Learning Accelerator ====
The other accelerators on-die is the deep learning accelerator (DLA) which is actually a physical implementation of the open source Nvidia NVDLA architecture. Xavier has two instances of NVDLA which can offer a peak theoretical performance of 5.7 [[teraFLOPS]] (half precision FP) or twice the throughput at 11.4 TOPS for int8.
+
The chip incorporates a deep learning accelerator (DLA) that implements a number of specific set of deep learning functions common to many applications. This allows them to read the highest possible energy efficiency for those operations. The DLA has a peak performance of 5 [[TOPS]] for 16-bit integers or 10 [[TOPS]] for 8-bit integer.
 
 
{| class="wikitable"
 
|-
 
! Throughput
 
|-
 
| 11.4 TOPS (int8)
 
|-
 
| 5.7 TFLOPS (FP16)
 
|}
 
 
 
==== Stereo & Optical Flow Engine ====
 
Xavier has dedicated engines for stereo and optical flow
 
 
 
{| class="wikitable"
 
|-
 
! Throughput
 
|-
 
| 12.4 TOPS 8-bit
 
|-
 
| 6.2 TOPS 16-bit
 
|}
 
  
 
== Memory controller ==
 
== Memory controller ==
 
{{memory controller
 
{{memory controller
|type=LPDDR4X-4266
+
|type=LPDDR4-2133
 
|ecc=Yes
 
|ecc=Yes
 +
|width=32 bit
 
|channels=8
 
|channels=8
|width=32 bit
 
 
|max bandwidth=127.1 GiB/s
 
|max bandwidth=127.1 GiB/s
 
}}
 
}}
  
 
== I/O ==
 
== I/O ==
* [[NVLINK]]
+
* 16 [[Camera Serial Interface|CSI]] channels
** NVLINK 1.0 (20 GB/s)
+
** 109 Gbps total bandwidth
* [[PCIE]]
+
* 1x gE interface
** Multiple 16GT/s gen4 controllers
+
* 1x 10gE interface
** x8, x4, x2, x1 configurations
 
** Root port + Endpoint
 
* USB
 
** 3x USB3.1 (10 GT/s) ports
 
** 4x USB2.0 ports
 
* DISPLAY
 
** 4x DP/HDMI/eDP
 
** 4K @ 60 Hz
 
** DP HBR3
 
** HDMI 2.0
 
* CAMERA
 
** 16 CSI lanes
 
** 40 Gbps in DPHY 1.2 Mode
 
** 109 Gbps in CPHY 1.1 Mode
 
* OTHER
 
** Ethernet
 
** CAN
 
** UART
 
** UFS
 
** SDMMC
 
** SPIO
 
** I2C
 
** I2S
 
** GPIO
 
  
 
== Die ==
 
== Die ==
Line 164: Line 82:
 
* ~89.2 mm² silicon area
 
* ~89.2 mm² silicon area
  
:[[File:xavier die volta gpu.png|600px]]
+
:[[File:xavier die volta gpu.png|500px]]
  
  
:[[File:xavier die volta gpu (annotated).png|600px]]
+
:[[File:xavier die volta gpu (annotated).png|500px]]
  
 
=== CPU ===
 
=== CPU ===
Line 173: Line 91:
  
 
=== PVA ===
 
=== PVA ===
:[[File:xavier die pva.png|750px]]
+
:[[File:xavier die pva.png|400px]]
 
 
:[[File:xavier die pva (annotated).png|750px]]
 
  
 
=== MM Engine / DLA ===
 
=== MM Engine / DLA ===
 
* ~21.75 mm² silicon area
 
* ~21.75 mm² silicon area
  
:[[File:xavier die mm-dl accel.png|750px]]
+
:[[File:xavier die mm-dl accel.png|400px]]
 
 
== Board ==
 
<gallery mode=packed-hover heights="300px" widths="300px">
 
jetson_xavier_(front).png|Jetson Xavier, front
 
jetson_xavier_(back).png|Jetson Xavier, back
 
</gallery>
 
  
 
== Documents ==
 
== Documents ==
 
* [[:File:ces2018 - nvidia drive xavier.pdf|CES 2018: Nvidia Drive Xavier]]
 
* [[:File:ces2018 - nvidia drive xavier.pdf|CES 2018: Nvidia Drive Xavier]]
 
== See also ==
 
* Tesla {{teslacar|FSD Chip}}
 
  
 
== Bibliography ==
 
== Bibliography ==
 
* IEEE Hot Chips 30 Symposium (HCS) 2018.
 
* IEEE Hot Chips 30 Symposium (HCS) 2018.
* Schor, David. (September, 2018). "[https://fuse.wikichip.org/news/1618/hot-chips-30-nvidia-xavier-soc/ Hot Chips 30: Nvidia Xavier SoC]"
 

Please note that all contributions to WikiChip may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see WikiChip:Copyrights for details). Do not submit copyrighted work without permission!

Cancel | Editing help (opens in new window)
Facts about "Tegra Xavier - Nvidia"
core count8 +
core nameCarmel +
designerNvidia +
die area350 mm² (0.543 in², 3.5 cm², 350,000,000 µm²) +
familyTegra +
first announcedJanuary 8, 2018 +
first launchedJune 2018 +
full page namenvidia/tegra/xavier +
has ecc memory supporttrue +
instance ofmicroprocessor +
isaARMv8 +
isa familyARM +
ldateJune 2018 +
main imageFile:xavier soc chip.png +
manufacturerTSMC +
market segmentArtificial Intelligence + and Embedded +
max cpu count4 +
max memory bandwidth127.1 GiB/s (130,150.4 MiB/s, 136.473 GB/s, 136,472.586 MB/s, 0.124 TiB/s, 0.136 TB/s) +
max memory channels8 +
microarchitectureCarmel + and Volta +
model numberTegra194 +
nameXavier +
process12 nm (0.012 μm, 1.2e-5 mm) +
smp max ways4 +
supported memory typeLPDDR4X-4266 +
tdp30 W (30,000 mW, 0.0402 hp, 0.03 kW) +
tdp (typical)20 W (20,000 mW, 0.0268 hp, 0.02 kW) +
technologyCMOS +
thread count8 +
transistor count9,000,000,000 +
word size64 bit (8 octets, 16 nibbles) +