From WikiChip
Editing intel/dl boost

Warning: You are not logged in. Your IP address will be publicly visible if you make any edits. If you log in or create an account, your edits will be attributed to your username, along with other benefits.

The edit can be undone. Please check the comparison below to verify that this is what you want to do, and then save the changes below to finish undoing the edit.

This page supports semantic in-text annotations (e.g. "[[Is specified as::World Heritage Site]]") to build structured and queryable content provided by Semantic MediaWiki. For a comprehensive description on how to use annotations or the #ask parser function, please have a look at the getting started, in-text annotation, or inline queries help pages.

Latest revision Your text
Line 1: Line 1:
 
{{intel title|DL Boost}}
 
{{intel title|DL Boost}}
'''DL Boost Technology''' ('''deep learning boost''') is an umbrella marketing term used by [[Intel]] for a collection of technologies designed for the [[acceleration]] of AI workloads, including both inference and training.
+
'''DL Boost''' is a marketing term used by [[Intel]] that encompasses a number of [[x86]] technologies designed to [[accelerate]] AI workloads.
 
 
== Overview ==
 
'''DL Boost''' is a term used by [[Intel]] to describe a set of features on their microprocessors designed to accelerate AI workloads. The term was first introduced with {{intel|Cascade Lake|l=arch}} but has since been extended further with more capabilities in newer microarchitectures.
 
 
 
DL Boost includes the following features:
 
 
 
* {{x86|AVX512_VNNI|AVX-512 Vector Neural Network Instructions}} (AVX512_VNNI) - an instruction set extension that introduces reduced-precision (8-bit and 16-bit) multiply-accumulate for the acceleration of inference. VNNI was first introduced with {{intel|Cascade Lake|l=arch}} (server) and {{intel|Ice Lake (Client)|Ice Lake|l=arch}} (Client)
 
* {{x86|AVX512_BF16|AVX-512 BFloat16 Instructions}} (AVX512_BF16) - an instruction set extension for converting to [[bfloat16]] and then performing multiply-accumulate on such values for the acceleration of both inference and training. BG16 was first introduced with {{intel|Cooper Lake|l=arch}}.
 
* {{x86|AMX|Advanced Matrix Extension}} (AMX) - an extension that introduces a matrix register file and matrix operations. AMX was first introduced with {{intel|Sapphire Rapids|l=arch}}.
 
 
 
== Implementations ==
 
{| class="wikitable"
 
|-
 
! Microarchitecture !! {{x86|AVX512_VNNI}} !! {{x86|AVX512_BF16}} !! {{x86|AMX}}
 
|-
 
! colspan="4" | Client
 
|-
 
| {{intel|Ice Lake (Client)|l=arch}} || {{tchk|yes}} || {{tchk|no}} || {{tchk|no}}
 
|-
 
! colspan="4" | Server
 
|-
 
| {{intel|Cascade Lake|l=arch}} || {{tchk|yes}} || {{tchk|no}} || {{tchk|no}}
 
|-
 
| {{intel|Cooper Lake|l=arch}} || {{tchk|yes}} || {{tchk|yes}} || {{tchk|no}}
 
|-
 
| {{intel|Ice Lake (Server)|l=arch}} || {{tchk|yes}} || {{tchk|no}} || {{tchk|no}}
 
|-
 
| {{intel|Sapphire Rapids|l=arch}} || {{tchk|yes}} || {{tchk|yes}} || {{tchk|yes}}
 
|-
 
| {{intel|Granite Rapids|l=arch}} || {{tchk|some|TBD}} || {{tchk|some|TBD}} || {{tchk|some|TBD}}
 
|-
 
| {{intel|Diamond Rapids|l=arch}} || {{tchk|some|TBD}} || {{tchk|some|TBD}} || {{tchk|some|TBD}}
 
|}
 
 
 
== See also ==
 
* [[bfloat16]]
 
* [[Neural processors]]
 
* [[Acceleration]]
 
  
 
[[category:intel]]
 
[[category:intel]]

Please note that all contributions to WikiChip may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see WikiChip:Copyrights for details). Do not submit copyrighted work without permission!

Cancel | Editing help (opens in new window)