From WikiChip
Editing nervana/nnp
Warning: You are not logged in. Your IP address will be publicly visible if you make any edits. If you log in or create an account, your edits will be attributed to your username, along with other benefits.
The edit can be undone.
Please check the comparison below to verify that this is what you want to do, and then save the changes below to finish undoing the edit.
This page supports semantic in-text annotations (e.g. "[[Is specified as::World Heritage Site]]") to build structured and queryable content provided by Semantic MediaWiki. For a comprehensive description on how to use annotations or the #ask parser function, please have a look at the getting started, in-text annotation, or inline queries help pages.
Latest revision | Your text | ||
Line 66: | Line 66: | ||
[[File:nnp-l-1000 announcement.png|thumb|right|NNP T-1000]] | [[File:nnp-l-1000 announcement.png|thumb|right|NNP T-1000]] | ||
{{main|nervana/microarchitectures/spring_crest|l1=Spring Crest µarch}} | {{main|nervana/microarchitectures/spring_crest|l1=Spring Crest µarch}} | ||
− | + | Second-generation NNP-Ts are branded as the NNP T-1000 series and are the first chips to be productized. Fabricated [[TSMC]]'s [[16 nm process]] based on the {{nervana|Spring Crest|Spring Crest microarchitecture|l=arch}}, those chips feature a number of enhancements and refinments over the prior generation including a shift from [[Flexpoint]] to [[Bfloat16]]. Intel claims that these chips have about 3-4x the training performance of first generation. Those chips come with 32 GiB of four [[HBM2]] stacks and are [[packaged]] in two forms - [[PCIe x16 Gen 3 Card]] and an [[OCP OAM]]. | |
[[File:spring_crest_ocp_board_(front).png|right|thumb|NNP-T 1400 [[OAM Module]].]] | [[File:spring_crest_ocp_board_(front).png|right|thumb|NNP-T 1400 [[OAM Module]].]] | ||
Line 107: | Line 107: | ||
[[File:ai hw summit supermicro ref pod rack.jpeg|right|thumb|POD Rack]] | [[File:ai hw summit supermicro ref pod rack.jpeg|right|thumb|POD Rack]] | ||
Along with the launch of the NNP-T 1000 series, Intel also introduced the POD reference design. Those systems were intended for large-scale out systems for the processing of very large neural networks. The POD reference design featured 10 racks with 6 nodes per rack. Each of the nodes features eight interconnected OAM cards, producing a system with a total of 480 NNP-Ts. | Along with the launch of the NNP-T 1000 series, Intel also introduced the POD reference design. Those systems were intended for large-scale out systems for the processing of very large neural networks. The POD reference design featured 10 racks with 6 nodes per rack. Each of the nodes features eight interconnected OAM cards, producing a system with a total of 480 NNP-Ts. | ||
− | |||
:[[File:ai hw summit supermicro ref pod.jpeg|500px]] | :[[File:ai hw summit supermicro ref pod.jpeg|500px]] |
Facts about "Neural Network Processors (NNP) - Intel Nervana"
designer | Intel + |
first announced | May 23, 2018 + |
first launched | 2019 + |
full page name | nervana/nnp + |
instance of | integrated circuit family + |
main designer | Intel + |
manufacturer | Intel + and TSMC + |
name | NNP + |
package | PCIe x16 Gen 3 Card +, OCP OAM + and M.2 + |
process | 28 nm (0.028 μm, 2.8e-5 mm) +, 16 nm (0.016 μm, 1.6e-5 mm) + and 10 nm (0.01 μm, 1.0e-5 mm) + |
technology | CMOS + |