From WikiChip
Difference between revisions of "intel/etann"
< intel

(Created page with "{{intel title|ETANN}} '''ETANN''' ('''Electronically Trainable Analog Neural Network''') was the first commercial neural processor, introduced by Intel in 1992...")
 
(Architecture)
(8 intermediate revisions by the same user not shown)
Line 1: Line 1:
 
{{intel title|ETANN}}
 
{{intel title|ETANN}}
'''ETANN''' ('''Electronically Trainable Analog Neural Network''') was the [[first]] commercial [[neural processor]], introduced by [[Intel]] in [[1992]]. Implemented on a [[1.0 µm process]], this chip incorporated 64 analog neurons and 10,240 analog synapses.
+
{{mpu
 +
|name=ETANN
 +
|designer=Intel
 +
|manufacturer=Intel
 +
|model number=80170NX
 +
|part number=80170NX
 +
|market=Artificial Intelligence
 +
|first announced=1989
 +
|first launched=1989
 +
|process=1 µm
 +
|technology=FGMOS
 +
|die area=93.15 mm²
 +
|die length=8.1 mm
 +
|die width=11.5 mm
 +
|power=2 W
 +
|average power=1.5 W
 +
|v core=5 V
 +
}}
 +
'''ETANN''' ('''Electronically Trainable Analog Neural Network''') was one of the [[first]] commercial [[neural processor]], introduced by [[Intel]] around 1989. Implemented on a [[1.0 µm process]], this chip incorporated 64 analog neurons and 10,240 analog synapses. The ETANN is also the first commercial [[analog]] neural processor and is considered to be the first successful commercial neural network chip.
 +
 
 +
== Overview ==
 +
The ETANN was originally announced at the [[1989]] International Joint Conference on Neural Networks (IJCNN). The chip was implemented using an analog nonvolatile floating gate technology on Intel's CHMOS-III 1µm nonvolatile memory technology. The chip  integrates a total of 64 analog neurons and 1024 analog nonvolatile synapses. The network calculated the [[dot product]] between the 64x64 nonvolatile EEPROM analog synaptic weight array and a 64-element analog input vector. The chip was reported the calculations to reach 2000 MCPs (million connections per second).
 +
 
 +
== Architecture ==
 +
The chip has two synapse weight arrays, both consists of 4096 normal weights and 1024 (64*16) bias weights. There are 16 bias weights per neuron, allowing for sufficient influence in the sum. With two synapse weight arrays, it's possible to make two layers of 64x64. The output of the 64 neurons in the first layer are stored in [[buffers]], which are then fed as inputs to the second synapse array. The same neurons are used again for the second slayer. Despite being described as having 64 neurons, effectively, this chip has 128 neurons. It's worth pointing out that weights are stored in EEPROMs, meaning Intel was able to eliminate the refresh circuitry that would otherwise be required to retain the weight values, wasting precious die area. There are some pretty big disadvantages with the use of [[EEPROM]] as well such as longer update time (100s of microseconds). It also means that the chip is not particularly suitable for applications that require frequency reprogramming. There is also a practical number of times that the weighted values can be programmed before degenerations is observed. Once the learning phase is complete and the weights are constants (i.e., no EEPROM updating), Intel claimed the chip is capable of reaching 2,000 MCPs or million connections per second.
 +
 
 +
Synapse multiplication is performed by a [[four-quadrant multiplier|four-quadrant]] differential [[Gilbert multiplier]] (an [[analog multiplier]]). Summing was done by collecting the differential output currents from the multipliers using a simple [[summing wire]]. The [[sigmoid functions]] are provided separately as 64 sigmoid circuits. The weighted sum have a 6-bit precision and are limited to around ± 2.5.
 +
 
 +
::[[File:ETANN block diagram.png|750px]]
 +
 
 +
=== Bit Slicing ===
 +
The chip is designed as a [[bit-slice microprocessor]]. Intel's own board allowed for up to 8 ETANNs chips to be wired together to form a larger network.
 +
 
 +
=== Training ===
 +
The '''Intel Neural Network Training System''' ('''iNNTS''') is a training device, accessible by a standard Intel PC. Intel provided a set of software and rivers for controlling the iNNTS.
 +
 
 +
== Die ==
 +
* CHMOS-III 1µm nonvolatile memory technology
 +
* 93.15 mm² die size
 +
** 8.1 mm x 11.5 mm
 +
 
 +
Floor plan:
 +
 
 +
::[[File:ETANN floor plan.png|500px]]
 +
 
 +
== References ==
 +
* Holler, Mark, et al. "An electrically trainable artificial neural network (ETANN) with 10240 floating gate synapses." International Joint Conference on Neural Networks. Vol. 2. 1989.
 +
* Castro, Hernan A., Simon M. Tam, and Mark A. Holler. "Implementation and performance of an analog nonvolatile neural network." Analog Integrated Circuits and Signal Processing 4.2 (1993): 97-113.

Revision as of 21:34, 20 November 2017

Template:mpu ETANN (Electronically Trainable Analog Neural Network) was one of the first commercial neural processor, introduced by Intel around 1989. Implemented on a 1.0 µm process, this chip incorporated 64 analog neurons and 10,240 analog synapses. The ETANN is also the first commercial analog neural processor and is considered to be the first successful commercial neural network chip.

Overview

The ETANN was originally announced at the 1989 International Joint Conference on Neural Networks (IJCNN). The chip was implemented using an analog nonvolatile floating gate technology on Intel's CHMOS-III 1µm nonvolatile memory technology. The chip integrates a total of 64 analog neurons and 1024 analog nonvolatile synapses. The network calculated the dot product between the 64x64 nonvolatile EEPROM analog synaptic weight array and a 64-element analog input vector. The chip was reported the calculations to reach 2000 MCPs (million connections per second).

Architecture

The chip has two synapse weight arrays, both consists of 4096 normal weights and 1024 (64*16) bias weights. There are 16 bias weights per neuron, allowing for sufficient influence in the sum. With two synapse weight arrays, it's possible to make two layers of 64x64. The output of the 64 neurons in the first layer are stored in buffers, which are then fed as inputs to the second synapse array. The same neurons are used again for the second slayer. Despite being described as having 64 neurons, effectively, this chip has 128 neurons. It's worth pointing out that weights are stored in EEPROMs, meaning Intel was able to eliminate the refresh circuitry that would otherwise be required to retain the weight values, wasting precious die area. There are some pretty big disadvantages with the use of EEPROM as well such as longer update time (100s of microseconds). It also means that the chip is not particularly suitable for applications that require frequency reprogramming. There is also a practical number of times that the weighted values can be programmed before degenerations is observed. Once the learning phase is complete and the weights are constants (i.e., no EEPROM updating), Intel claimed the chip is capable of reaching 2,000 MCPs or million connections per second.

Synapse multiplication is performed by a four-quadrant differential Gilbert multiplier (an analog multiplier). Summing was done by collecting the differential output currents from the multipliers using a simple summing wire. The sigmoid functions are provided separately as 64 sigmoid circuits. The weighted sum have a 6-bit precision and are limited to around ± 2.5.

ETANN block diagram.png

Bit Slicing

The chip is designed as a bit-slice microprocessor. Intel's own board allowed for up to 8 ETANNs chips to be wired together to form a larger network.

Training

The Intel Neural Network Training System (iNNTS) is a training device, accessible by a standard Intel PC. Intel provided a set of software and rivers for controlling the iNNTS.

Die

  • CHMOS-III 1µm nonvolatile memory technology
  • 93.15 mm² die size
    • 8.1 mm x 11.5 mm

Floor plan:

ETANN floor plan.png

References

  • Holler, Mark, et al. "An electrically trainable artificial neural network (ETANN) with 10240 floating gate synapses." International Joint Conference on Neural Networks. Vol. 2. 1989.
  • Castro, Hernan A., Simon M. Tam, and Mark A. Holler. "Implementation and performance of an analog nonvolatile neural network." Analog Integrated Circuits and Signal Processing 4.2 (1993): 97-113.