| Literature DB >> 35167265 |
Guilherme Migliato Marega1,2, Zhenyu Wang1,2, Maksym Paliy3, Gino Giusi4, Sebastiano Strangio3, Francesco Castiglione5, Christian Callegari5, Mukesh Tripathi1,2, Aleksandra Radenovic6, Giuseppe Iannaccone3,5, Andras Kis1,2.
Abstract
Machine learning and signal processing on the edge are poised to influence our everyday lives with devices that will learn and infer from data generated by smart sensors and other devices for the Internet of Things. The next leap toward ubiquitous electronics requires increased energy efficiency of processors for specialized data-driven applications. Here, we show how an in-memory processor fabricated using a two-dimensional materials platform can potentially outperform its silicon counterparts in both standard and nontraditional Von Neumann architectures for artificial neural networks. We have fabricated a flash memory array with a two-dimensional channel using wafer-scale MoS2. Simulations and experiments show that the device can be scaled down to sub-micrometer channel length without any significant impact on its memory performance and that in simulation a reasonable memory window still exists at sub-50 nm channel lengths. Each device conductance in our circuit can be tuned with a 4-bit precision by closed-loop programming. Using our physical circuit, we demonstrate seven-segment digit display classification with a 91.5% accuracy with training performed ex situ and transferred from a host. Further simulations project that at a system level, the large memory arrays can perform AlexNet classification with an upper limit of 50 000 TOpS/W, potentially outperforming neural network integrated circuits based on double-poly CMOS technology.Entities:
Keywords: MoS2; beyond-Moore; in-memory computing; nanoelectronics; two-dimensional materials; two-dimensional semiconductors
Year: 2022 PMID: 35167265 PMCID: PMC8945700 DOI: 10.1021/acsnano.1c07065
Source DB: PubMed Journal: ACS Nano ISSN: 1936-0851 Impact factor: 15.881
Figure 1Device structure and characterization. (a) 3D schematic representation of the MoS2 memory device array and the corresponding circuit schematic for the multiplication-accumulation operation. (b) Optical image of an array of memories connected in parallel (scale bar: 50 μm). (c) IDS as a function of VG for constant drain-source voltage, VDS = 50 mV. (d) IDS as a function of VDS for different programming voltages, showing the programmable conductance behavior. The device is read using VG(READ) = 0 V and VDS = 50 mV.
Figure 2Device scaling. (a) Simulated hysteresis cycle as the device gate length is scaled from L = 1 μm to L = 50 nm. (b) Calculated threshold voltage shift (for IDS = 10–10A·μm–1) as a function of programming time tPROG. (c) Calculated threshold voltage shift for different channel lengths with a program time of 1 μs. (d) Experimental hysteresis cycle (IDS versus VG with VDS = 500 mV) of devices with 950, 430, and 180 nm gate length. The curves shown were select as the median behavior from the experimental data set. (e) Experimental variation of the ON current for different devices with gate lengths demonstrated in (d). Triangle: experimental data. Dot: average value. Error bar: confidence interval with 95% certainty.
Figure 3Closed-loop programming. (a) Block diagram explaining the closed-loop programming procedure. (b) Convergence map for overshoot of the weight and progressively decreasing the weight until the correct value has been reached.
Figure 4In-memory dot product. (a) Realization of the dot-product operation using two memories connected in parallel. (b–d) Data surface showing the equivalent multiplication-sum planes of a dot-product with the following weights: (b) w1 = 1, w2 = 0; (c) w1 = 0.4, w2 = 0.6; (d) w1 = 0, w2 = 1.
Figure 5Classification of a seven-segment digit in memory. (a) Representation of a seven-segment display. (b) One-layer perceptron network for seven-segment figure classification. (c) Transfer of learning of the theoretical weight matrix to proportional conductance values of individual memories. (d) Sample of inference operations after different test signals are sent to the input layer and measured in one of the neurons. (e) Effect of the signal noise on the classification accuracy. (f) Effect of the programming resolution on the classification accuracy.
Figure 6System-level analysis. (a) Analogue vector-matrix multiplier circuit with floating gate memory devices. (b) Transfer characteristics of the memory cells and of the extracted SPICE models in inversion. (c) Transfer characteristics of the memory cells and of the extracted SPICE models in the subthreshold. (d) Achievable ENOB of the multiplier as a function of the cell voltage bias. (e) Error rate in Imagenet classification for an analogue neural network as a function of the signal-to-noise-and-distortion ratio (SINAD) and of the number of bits.