| Literature DB >> 32637614 |
Chen-Yu Wang1, Shi-Jun Liang1, Shuang Wang1, Pengfei Wang1, Zhu'an Li1, Zhongrui Wang2, Anyuan Gao1, Chen Pan1, Chuan Liu3, Jian Liu3, Huafeng Yang3, Xiaowei Liu1, Wenhao Song2, Cong Wang1, Bin Cheng1, Xiaomu Wang3, Kunji Chen3, Zhenlin Wang1, Kenji Watanabe4, Takashi Taniguchi4, J Joshua Yang2, Feng Miao1.
Abstract
Early processing of visual information takes place in the human retina. Mimicking neurobiological structures and functionalities of the retina provides a promising pathway to achieving vision sensor with highly efficient image processing. Here, we demonstrate a prototype vision sensor that operates via the gate-tunable positive and negative photoresponses of the van der Waals (vdW) vertical heterostructures. The sensor emulates not only the neurobiological functionalities of bipolar cells and photoreceptors but also the unique connectivity between bipolar cells and photoreceptors. By tuning gate voltage for each pixel, we achieve reconfigurable vision sensor for simultaneous image sensing and processing. Furthermore, our prototype vision sensor itself can be trained to classify the input images by updating the gate voltages applied individually to each pixel in the sensor. Our work indicates that vdW vertical heterostructures offer a promising platform for the development of neural network vision sensor.Entities:
Year: 2020 PMID: 32637614 PMCID: PMC7314516 DOI: 10.1126/sciadv.aba6173
Source DB: PubMed Journal: Sci Adv ISSN: 2375-2548 Impact factor: 14.136
Fig. 1Retinal and artificial retinal structures.
(A) Profile of a biological retina. (B) Biological working mechanism and photoresponse of OFF bipolar cells [with α-amino-3-hydroxy-5-methyl-4-isoxazolepropionic acid (AMPA)] and ON bipolar cells [with metabotropic glutamate receptor 6 (mGluR6)]. Black bars in the photoresponse of bipolar cells represent the moment of light illumination. (C) Optical image of a retinomorphic device based on a vdW vertical heterostructure. (D) Operating mechanism and photoresponse of the ON- and OFF-photoresponse devices at zero and negative gate voltages, respectively. The positive (negative) ∆Ids corresponds to ON-photoresponse (OFF-photoresponse). Shadow areas correspond to the duration of light illumination. (E) OFF-photoresponse at different bias voltages and light intensities (indicated by shadow areas). OFF-photoresponse of the device remains retained at extremely low bias voltage (10 mV), which allows the operation of low power consumption.
Fig. 2The retinomorphic vision sensor based on a vdW vertical heterostructure for simultaneous image sensing and processing.
(A) RF with green center and pink surround areas. Left panel: DGM of the RF characterizes the distribution of responsivity. Center panel: Vision sensor and its output. An OFF-photoresponse device in the center is surrounded by ON-photoresponse devices. The output of vision sensor is the current summation over all devices. Right panel: Outputs of the artificial RF with a contrast-reversing edge moving from the left side to the right side. The upper circle array represents light sources. Light is on for solid circles and off for circles. (B) Vision sensor for simultaneous image processing and processing. (C) Edge enhancement of the letter N. Left panel: Original 8 × 8 binary image of the letter N. Middle and right panels: Simulation and experimental results.
Fig. 3Reconfigurable retinomorphic vision sensor.
(A) Demonstration of image processing with three different operations (i.e., image stylization, edge enhancement, and contrast correction). These operations are realized by controlling the photoresponse of each pixel in the sensor by varying Vg independently. (B) Image stylization. (C) Edge enhancement. (D) Contrast correction. Original images correspond to the images to be processed by different operations. Experimental results by distinct convolution operations are compared with simulations. Photo credits: C.Y. Wang, Nanjing University.
Fig. 4Implementation of convolutional neural network with the retinomorphic vision sensor.
(A) Three different patterns of each specific letter (n, j, and u). (B) Training process of the vision sensor at each epoch. The different color maps correspond to different convolution kernels. k is the number of training. i and j denote the location of each pixel in the sensor. (C) Accuracy of recognition over the epoch; the inset shows the weight distribution of vision sensor, corresponding to initial (yellow) and final training (blue). (D) Measured average output signals for each epoch for a specific input letter. The curves with largest values (f1, f2, and f3, respectively) represent the recognition results of the target letters.