| Literature DB >> 35379848 |
Angelina Totovic1, George Giamougiannis2, Apostolos Tsakyridis2, David Lazovsky3, Nikos Pleros2.
Abstract
Neuromorphic photonics has relied so far either solely on coherent or Wavelength-Division-Multiplexing (WDM) designs for enabling dot-product or vector-by-matrix multiplication, which has led to an impressive variety of architectures. Here, we go a step further and employ WDM for enriching the layout with parallelization capabilities across fan-in and/or weighting stages instead of serving the computational purpose and present, for the first time, a neuron architecture that combines coherent optics with WDM towards a multifunctional programmable neural network platform. Our reconfigurable platform accommodates four different operational modes over the same photonic hardware, supporting multi-layer, convolutional, fully-connected and power-saving layers. We validate mathematically the successful performance along all four operational modes, taking into account crosstalk, channel spacing and spectral dependence of the critical optical elements, concluding to a reliable operation with MAC relative error [Formula: see text].Entities:
Year: 2022 PMID: 35379848 PMCID: PMC8980092 DOI: 10.1038/s41598-022-09370-y
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1(a) Schematic representation of PPNN showing M laser diodes (LDs), a MUX, a 3dB X-splitter followed by a bias branch () and a reconfigurable OLAU encompassing 1-to-N splitting stage, input () and weight () modulator banks and an N-to-1 combiner stage, the output of which is brought to interfere with the bias signal within 3dB X-coupler and sent to the DEMUX. Closer look into (b) 1-to-N splitting and (d) its -rotated N-to-1 coupling stage. Zoom-in into the (c) bias branch wavelength selective weights and phase modulators and (e) an axon of the OLAU consisting of switches for signal routing and modulators for inputs () and weights ().
PPNN modes of operation and the corresponding switch states.
| Mode | ||||
|---|---|---|---|---|
| #1 | Multi-neuron | 1 (up) | 1 (bar) | 1 (up) |
| #2 | Convolutional | 1 (up) | 0 (cross) | 0 (down) |
| #3 | Fully-connected | 0 (down) | 0 (cross) | 1 (up) |
| #4 | Power-saving | 0 (down) | 1 (bar) | 0 (down) |
Figure 2(a) Simplified CNN inspired by LeNet-5, employed in image classification. (b) Schematic of a convolutional layer with color coded input/output pairs and (c) its implementation over PPNN in mode #2 where each channel m corresponds to one input/output pair.
Figure 3(b) Schematic of an autoencoder and (a), (c) its two FC layers implemented over PPNN in mode #3 where channels correspond to unique weight vectors and outputs . Based on the connectivity graph from (b), the implementation assumes the use of (a) 4 branches and 2 wavelengths in the first layer and (c) 2 branches and 4 wavelengths in the second one. If the number of available branches N is greater than needed, all the excess branches will have the inputs set to 0 (observe the Nth branch in (a), (c), where the condition and is imposed, respectively). Index n in the implementation (a) is set to to denote that the lit nth branch carries a non-zero input. Similarly, if the number of available wavelengths M exceeds the number of required ones, the excess LDs are powered off.
Input and weight matrices of the nth axon.
| Mode | ||
|---|---|---|
| #1 | diag | diag |
| #2 | diag | |
| #3 | diag | |
| #4 |
Figure 4Comparison between the convolutional (#2, left-hand-side) and the fully-connected (#3, right-hand-side) mode of PPNN operation with channels, optimized for operation at channel , and axons for and . Channel-wise color coded 2-D scatter plots of the targeted matrix element and (a), (b) the magnitude and (c), (d) the argument of the experimental matrix element and (e), (f) the algebraic magnitude of the absolute deviation of the experimental from targeted matrix element, , with , all with displayed univariate kernel probability density plots on the corresponding horizontal and vertical axes of the scatter plots.
Figure 5Mean relative errors of the matrix element (given in percent) with to confidence bounds for (a), (b) multi-neuron, (c), (d) convolutional, and (e), (f) FC mode of operation, depending on (a), (c), (e) channel spacing for and (b), (d), (f) AWG crosstalk for .
Figure 6Relative error 5–95% confidence interval (given in %) versus the neuron fan-in N at and for (a) convolutional and (b) fully-connected mode.