Photonics Research, Volume. 12, Issue 11, 2691(2024)

Scalable parallel photonic processing unit for various neural network accelerations

Shiyin Du1, Jun Zhang2, Hao Ouyang3, Zilong Tao1, Qiuquan Yan1, Hao Hao4,5, Junhu Zhou2, Jie You2,6, Yuhua Tang1, and Tian Jiang4、*
Author Affiliations
  • 1State Key Laboratory of High-Performance Computing, College of Computer, National University of Defense Technology, Changsha 410073, China
  • 2National Innovation Institute of Defense Technology, Academy of Military Sciences PLA China, Beijing 100071, China
  • 3College of Advanced Interdisciplinary Studies, National University of Defense Technology, Changsha 410073, China
  • 4Institute for Quantum Science and Technology, College of Science, National University of Defense Technology, Changsha 410073, China
  • 5e-mail: HH65637917@163.com
  • 6e-mail: jieyou1991@hotmail.com
  • show less
    Figures & Tables(9)
    Conceptual architecture diagram of SPPU on chip. (a) Basic process of SPPU inference for neural networks with different structures. (b) Schematic illustration of SPPU chip. This chip includes eight independent computational channels, with each channel consisting of an edge coupler (EC), two phase modulators (PMs), two microring resonators (MRRs), a balanced photodetector (BPD), and two grating couplers (GCs).
    Optical micrograph of SPPU. (a) Microscopic image of the SPPU chip. (b) Image of SPPU unit after photoelectric hybrid package. Microscopic images of (c) phase modulator, (d) add-drop microring resonator, and (e) photodetector.
    Characterization of key photonic components on chip. (a) Measured bandwidth of the on-chip silicon phase modulator. (b) Normalized weights and the voltage applied to the on-ring heater for one microring resonator. (c) Scatter plots of the measured weights versus calculated expectations. (d) Measured bandwidth of the on-chip photodetector. (e) Responsivity of the on-chip photodetector measured at −1 V bias. (f) Dark current measured by the on-chip photodetector at different bias voltages.
    Working flow of one SPPU channel. (a) Schematic of an SPPU channel performing optical neural network computation. (b) Sequence calculation results when the weight is one and the input data vector is [1, 0.75, 0.5, 0.25, 0, −0.25, −0.5, −0.75, −1], which is used for SPPU channel calibration. (c) Ideal (orange line) and experimental (blue curve) convolution output waveform with 4-bit coding.
    SPPU for a CNN. (a) Structure of the CNN for handwritten digits classification, which consists of two convolutional layers and a fully connected layer. Here, each convolutional layer can be subdivided into convolutional operations and pooling operations. (b) Variation in calculation accuracy and loss of validation dataset during the training. (c) Confusion matrices of digit classification. The experimental accuracy (97.9%) is in good agreement with the calculated results (98.3%).
    SPPU for a residual CNN. (a) Structure of the residual CNN for signal modulation format classification. The CNN contains a residual block and two fully connected layers, where the residual block can be subdivided into three convolutional layers. (b) Variation in calculation accuracy and loss of validation dataset during the training. (c) Confusion matrices of modulation format classification.
    SPPU for a simple RNN. (a) Structure of the simple RNN for emotion recognition of reviews. It contains an embedding layer, a recurrent layer, and a fully connected layer. Variation in calculation accuracy and loss for both (b) Amazon Fine Food Reviews dataset and (c) IMDBMovie Reviews dataset, during the training. Confusion matrices of emotion recognition for the (d) Amazon Fine Food Reviews dataset and (e) IMDB Movie Reviews dataset.
    • Table 1. Computing Paradigm of a Single SPPU Channel in Different Computing Modes

      View table
      View in Article

      Table 1. Computing Paradigm of a Single SPPU Channel in Different Computing Modes

      BiasSingle PMTwo PMs
      Δφ=2kπy=wxy=w(x1+x2)
      Δφ=2kπ+π/2y=wcos(x)y=wcos(x1+x2)
      Δφ=2kππ/2y=wcos(x)y=wcos(x1+x2)
      Δφ=(2k+1)πy=wxy=w(x1+x2)
    • Table 2. Recognition Accuracy of Neural Networks with Different Activation Functions

      View table
      View in Article

      Table 2. Recognition Accuracy of Neural Networks with Different Activation Functions

      Accuracy (%)
      FunctionMNISTRML201610aAmazonIMDB
      Calculated (cos)98.396.892.584.3
      Experiment (cos)97.992.892.383.6
      Calculated (relu)98.689.6
      Calculated (tanh)92.480.3
    Tools

    Get Citation

    Copy Citation Text

    Shiyin Du, Jun Zhang, Hao Ouyang, Zilong Tao, Qiuquan Yan, Hao Hao, Junhu Zhou, Jie You, Yuhua Tang, Tian Jiang, "Scalable parallel photonic processing unit for various neural network accelerations," Photonics Res. 12, 2691 (2024)

    Download Citation

    EndNote(RIS)BibTexPlain Text
    Save article for my favorites
    Paper Information

    Category: Silicon Photonics

    Received: Apr. 22, 2024

    Accepted: Sep. 2, 2024

    Published Online: Nov. 1, 2024

    The Author Email: Tian Jiang (tjiang@nudt.edu.cn)

    DOI:10.1364/PRJ.527940

    Topics