SqueezeLight: A Multi-Operand Ring-Based Optical Neural Network With Cross-Layer Scalability

Optical neural networks (ONNs) are promising hardware platforms for next-generation artificial intelligence acceleration with ultrafast speed and low-energy consumption. However, previous ONN designs are bounded by one multiply–accumulate operation per device, showing unsatisfying scalability. In this work, we propose a scalable ONN architecture, dubbed SqueezeLight. We propose a nonlinear optical neuron based on multioperand ring resonators (MORRs) to squeeze vector dot-product into a single device with low wavelength usage and built-in nonlinearity. A block-level squeezing technique with structured sparsity is exploited to support higher scalability. We adopt a robustness-aware training algorithm to guarantee variation tolerance. To enable a truly scalable ONN architecture, we extend SqueezeLight to a separable optical CNN architecture that further squeezes in the layer level. Two orthogonal convolutional layers are mapped to one MORR array, leading to order-of-magnitude higher software training scalability. We further explore augmented representability for SqueezeLight by introducing parametric MORR neurons with trainable nonlinearity, together with a nonlinearity-aware initialization method to stabilize convergence. Experimental results show that SqueezeLight achieves one-order-of-magnitude better compactness and efficiency than previous designs with high fidelity, trainability, and robustness. Our open-source codes are available at https://github.com/JeremieMelo/SqueezeLight.

[1]  Febin P. Sunny,et al.  Silicon Photonic Microring Resonators: A Comprehensive Design-Space Exploration and Optimization Under Fabrication-Process Variations , 2022, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[2]  A. Tait Quantifying power use in silicon photonic neural networks , 2021, Physical Review Applied.

[3]  David Z. Pan,et al.  L2ight: Enabling On-Chip Learning for Optical Neural Networks via Efficient in-situ Subspace Optimization , 2021, NeurIPS.

[4]  F. Yaman,et al.  A silicon photonic–electronic neural network for fibre nonlinearity compensation , 2021, Nature Electronics.

[5]  Ray T. Chen,et al.  Toward Hardware-Efficient Optical Neural Networks: Beyond FFT Architecture via Joint Learnability , 2021, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[6]  Mahdi Nikdast,et al.  CrossLight: A Cross-Layer Optimized Silicon Photonic Neural Network Accelerator , 2021, 2021 58th ACM/IEEE Design Automation Conference (DAC).

[7]  Ray T. Chen,et al.  SqueezeLight: Towards Scalable Optical Neural Networks with Multi-Operand Ring Resonators , 2021, 2021 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[8]  David Z. Pan,et al.  Silicon photonic subspace neural chip for hardware-efficient deep learning , 2021, ArXiv.

[9]  Huaxi Gu,et al.  Countering Variations and Thermal Effects for Accurate Optical Neural Networks , 2020, 2020 IEEE/ACM International Conference On Computer Aided Design (ICCAD).

[10]  Bhavin J. Shastri,et al.  Photonics for artificial intelligence and neuromorphic computing , 2020, Nature Photonics.

[11]  Jiaqi Gu,et al.  FLOPS: EFficient On-Chip Learning for OPtical Neural Networks Through Stochastic Zeroth-Order Optimization , 2020, 2020 57th ACM/IEEE Design Automation Conference (DAC).

[12]  Weichen Liu,et al.  LightBulb: A Photonic-Nonvolatile-Memory-based Accelerator for Binarized Convolutional Neural Networks , 2020, 2020 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[13]  Jiaqi Gu,et al.  ROQ: A Noise-Aware Quantization Scheme Towards Robust Optical Neural Networks with Low-bit Controls , 2020, 2020 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[14]  Pengfei Xu,et al.  DNN-Chip Predictor: An Analytical Performance Predictor for DNN Accelerators with Various Dataflows and Hardware Architectures , 2020, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[15]  Luca P. Carloni,et al.  Silicon Photonics Codesign for Deep Learning , 2020, Proceedings of the IEEE.

[16]  Mario Miscuglio,et al.  Photonic tensor cores for machine learning , 2020, Applied Physics Reviews.

[17]  Jiaqi Gu,et al.  Towards Area-Efficient Optical Neural Networks: An FFT-based Architecture , 2020, 2020 25th Asia and South Pacific Design Automation Conference (ASP-DAC).

[18]  Paul R. Prucnal,et al.  Photonic Multiply-Accumulate Operations for Neural Networks , 2020, IEEE Journal of Selected Topics in Quantum Electronics.

[19]  Lu Yuan,et al.  Dynamic Convolution: Attention Over Convolution Kernels , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Jiaqi Gu,et al.  Design Technology for Scalable and Robust Photonic Integrated Circuits: Invited Paper , 2019, 2019 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).

[21]  Richard A. Soref,et al.  Integrated multi-operand electro-optic logic gates for optical computing , 2019, Applied Physics Letters.

[22]  Weichen Liu,et al.  HolyLight: A Nanophotonic Accelerator for Deep Learning in Data Centers , 2019, 2019 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[23]  Andrea Melloni,et al.  Canceling Thermal Cross-Talk Effects in Photonic Integrated Circuits , 2019, Journal of Lightwave Technology.

[24]  Derong Liu,et al.  Hardware-software co-design of slimmed optical neural networks , 2019, ASP-DAC.

[25]  H. Rong,et al.  A 128 Gb/s PAM4 Silicon Microring Modulator With Integrated Thermo-Optic Resonance Tuning , 2019, Journal of Lightwave Technology.

[26]  Le Song,et al.  Decoupled Networks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[27]  Michael R. Watts,et al.  AIM Process Design Kit (AIMPDKv2.0): Silicon Photonics Passive and Active Component Libraries on a 300mm Wafer , 2018, 2018 Optical Fiber Communications Conference and Exposition (OFC).

[28]  Chao Wang,et al.  CirCNN: Accelerating and Compressing Deep Neural Networks Using Block-Circulant Weight Matrices , 2017, 2017 50th Annual IEEE/ACM International Symposium on Microarchitecture (MICRO).

[29]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[30]  P. Prucnal,et al.  Neuromorphic photonic networks using silicon photonic weight banks , 2017, Scientific Reports.

[31]  Dirk Englund,et al.  Deep learning with coherent nanophotonic circuits , 2017, 2017 Fifth Berkeley Symposium on Energy Efficient Electronic Systems & Steep Transistors Workshop (E3S).

[32]  V. Sze,et al.  Eyeriss: An Energy-Efficient Reconfigurable Accelerator for Deep Convolutional Neural Networks , 2016, IEEE Journal of Solid-State Circuits.

[33]  William Shieh,et al.  End-to-End Energy Modeling and Analysis of Long-Haul Coherent Transmission Systems , 2014, Journal of Lightwave Technology.

[34]  Cordelia Schmid,et al.  Convolutional Kernel Networks , 2014, NIPS.

[35]  Y Fainman,et al.  Towards 100 channel dense wavelength division multiplexing with 100GHz spacing on silicon. , 2014, Optics express.

[36]  N. Harris,et al.  Efficient, compact and low loss thermo-optic phase shifter in silicon. , 2014, Optics express.

[37]  Wenqing Wu,et al.  Cross-layer racetrack memory design for ultra high density and low power consumption , 2013, 2013 50th ACM/EDAC/IEEE Design Automation Conference (DAC).

[38]  Andrew Y. Ng,et al.  Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .

[39]  Xiang Zhou,et al.  Ultra-High-Capacity DWDM transmission system for 100G and beyond , 2010, IEEE Communications Magazine.

[40]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[41]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .