scispace - formally typeset
Journal ArticleDOI

Silicon Photonics Codesign for Deep Learning

Reads0
Chats0
TLDR
The detailed analysis of a silicon photonic integrated circuit shows that a codesigned implementation based on the decomposition of large matrix-vector multiplication into smaller instances and the use of nonnegative weights could significantly simplify the photonic implementation of the matrix multiplier and allow increased scalability.
Abstract
Deep learning is revolutionizing many aspects of our society, addressing a wide variety of decision-making tasks, from image classification to autonomous vehicle control. Matrix multiplication is an essential and computationally intensive step of deep-learning calculations. The computational complexity of deep neural networks requires dedicated hardware accelerators for additional processing throughput and improved energy efficiency in order to enable scaling to larger networks in the upcoming applications. Silicon photonics is a promising platform for hardware acceleration due to recent advances in CMOS-compatible manufacturing capabilities, which enable efficient exploitation of the inherent parallelism of optics. This article provides a detailed description of recent implementations in the relatively new and promising platform of silicon photonics for deep learning. Opportunities for multiwavelength microring silicon photonic architectures codesigned with field-programmable gate array (FPGA) for pre- and postprocessing are presented. The detailed analysis of a silicon photonic integrated circuit shows that a codesigned implementation based on the decomposition of large matrix-vector multiplication into smaller instances and the use of nonnegative weights could significantly simplify the photonic implementation of the matrix multiplier and allow increased scalability. We conclude this article by presenting an overview and a detailed analysis of design parameters. Insights for ways forward are explored.

read more

Citations
More filters
Journal ArticleDOI

Efficient Processing of Deep Neural Networks

TL;DR: This book provides a structured treatment of the key principles and techniques for enabling efficient processing of deep neural networks (DNNs).
Posted Content

A Survey on Silicon Photonics for Deep Learning

TL;DR: The landscape of silicon photonics to accelerate deep learning is surveyed, with a coverage of developments across design abstractions in a bottom-up manner, to convey both the capabilities and limitations of the silicon Photonics paradigm in the context of deep learning acceleration.
Journal ArticleDOI

A Survey on Silicon Photonics for Deep Learning

TL;DR: Deep learning has led to unprecedented successes in solving some very difficult problems in domains such as computer vision, natural language processing, and general pattern recognition as mentioned in this paper. But deep learning has not yet shown great success in solving problems in other domains.
Proceedings ArticleDOI

Modeling Silicon-Photonic Neural Networks under Uncertainties

TL;DR: In this article, the impact of random uncertainties on the classification accuracy of a Mach-Zehnder Interferometer (MZI)-based silicon-photonic neural networks (SPNNs) is investigated.
Proceedings ArticleDOI

Albireo: energy-efficient acceleration of convolutional neural networks via silicon photonics

TL;DR: Albireo as mentioned in this paper proposes an analog photonic architecture for scaling DNN acceleration by characterizing photonic devices such as microring resonators and Mach-Zehnder modulators using photonic simulators.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Journal ArticleDOI

Gradient-based learning applied to document recognition

TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Journal ArticleDOI

Multilayer feedforward networks are universal approximators

TL;DR: It is rigorously established that standard multilayer feedforward networks with as few as one hidden layer using arbitrary squashing functions are capable of approximating any Borel measurable function from one finite dimensional space to another to any desired degree of accuracy, provided sufficiently many hidden units are available.
Related Papers (5)