Journal ArticleDOI
Silicon Photonics Codesign for Deep Learning
Qixiang Cheng,Jihye Kwon,Madeleine Glick,Meisam Bahadori,Luca P. Carloni,Keren Bergman +5 more
- Vol. 108, Iss: 8, pp 1261-1282
Reads0
Chats0
TLDR
The detailed analysis of a silicon photonic integrated circuit shows that a codesigned implementation based on the decomposition of large matrix-vector multiplication into smaller instances and the use of nonnegative weights could significantly simplify the photonic implementation of the matrix multiplier and allow increased scalability.Abstract:
Deep learning is revolutionizing many aspects of our society, addressing a wide variety of decision-making tasks, from image classification to autonomous vehicle control. Matrix multiplication is an essential and computationally intensive step of deep-learning calculations. The computational complexity of deep neural networks requires dedicated hardware accelerators for additional processing throughput and improved energy efficiency in order to enable scaling to larger networks in the upcoming applications. Silicon photonics is a promising platform for hardware acceleration due to recent advances in CMOS-compatible manufacturing capabilities, which enable efficient exploitation of the inherent parallelism of optics. This article provides a detailed description of recent implementations in the relatively new and promising platform of silicon photonics for deep learning. Opportunities for multiwavelength microring silicon photonic architectures codesigned with field-programmable gate array (FPGA) for pre- and postprocessing are presented. The detailed analysis of a silicon photonic integrated circuit shows that a codesigned implementation based on the decomposition of large matrix-vector multiplication into smaller instances and the use of nonnegative weights could significantly simplify the photonic implementation of the matrix multiplier and allow increased scalability. We conclude this article by presenting an overview and a detailed analysis of design parameters. Insights for ways forward are explored.read more
Citations
More filters
Journal ArticleDOI
Efficient Processing of Deep Neural Networks
TL;DR: This book provides a structured treatment of the key principles and techniques for enabling efficient processing of deep neural networks (DNNs).
Posted Content
A Survey on Silicon Photonics for Deep Learning
TL;DR: The landscape of silicon photonics to accelerate deep learning is surveyed, with a coverage of developments across design abstractions in a bottom-up manner, to convey both the capabilities and limitations of the silicon Photonics paradigm in the context of deep learning acceleration.
Journal ArticleDOI
A Survey on Silicon Photonics for Deep Learning
TL;DR: Deep learning has led to unprecedented successes in solving some very difficult problems in domains such as computer vision, natural language processing, and general pattern recognition as mentioned in this paper. But deep learning has not yet shown great success in solving problems in other domains.
Proceedings ArticleDOI
Modeling Silicon-Photonic Neural Networks under Uncertainties
TL;DR: In this article, the impact of random uncertainties on the classification accuracy of a Mach-Zehnder Interferometer (MZI)-based silicon-photonic neural networks (SPNNs) is investigated.
Proceedings ArticleDOI
Albireo: energy-efficient acceleration of convolutional neural networks via silicon photonics
TL;DR: Albireo as mentioned in this paper proposes an analog photonic architecture for scaling DNN acceleration by characterizing photonic devices such as microring resonators and Mach-Zehnder modulators using photonic simulators.
References
More filters
Proceedings Article
Adam: A Method for Stochastic Optimization
Diederik P. Kingma,Jimmy Ba +1 more
TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article
Very Deep Convolutional Networks for Large-Scale Image Recognition
Karen Simonyan,Andrew Zisserman +1 more
TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings Article
Very Deep Convolutional Networks for Large-Scale Image Recognition
Karen Simonyan,Andrew Zisserman +1 more
TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.
Journal ArticleDOI
Gradient-based learning applied to document recognition
Yann LeCun,Léon Bottou,Léon Bottou,Yoshua Bengio,Yoshua Bengio,Yoshua Bengio,Patrick Haffner +6 more
TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Journal ArticleDOI
Multilayer feedforward networks are universal approximators
TL;DR: It is rigorously established that standard multilayer feedforward networks with as few as one hidden layer using arbitrary squashing functions are capable of approximating any Borel measurable function from one finite dimensional space to another to any desired degree of accuracy, provided sufficiently many hidden units are available.