T
Thierry Moreau
Researcher at University of Washington
Publications - 29
Citations - 2182
Thierry Moreau is an academic researcher from University of Washington. The author has contributed to research in topics: Compiler & Deep learning. The author has an hindex of 14, co-authored 29 publications receiving 1346 citations.
Papers
More filters
Proceedings ArticleDOI
TVM: an automated end-to-end optimizing compiler for deep learning
Tianqi Chen,Thierry Moreau,Ziheng Jiang,Lianmin Zheng,Eddie Yan,Meghan Cowan,Haichen Shen,Leyuan Wang,Yuwei Hu,Luis Ceze,Carlos Guestrin,Arvind Krishnamurthy +11 more
TL;DR: TVM as discussed by the authors is a compiler that exposes graph-level and operator-level optimizations to provide performance portability to deep learning workloads across diverse hardware back-ends, such as mobile phones, embedded devices, and accelerators.
Posted Content
TVM: End-to-End Optimization Stack for Deep Learning
Tianqi Chen,Thierry Moreau,Ziheng Jiang,Haichen Shen,Eddie Yan,Leyuan Wang,Yuwei Hu,Luis Ceze,Carlos Guestrin,Arvind Krishnamurthy +9 more
TL;DR: TVM is proposed, an end-to-end optimization stack that exposes graph-level and operator-level optimizations to provide performance portability to deep learning workloads across diverse hardware back-ends and discusses the optimization challenges specific toDeep learning that TVM solves.
Posted Content
Learning to Optimize Tensor Programs
Tianqi Chen,Lianmin Zheng,Eddie Yan,Ziheng Jiang,Thierry Moreau,Luis Ceze,Carlos Guestrin,Arvind Krishnamurthy +7 more
TL;DR: In this article, a learning-based framework is introduced to optimize tensor programs for deep learning workloads, such as matrix multiplication and high dimensional convolution, which are key enablers of effective deep learning systems.
Proceedings ArticleDOI
SNNAP: Approximate computing on programmable SoCs via neural acceleration
TL;DR: SNNAP is designed to work with a compiler workflow that configures the neural network's topology and weights instead of the programmable logic of the FPGA itself, which enables effective use of neural acceleration in commercially available devices and accelerates different applications without costly FPGAs reconfigurations.
Posted Content
TVM: An Automated End-to-End Optimizing Compiler for Deep Learning
Tianqi Chen,Thierry Moreau,Ziheng Jiang,Lianmin Zheng,Eddie Yan,Meghan Cowan,Haichen Shen,Leyuan Wang,Yuwei Hu,Luis Ceze,Carlos Guestrin,Arvind Krishnamurthy +11 more
TL;DR: TVM is a compiler that exposes graph-level and operator-level optimizations to provide performance portability to deep learning workloads across diverse hardware back-ends and automates optimization of low-level programs to hardware characteristics by employing a novel, learning-based cost modeling method for rapid exploration of code optimizations.