scispace - formally typeset
T

Thierry Moreau

Researcher at University of Washington

Publications -  29
Citations -  2182

Thierry Moreau is an academic researcher from University of Washington. The author has contributed to research in topics: Compiler & Deep learning. The author has an hindex of 14, co-authored 29 publications receiving 1346 citations.

Papers
More filters
Proceedings ArticleDOI

TVM: an automated end-to-end optimizing compiler for deep learning

TL;DR: TVM as discussed by the authors is a compiler that exposes graph-level and operator-level optimizations to provide performance portability to deep learning workloads across diverse hardware back-ends, such as mobile phones, embedded devices, and accelerators.
Posted Content

TVM: End-to-End Optimization Stack for Deep Learning

TL;DR: TVM is proposed, an end-to-end optimization stack that exposes graph-level and operator-level optimizations to provide performance portability to deep learning workloads across diverse hardware back-ends and discusses the optimization challenges specific toDeep learning that TVM solves.
Posted Content

Learning to Optimize Tensor Programs

TL;DR: In this article, a learning-based framework is introduced to optimize tensor programs for deep learning workloads, such as matrix multiplication and high dimensional convolution, which are key enablers of effective deep learning systems.
Proceedings ArticleDOI

SNNAP: Approximate computing on programmable SoCs via neural acceleration

TL;DR: SNNAP is designed to work with a compiler workflow that configures the neural network's topology and weights instead of the programmable logic of the FPGA itself, which enables effective use of neural acceleration in commercially available devices and accelerates different applications without costly FPGAs reconfigurations.
Posted Content

TVM: An Automated End-to-End Optimizing Compiler for Deep Learning

TL;DR: TVM is a compiler that exposes graph-level and operator-level optimizations to provide performance portability to deep learning workloads across diverse hardware back-ends and automates optimization of low-level programs to hardware characteristics by employing a novel, learning-based cost modeling method for rapid exploration of code optimizations.