scispace - formally typeset
A

Ashish Vaswani

Researcher at Google

Publications -  73
Citations -  70493

Ashish Vaswani is an academic researcher from Google. The author has contributed to research in topics: Machine translation & Transformer (machine learning model). The author has an hindex of 34, co-authored 70 publications receiving 35599 citations. Previous affiliations of Ashish Vaswani include Information Sciences Institute & University of Southern California.

Papers
More filters
Proceedings Article

Attention is All you Need

TL;DR: This paper proposed a simple network architecture based solely on an attention mechanism, dispensing with recurrence and convolutions entirely and achieved state-of-the-art performance on English-to-French translation.
Posted Content

Attention Is All You Need

TL;DR: A new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely is proposed, which generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.
Proceedings ArticleDOI

Self-Attention with Relative Position Representations

TL;DR: This article extended the self-attention mechanism to consider representations of the relative positions, or distances between sequence elements, and showed that relative and absolute position representations yields no further improvement in translation quality.
Posted Content

Image Transformer

TL;DR: In this article, a self-attention mechanism is used to attend to local neighborhoods to increase the size of images generated by the model, despite maintaining significantly larger receptive fields per layer than typical CNNs.