scispace - formally typeset
K

Kelvin Xu

Researcher at University of California, Berkeley

Publications -  34
Citations -  17341

Kelvin Xu is an academic researcher from University of California, Berkeley. The author has contributed to research in topics: Reinforcement learning & Task (project management). The author has an hindex of 16, co-authored 28 publications receiving 15203 citations. Previous affiliations of Kelvin Xu include University of Toronto & Université de Montréal.

Papers
More filters
Proceedings Article

Show, Attend and Tell: Neural Image Caption Generation with Visual Attention

TL;DR: An attention based model that automatically learns to describe the content of images is introduced that can be trained in a deterministic manner using standard backpropagation techniques and stochastically by maximizing a variational lower bound.
Posted Content

Show, Attend and Tell: Neural Image Caption Generation with Visual Attention

TL;DR: This paper proposed an attention-based model that automatically learns to describe the content of images by focusing on salient objects while generating corresponding words in the output sequence, which achieved state-of-the-art performance on three benchmark datasets: Flickr8k, Flickr30k and MS COCO.
Posted Content

Theano: A Python framework for fast computation of mathematical expressions

Rami Al-Rfou, +111 more
TL;DR: The performance of Theano is compared against Torch7 and TensorFlow on several machine learning models and recently-introduced functionalities and improvements are discussed.
Journal Article

On using monolingual corpora in neural machine translation

TL;DR: This work investigates how to leverage abundant monolingual corpora for neural machine translation to improve results for En-Fr and En-De translation and extends to high resource languages such as Cs-En and De-En.
Proceedings Article

Bridging the Gap Between Value and Policy Based Reinforcement Learning

TL;DR: A new RL algorithm, Path Consistency Learning (PCL), is developed that minimizes a notion of soft consistency error along multi-step action sequences extracted from both on- and off-policy traces and significantly outperforms strong actor-critic and Q-learning baselines across several benchmarks.