Y
Yoshua Bengio
Researcher at Université de Montréal
Publications - 1146
Citations - 534376
Yoshua Bengio is an academic researcher from Université de Montréal. The author has contributed to research in topics: Artificial neural network & Deep learning. The author has an hindex of 202, co-authored 1033 publications receiving 420313 citations. Previous affiliations of Yoshua Bengio include McGill University & Centre de Recherches Mathématiques.
Papers
More filters
Proceedings ArticleDOI
HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering
Zhilin Yang,Peng Qi,Saizheng Zhang,Yoshua Bengio,William W. Cohen,Ruslan Salakhutdinov,Christopher D. Manning +6 more
TL;DR: HotpotQA as discussed by the authors is a dataset with 113k Wikipedia-based question-answer pairs with four key features: finding and reasoning over multiple supporting documents to answer; the questions are diverse and not constrained to any pre-existing knowledge bases or knowledge schemas; providing sentence-level supporting facts required for reasoning; and offering a new type of factoid comparison questions to test QA systems' ability to extract relevant facts and perform necessary comparison.
Posted Content
Deep Graph Infomax.
TL;DR: Deep Graph Infomax (DGI) is presented, a general approach for learning node representations within graph-structured data in an unsupervised manner that is readily applicable to both transductive and inductive learning setups.
Proceedings Article
Mutual Information Neural Estimation.
Mohamed Ishmael Belghazi,Aristide Baratin,Sai Rajeshwar,Sherjil Ozair,Yoshua Bengio,Aaron Courville,Devon Hjelm +6 more
TL;DR: A Mutual Information Neural Estimator (MINE) is presented that is linearly scalable in dimensionality as well as in sample size, trainable through back-prop, and strongly consistent, and applied to improve adversarially trained generative models.
Posted Content
A Recurrent Latent Variable Model for Sequential Data
TL;DR: In this article, the authors explore the use of latent random variables into the dynamic hidden state of a recurrent neural network (RNN) by combining elements of the variational autoencoder.
Journal ArticleDOI
Representational power of restricted boltzmann machines and deep belief networks
Nicolas Le Roux,Yoshua Bengio +1 more
TL;DR: This work proves that adding hidden units yields strictly improved modeling power, while a second theorem shows that RBMs are universal approximators of discrete distributions and suggests a new and less greedy criterion for training RBMs within DBNs.