G
Geoffrey E. Hinton
Researcher at Google
Publications - 426
Citations - 501778
Geoffrey E. Hinton is an academic researcher from Google. The author has contributed to research in topics: Artificial neural network & Generative model. The author has an hindex of 157, co-authored 414 publications receiving 409047 citations. Previous affiliations of Geoffrey E. Hinton include Canadian Institute for Advanced Research & Max Planck Society.
Papers
More filters
Journal ArticleDOI
Variational Learning for Switching State-Space Models
TL;DR: A new statistical model for time series that iteratively segments data into regimes with approximately linear dynamics and learns the parameters of each of these linear regimes is introduced and the results suggest that variational approximations are a viable method for inference and learning in switching state-space models.
Dissertation
Evaluation of gaussian processes and other methods for non-linear regression
TL;DR: It is shown that a Bayesian approach to learning in multi-layer perceptron neural networks achieves better performance than the commonly used early stopping procedure, even for reasonably short amounts of computation time.
Dissertation
Machine learning for aerial image labeling
TL;DR: It is shown how deep neural networks implemented on modern GPUs can be used to efficiently learn highly discriminative image features and two ways of improving the predictions of the system by introducing structure into the outputs of the neural networks are proposed.
Journal ArticleDOI
An efficient learning procedure for deep boltzmann machines
TL;DR: A new learning algorithm for Boltzmann machines that contain many layers of hidden variables is presented and results on the MNIST and NORB data sets are presented showing that deep BoltZmann machines learn very good generative models of handwritten digits and 3D objects.
Proceedings Article
The Recurrent Temporal Restricted Boltzmann Machine
TL;DR: The Recurrent TRBM is introduced, which is a very slight modification of the TRBM for which exact inference is very easy and exact gradient learning is almost tractable.