scispace - formally typeset
Search or ask a question
Author

Yann Ollivier

Bio: Yann Ollivier is an academic researcher from University of Paris-Sud. The author has contributed to research in topics: Ricci curvature & Random group. The author has an hindex of 19, co-authored 38 publications receiving 1780 citations. Previous affiliations of Yann Ollivier include Facebook & École normale supérieure de Lyon.

Papers
More filters
Journal ArticleDOI
TL;DR: In this article, the authors define the Ricci curvature of metric spaces in terms of how much small balls are closer (in Wasserstein transportation distance) than their centers are.

728 citations

Journal ArticleDOI
TL;DR: Ollivier et al. as discussed by the authors define a notion of Ricci curvature in metric spaces equipped with a measure or a random walk, which is compatible with Bakry-Emery theory and is robust and very easy to implement in concrete examples such as graphs.

177 citations

Proceedings ArticleDOI
01 Jan 2010
TL;DR: In this paper, the authors present a notion of Ricci curvature valid on arbitrary metric spaces, such as graphs, and generalize a series of classical theorems in positive Ricci curve, including spectral gap estimates, concentration of measure or log-Sobolev inequalities.
Abstract: This text is a presentation of the general context and results of [Oll07] and [Oll09], with comments on related work. The goal is to present a notion of Ricci curvature valid on arbitrary metric spaces, such as graphs, and to generalize a series of classical theorems in positive Ricci curvature, such as spectral gap estimates, concentration of measure or log-Sobolev inequalities. The necessary background (concentration of measure, curvature in Riemannian geometry, convergence of Markov chains) is covered in the first section. Special emphasis is put on open questions of varying difficulty.

130 citations

Journal ArticleDOI
TL;DR: In this paper, the authors provide nonasymptotic estimates for the rate of convergence of empirical means of Markov chains, together with a Gaussian or exponential control on the deviations of empirical mean.
Abstract: We provide explicit nonasymptotic estimates for the rate of convergence of empirical means of Markov chains, together with a Gaussian or exponential control on the deviations of empirical means. These estimates hold under a "positive curvature" assumption expressing a kind of metric ergodicity, which generalizes the Ricci curvature from differential geometry and, on finite graphs, amounts to contraction under path coupling.

118 citations

Journal ArticleDOI
TL;DR: In this article, the authors provide nonasymptotic estimates for the rate of convergence of empirical means of Markov chains, together with a Gaussian or exponential control on the deviations of empirical mean.
Abstract: We provide explicit nonasymptotic estimates for the rate of convergence of empirical means of Markov chains, together with a Gaussian or exponential control on the deviations of empirical means. These estimates hold under a "positive curvature" assumption expressing a kind of metric ergodicity, which generalizes the Ricci curvature from differential geometry and, on finite graphs, amounts to contraction under path coupling.

106 citations


Cited by
More filters
Christopher M. Bishop1
01 Jan 2006
TL;DR: Probability distributions of linear models for regression and classification are given in this article, along with a discussion of combining models and combining models in the context of machine learning and classification.
Abstract: Probability Distributions.- Linear Models for Regression.- Linear Models for Classification.- Neural Networks.- Kernel Methods.- Sparse Kernel Machines.- Graphical Models.- Mixture Models and EM.- Approximate Inference.- Sampling Methods.- Continuous Latent Variables.- Sequential Data.- Combining Models.

10,141 citations

01 Jan 2015
TL;DR: This compact, informal introduction for graduate students and advanced undergraduates presents the current state-of-the-art filtering and smoothing methods in a unified Bayesian framework and learns what non-linear Kalman filters and particle filters are, how they are related, and their relative advantages and disadvantages.
Abstract: Filtering and smoothing methods are used to produce an accurate estimate of the state of a time-varying system based on multiple observational inputs (data). Interest in these methods has exploded in recent years, with numerous applications emerging in fields such as navigation, aerospace engineering, telecommunications, and medicine. This compact, informal introduction for graduate students and advanced undergraduates presents the current state-of-the-art filtering and smoothing methods in a unified Bayesian framework. Readers learn what non-linear Kalman filters and particle filters are, how they are related, and their relative advantages and disadvantages. They also discover how state-of-the-art Bayesian parameter estimation methods can be combined with state-of-the-art filtering and smoothing algorithms. The book’s practical and algorithmic approach assumes only modest mathematical prerequisites. Examples include MATLAB computations, and the numerous end-of-chapter exercises include computational assignments. MATLAB/GNU Octave source code is available for download at www.cambridge.org/sarkka, promoting hands-on work with the methods.

1,102 citations

Book ChapterDOI
01 Jan 1985
TL;DR: The first group of results in fixed point theory were derived from Banach's fixed point theorem as discussed by the authors, which is a nice result since it contains only one simple condition on the map F, since it is easy to prove and since it nevertheless allows a variety of applications.
Abstract: Formally we have arrived at the middle of the book. So you may need a pause for recovering, a pause which we want to fill up by some fixed point theorems supplementing those which you already met or which you will meet in later chapters. The first group of results centres around Banach’s fixed point theorem. The latter is certainly a nice result since it contains only one simple condition on the map F, since it is so easy to prove and since it nevertheless allows a variety of applications. Therefore it is not astonishing that many mathematicians have been attracted by the question to which extent the conditions on F and the space Ω can be changed so that one still gets the existence of a unique or of at least one fixed point. The number of results produced this way is still finite, but of a statistical magnitude, suggesting at a first glance that only a random sample can be covered by a chapter or even a book of the present size. Fortunately (or unfortunately?) most of the modifications have not found applications up to now, so that there is no reason to write a cookery book about conditions but to write at least a short outline of some ideas indicating that this field can be as interesting as other chapters. A systematic account of more recent ideas and examples in fixed point theory should however be written by one of the true experts. Strange as it is, such a book does not seem to exist though so many people are puzzling out so many results.

994 citations

Posted Content
TL;DR: This paper proposes the weight-dropped LSTM which uses DropConnect on hidden-to-hidden weights as a form of recurrent regularization and introduces NT-ASGD, a variant of the averaged stochastic gradient method, wherein the averaging trigger is determined using a non-monotonic condition as opposed to being tuned by the user.
Abstract: Recurrent neural networks (RNNs), such as long short-term memory networks (LSTMs), serve as a fundamental building block for many sequence learning tasks, including machine translation, language modeling, and question answering. In this paper, we consider the specific problem of word-level language modeling and investigate strategies for regularizing and optimizing LSTM-based models. We propose the weight-dropped LSTM which uses DropConnect on hidden-to-hidden weights as a form of recurrent regularization. Further, we introduce NT-ASGD, a variant of the averaged stochastic gradient method, wherein the averaging trigger is determined using a non-monotonic condition as opposed to being tuned by the user. Using these and other regularization strategies, we achieve state-of-the-art word level perplexities on two data sets: 57.3 on Penn Treebank and 65.8 on WikiText-2. In exploring the effectiveness of a neural cache in conjunction with our proposed model, we achieve an even lower state-of-the-art perplexity of 52.8 on Penn Treebank and 52.0 on WikiText-2.

899 citations

Journal ArticleDOI
TL;DR: In this article, the authors define the Ricci curvature of metric spaces in terms of how much small balls are closer (in Wasserstein transportation distance) than their centers are.

728 citations