scispace - formally typeset
Search or ask a question
Topic

Gaussian process

About: Gaussian process is a research topic. Over the lifetime, 18944 publications have been published within this topic receiving 486645 citations. The topic is also known as: Gaussian stochastic process.


Papers
More filters
Proceedings Article
14 Jul 2011
TL;DR: In this article, the authors adopt a portfolio of acquisition functions governed by an online multi-armed bandit strategy, and propose several portfolio strategies, the best of which is called GP-Hedge, and show that this method outperforms the best individual acquisition function.
Abstract: Bayesian optimization with Gaussian processes has become an increasingly popular tool in the machine learning community. It is efficient and can be used when very little is known about the objective function, making it popular in expensive black-box optimization scenarios. It uses Bayesian methods to sample the objective efficiently using an acquisition function which incorporates the posterior estimate of the objective. However, there are several different parameterized acquisition functions in the literature, and it is often unclear which one to use. Instead of using a single acquisition function, we adopt a portfolio of acquisition functions governed by an online multi-armed bandit strategy. We propose several portfolio strategies, the best of which we call GP-Hedge, and show that this method outperforms the best individual acquisition function. We also provide a theoretical bound on the algorithm's performance.

242 citations

Journal ArticleDOI
TL;DR: This paper addresses the design of an optimal transmit signal and its corresponding optimal detector for a radar or active sonar system with a focus on the temporal aspects of the waveform with the spatial aspects to be described in a future paper.
Abstract: In this paper, we address the design of an optimal transmit signal and its corresponding optimal detector for a radar or active sonar system. The focus is on the temporal aspects of the waveform with the spatial aspects to be described in a future paper. The assumptions involved in modeling the clutter/reverberation return are crucial to the development of the optimal detector and its consequent optimal signal design. In particular, the target is assumed to be a Gaussian point target and the clutter/reverberation a stationary Gaussian random process. In practice, therefore, the modeling will need to be assessed and possibly extended, and additionally a means of measuring the "in-situ" clutter/reverberation spectrum will be required. The advantages of our approach are that a simple analytical result is obtained which is guaranteed to be optimal, and also the extension to spatial-temporal signal design is immediate using ideas of frequency-wavenumber representations. Some examples are given to illustrate the signal design procedure as well as the calculation of the increase in processing gain. Finally, the results are shown to be an extension of the usual procedure which places the signal energy in the noise band having minimum power

242 citations

Proceedings Article
01 Jan 2019
TL;DR: This work derives an analogous equivalence for multi-layer convolutional neural networks (CNNs) both with and without pooling layers, and introduces a Monte Carlo method to estimate the GP corresponding to a given neural network architecture, even in cases where the analytic form has too many terms to be computationally feasible.
Abstract: There is a previously identified equivalence between wide fully connected neural networks (FCNs) and Gaussian processes (GPs). This equivalence enables, for instance, test set predictions that would have resulted from a fully Bayesian, infinitely wide trained FCN to be computed without ever instantiating the FCN, but by instead evaluating the corresponding GP. In this work, we derive an analogous equivalence for multi-layer convolutional neural networks (CNNs) both with and without pooling layers, and achieve state of the art results on CIFAR10 for GPs without trainable kernels. We also introduce a Monte Carlo method to estimate the GP corresponding to a given neural network architecture, even in cases where the analytic form has too many terms to be computationally feasible. Surprisingly, in the absence of pooling layers, the GPs corresponding to CNNs with and without weight sharing are identical. As a consequence, translation equivariance, beneficial in finite channel CNNs trained with stochastic gradient descent (SGD), is guaranteed to play no role in the Bayesian treatment of the infinite channel limit - a qualitative difference between the two regimes that is not present in the FCN case. We confirm experimentally, that while in some scenarios the performance of SGD-trained finite CNNs approaches that of the corresponding GPs as the channel count increases, with careful tuning SGD-trained CNNs can significantly outperform their corresponding GPs, suggesting advantages from SGD training compared to fully Bayesian parameter estimation.

241 citations

Journal ArticleDOI
TL;DR: Nonparametric statistics for comparing two mean frequency functions and for combining data on recurrent events and death, together with consistent variance estimators, are developed and an application to a cancer clinical trial is provided.
Abstract: This article is concerned with the analysis of recurrent events in the presence of a terminal event such as death. We consider the mean frequency function, defined as the marginal mean of the cumulative number of recurrent events over time. A simple nonparametric estimator for this quantity is presented. It is shown that the estimator, properly normalized, converges weakly to a zero-mean Gaussian process with an easily estimable covariance function. Nonparametric statistics for comparing two mean frequency functions and for combining data on recurrent events and death are also developed. The asymptotic null distributions of these statistics, together with consistent variance estimators, are derived. The small-sample properties of the proposed estimators and test statistics are examined through simulation studies. An application to a cancer clinical trial is provided.

241 citations

Journal ArticleDOI
TL;DR: In this paper, the authors provide a precise description of the asymptotic behavior of the product limit estimator, including a simple explicit form of the covariance structure, which also turns out to be the analogue of the variance structure of the Kaplan-Meier estimator.
Abstract: Many authors have considered the problem of estimating a distribution function when the observed data is subject to random truncation. A prominent role is played by the product limit estimator, which is the analogue of the Kaplan-Meier estimator of a distribution function under random censoring. Wang and Jewell (1985) and Woodroofe (1985) independently proved consistency results for this product limit estimator and showed weak convergence to a Gaussian process. Both papers left open the exact form of the covariance structure of the limiting process. Here we provide a precise description of the asymptotic behavior of the product limit estimator, including a simple explicit form of the asymptotic covariance structure, which also turns out to be the analogue of the covariance structure of the Kaplan-Meier estimator. Some applications are briefly discussed.

241 citations


Network Information
Related Topics (5)
Estimator
97.3K papers, 2.6M citations
87% related
Optimization problem
96.4K papers, 2.1M citations
85% related
Artificial neural network
207K papers, 4.5M citations
84% related
Support vector machine
73.6K papers, 1.7M citations
82% related
Deep learning
79.8K papers, 2.1M citations
82% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
2023502
20221,181
20211,132
20201,220
20191,119
2018978