scispace - formally typeset
Search or ask a question

Showing papers on "Upper and lower bounds published in 2015"


Proceedings Article
06 Jul 2015
TL;DR: This work introduces a new, efficient, principled and backpropagation-compatible algorithm for learning a probability distribution on the weights of a neural network, called Bayes by Backprop, and shows how the learnt uncertainty in the weights can be used to improve generalisation in non-linear regression problems.
Abstract: We introduce a new, efficient, principled and backpropagation-compatible algorithm for learning a probability distribution on the weights of a neural network, called Bayes by Backprop. It regularises the weights by minimising a compression cost, known as the variational free energy or the expected lower bound on the marginal likelihood. We show that this principled kind of regularisation yields comparable performance to dropout on MNIST classification. We then demonstrate how the learnt uncertainty in the weights can be used to improve generalisation in non-linear regression problems, and how this weight uncertainty can be used to drive the exploration-exploitation trade-off in reinforcement learning.

1,287 citations


Book ChapterDOI
31 Dec 2015

1,232 citations


Journal ArticleDOI
TL;DR: In this article, the volume-weighted neutral hydrogen fraction was derived from spectroscopy of bright quasars, and it was shown that the fraction of the Lyman alpha and Lyman-beta forests that is "dark" (with zero flux) provides the only model-independent upper limit on, requiring no assumptions about the physical conditions in the intergalactic medium or the quasar's unabsorbed UV continuum.
Abstract: We present new upper limits on the volume-weighted neutral hydrogen fraction, , at z~5-6 derived from spectroscopy of bright quasars. The fraction of the Lyman-alpha and Lyman-beta forests that is "dark" (with zero flux) provides the only model-independent upper limit on , requiring no assumptions about the physical conditions in the intergalactic medium or the quasar's unabsorbed UV continuum. In this work we update our previous results using a larger sample (22 objects) of medium-depth (~ few hours) spectra of high-redshift quasars obtained with the Magellan, MMT, and VLT. This significantly improves the upper bound on derived from dark pixel analysis to <= 0.04 + 0.05 (1{\sigma}) at z=5.6. These results provide robust constraints for theoretical models of reionization, and provide the strongest available evidence that reionization has completed (or is very nearly complete) by z~6.

410 citations


Journal ArticleDOI
TL;DR: This work identifies a “canonical sufficient condition” for an upper bound on the price of anarchy of pure Nash equilibria, and proves that in congestion games, smoothness arguments are “complete” in a proof-theoretic sense: despite their automatic generality, they are guaranteed to produce optimal worst-case upper bounds on thePrice of anarchy.
Abstract: The price of anarchy, defined as the ratio of the worst-case objective function value of a Nash equilibrium of a game and that of an optimal outcome, quantifies the inefficiency of selfish behavior. Remarkably good bounds on this measure are known for a wide range of application domains. However, such bounds are meaningful only if a game's participants successfully reach a Nash equilibrium. This drawback motivates inefficiency bounds that apply more generally to weaker notions of equilibria, such as mixed Nash equilibria and correlated equilibria, and to sequences of outcomes generated by natural experimentation strategies, such as successive best responses and simultaneous regret-minimization.We establish a general and fundamental connection between the price of anarchy and its seemingly more general relatives. First, we identify a “canonical sufficient condition” for an upper bound on the price of anarchy of pure Nash equilibria, which we call a smoothness argument. Second, we prove an “extension theorem”: every bound on the price of anarchy that is derived via a smoothness argument extends automatically, with no quantitative degradation in the bound, to mixed Nash equilibria, correlated equilibria, and the average objective function value of every outcome sequence generated by no-regret learners. Smoothness arguments also have automatic implications for the inefficiency of approximate equilibria, for bicriteria bounds, and, under additional assumptions, for polynomial-length best-response sequences. Third, we prove that in congestion games, smoothness arguments are “complete” in a proof-theoretic sense: despite their automatic generality, they are guaranteed to produce optimal worst-case upper bounds on the price of anarchy.

242 citations


Proceedings ArticleDOI
TL;DR: An algorithm for sparse Hamiltonian simulation whose complexity is optimal (up to log factors) as a function of all parameters of interest is presented, and a new lower bound is proved showing that no algorithm can have sub linear dependence on tau.
Abstract: We present an algorithm for sparse Hamiltonian simulation whose complexity is optimal (up to log factors) as a function of all parameters of interest. Previous algorithms had optimal or near-optimal scaling in some parameters at the cost of poor scaling in others. Hamiltonian simulation via a quantum walk has optimal dependence on the sparsity at the expense of poor scaling in the allowed error. In contrast, an approach based on fractional-query simulation provides optimal scaling in the error at the expense of poor scaling in the sparsity. Here we combine the two approaches, achieving the best features of both. By implementing a linear combination of quantum walk steps with coefficients given by Bessel functions, our algorithm's complexity (as measured by the number of queries and 2-qubit gates) is logarithmic in the inverse error, and nearly linear in the product $\tau$ of the evolution time, the sparsity, and the magnitude of the largest entry of the Hamiltonian. Our dependence on the error is optimal, and we prove a new lower bound showing that no algorithm can have sublinear dependence on $\tau$.

236 citations


Proceedings ArticleDOI
17 Oct 2015
TL;DR: In this article, the authors present an algorithm for sparse Hamiltonian simulation whose complexity is optimal (up to log factors) as a function of all parameters of interest, and prove that no algorithm can have sub linear dependence on tau.
Abstract: We present an algorithm for sparse Hamiltonian simulation whose complexity is optimal (up to log factors) as a function of all parameters of interest. Previous algorithms had optimal or near-optimal scaling in some parameters at the cost of poor scaling in others. Hamiltonian simulation via a quantum walk has optimal dependence on the sparsity at the expense of poor scaling in the allowed error. In contrast, an approach based on fractional-query simulation provides optimal scaling in the error at the expense of poor scaling in the sparsity. Here we combine the two approaches, achieving the best features of both. By implementing a linear combination of quantum walk steps with coefficients given by Bessel functions, our algorithm's complexity (as measured by the number of queries and 2-qubit gates) is logarithmic in the inverse error, and nearly linear in the product tau of the evolution time, the sparsity, and the magnitude of the largest entry of the Hamiltonian. Our dependence on the error is optimal, and we prove a new lower bound showing that no algorithm can have sub linear dependence on tau.

226 citations


Posted Content
TL;DR: In this article, the authors provide a new proof of the linear convergence of the alternating direction method of multipliers (ADMM) when one of the objective terms is strongly convex, based on a framework for analyzing optimization algorithms introduced in Lessard et al.
Abstract: We provide a new proof of the linear convergence of the alternating direction method of multipliers (ADMM) when one of the objective terms is strongly convex. Our proof is based on a framework for analyzing optimization algorithms introduced in Lessard et al. (2014), reducing algorithm convergence to verifying the stability of a dynamical system. This approach generalizes a number of existing results and obviates any assumptions about specific choices of algorithm parameters. On a numerical example, we demonstrate that minimizing the derived bound on the convergence rate provides a practical approach to selecting algorithm parameters for particular ADMM instances. We complement our upper bound by constructing a nearly-matching lower bound on the worst-case rate of convergence.

195 citations


Posted Content
TL;DR: It is shown that the only way to achieve a sufficiently good constant approximation of the optimal revenue is through a detailed understanding of bidders' valuation distributions, and introduces α-strongly regular distributions, which interpolate between the well-studied classes of regular and MHR distributions.
Abstract: In the design and analysis of revenue-maximizing auctions, auction performance is typically measured with respect to a prior distribution over inputs. The most obvious source for such a distribution is past data. The goal is to understand how much data is necessary and sufficient to guarantee near-optimal expected revenue. Our basic model is a single-item auction in which bidders' valuations are drawn independently from unknown and non-identical distributions. The seller is given $m$ samples from each of these distributions "for free" and chooses an auction to run on a fresh sample. How large does m need to be, as a function of the number k of bidders and eps > 0, so that a (1 - eps)-approximation of the optimal revenue is achievable? We prove that, under standard tail conditions on the underlying distributions, m = poly(k, 1/eps) samples are necessary and sufficient. Our lower bound stands in contrast to many recent results on simple and prior-independent auctions and fundamentally involves the interplay between bidder competition, non-identical distributions, and a very close (but still constant) approximation of the optimal revenue. It effectively shows that the only way to achieve a sufficiently good constant approximation of the optimal revenue is through a detailed understanding of bidders' valuation distributions. Our upper bound is constructive and applies in particular to a variant of the empirical Myerson auction, the natural auction that runs the revenue-maximizing auction with respect to the empirical distributions of the samples. Our sample complexity lower bound depends on the set of allowable distributions, and to capture this we introduce alpha-strongly regular distributions, which interpolate between the well-studied classes of regular (alpha = 0) and MHR (alpha = 1) distributions. We give evidence that this definition is of independent interest.

193 citations


Posted Content
TL;DR: In this article, the authors show that kernel-based quadrature rules for computing integrals can be seen as a special case of random feature expansions for positive definite kernels, for a particular decomposition that always exists for such kernels.
Abstract: We show that kernel-based quadrature rules for computing integrals can be seen as a special case of random feature expansions for positive definite kernels, for a particular decomposition that always exists for such kernels. We provide a theoretical analysis of the number of required samples for a given approximation error, leading to both upper and lower bounds that are based solely on the eigenvalues of the associated integral operator and match up to logarithmic terms. In particular, we show that the upper bound may be obtained from independent and identically distributed samples from a specific non-uniform distribution, while the lower bound if valid for any set of points. Applying our results to kernel-based quadrature, while our results are fairly general, we recover known upper and lower bounds for the special cases of Sobolev spaces. Moreover, our results extend to the more general problem of full function approximations (beyond simply computing an integral), with results in L2- and L$\infty$-norm that match known results for special cases. Applying our results to random features, we show an improvement of the number of random features needed to preserve the generalization guarantees for learning with Lipschitz-continuous losses.

184 citations


Journal ArticleDOI
TL;DR: It is shown through numerical examples that the stability criterion can provide a larger admissible maximum upper bound than stability criteria using a Jensen-type inequality approach and a free-weighting matrix approach.

181 citations


Proceedings Article
07 Dec 2015
TL;DR: This work shows the existence of a Locality-Sensitive Hashing (LSH) family for the angular distance that yields an approximate Near Neighbor Search algorithm with the asymptotically optimal running time exponent and establishes a fine-grained lower bound for the quality of any LSH family for angular distance.
Abstract: We show the existence of a Locality-Sensitive Hashing (LSH) family for the angular distance that yields an approximate Near Neighbor Search algorithm with the asymptotically optimal running time exponent. Unlike earlier algorithms with this property (e.g., Spherical LSH [1, 2]), our algorithm is also practical, improving upon the well-studied hyperplane LSH [3] in practice. We also introduce a multiprobe version of this algorithm and conduct an experimental evaluation on real and synthetic data sets. We complement the above positive results with a fine-grained lower bound for the quality of any LSH family for angular distance. Our lower bound implies that the above LSH family exhibits a trade-off between evaluation time and quality that is close to optimal for a natural class of LSH functions.

Journal ArticleDOI
TL;DR: In this article, a nearly optimal algorithm for denoising a mixture of sinusoids from noisy equispaced samples was derived by viewing line spectral estimation as a sparse recovery problem with a continuous, infinite dictionary.
Abstract: This paper establishes a nearly optimal algorithm for denoising a mixture of sinusoids from noisy equispaced samples. We derive our algorithm by viewing line spectral estimation as a sparse recovery problem with a continuous, infinite dictionary. We show how to compute the estimator via semidefinite programming and provide guarantees on its mean-squared error rate. We derive a complementary minimax lower bound on this estimation rate, demonstrating that our approach nearly achieves the best possible estimation error. Furthermore, we establish bounds on how well our estimator localizes the frequencies in the signal, showing that the localization error tends to zero as the number of samples grows. We verify our theoretical results in an array of numerical experiments, demonstrating that the semidefinite programming approach outperforms three classical spectral estimation techniques.

Proceedings ArticleDOI
17 Oct 2015
TL;DR: In this article, the sample complexity of learning threshold functions with differential privacy was shown to be at most 2(1+o(1)) log * |X| samples, which grows with the size of the domain.
Abstract: We prove new upper and lower bounds on the sample complexity of (a#x03B5;, a#x03B4;) differentially private algorithms for releasing approximate answers to threshold functions. A threshold function cx over a totally ordered domain X evaluates to cx(y) = 1 if y a#x2264; x, and evaluates to 0 otherwise. We give the first nontrivial lower bound for releasing thresholds with (a#x03B5;, a#x03B4;) differential privacy, showing that the task is impossible over an infinite domain X, and moreover requires sample complexity n a#x2265; (log* |X|), which grows with the size of the domain. Inspired by the techniques used to prove this lower bound, we give an algorithm for releasing thresholds with n a#x2264; 2(1+o(1)) log* |X| samples. This improves the previous best upper bound of 8(1+o(1)) log* |X| (Beimel et al., RANDOM'13). Our sample complexity upper and lower bounds also apply to the tasks of learning distributions with respect to Kolmogorov distance and of properly PAC learning thresholds with differential privacy. The lower bound gives the first separation between the sample complexity of properly learning a concept class with (a#x03B5;, a#x03B4;) differential privacy and learning without privacy. For properly learning thresholds in 'dimensions, this lower bound extends to n a#x2265; (l a#x2219; log* |X|). To obtain our results, we give reductions in both directions from releasing and properly learning thresholds and the simpler interior point problem. Given a database D of elements from X, the interior point problem asks for an element between the smallest and largest elements in D. We introduce new recursive constructions for bounding the sample complexity of the interior point problem, as well as further reductions and techniques for proving impossibility results for other basic problems in differential privacy.

Journal ArticleDOI
TL;DR: Results for a set of water network design problems show that the new algorithm can lead to orders of magnitude reduction in the optimality gap compared to commercial solvers.

Journal ArticleDOI
TL;DR: In this paper, the authors established the optimal rate of convergence for graphon estimation for the stochastic block model with k-clusters and showed that the optimal convergence rate is n −1/log k+k −2/n −2 +k−2 +n−2.
Abstract: Network analysis is becoming one of the most active research areas in statistics. Significant advances have been made recently on developing theories, methodologies and algorithms for analyzing networks. However, there has been little fundamental study on optimal estimation. In this paper, we establish optimal rate of convergence for graphon estimation. For the stochastic block model with $k$ clusters, we show that the optimal rate under the mean squared error is $n^{-1}\log k+k^{2}/n^{2}$. The minimax upper bound improves the existing results in literature through a technique of solving a quadratic equation. When $k\leq\sqrt{n\log n}$, as the number of the cluster $k$ grows, the minimax rate grows slowly with only a logarithmic order $n^{-1}\log k$. A key step to establish the lower bound is to construct a novel subset of the parameter space and then apply Fano’s lemma, from which we see a clear distinction of the nonparametric graphon estimation problem from classical nonparametric regression, due to the lack of identifiability of the order of nodes in exchangeable random graph models. As an immediate application, we consider nonparametric graphon estimation in a Holder class with smoothness $\alpha$. When the smoothness $\alpha\geq1$, the optimal rate of convergence is $n^{-1}\log n$, independent of $\alpha$, while for $\alpha\in(0,1)$, the rate is $n^{-2\alpha/(\alpha+1)}$, which is, to our surprise, identical to the classical nonparametric rate.

Journal ArticleDOI
TL;DR: This paper investigates event-triggered pinning control for the synchronization of complex networks of nonlinear dynamical systems described by time-varying weighted graphs and featuring generic linear interaction protocols.
Abstract: This paper investigates event-triggered pinning control for the synchronization of complex networks of nonlinear dynamical systems. We consider networks described by time-varying weighted graphs and featuring generic linear interaction protocols. Sufficient conditions for the absence of Zeno behavior are derived and exponential convergence of a global normed error function is proven. Static networks are considered as a special case, wherein the existence of a lower bound for interevent times is also proven. Numerical examples demonstrate the effectiveness of the proposed control strategy.

Journal ArticleDOI
TL;DR: In this paper, the authors obtained sharp probabilistic lower bounds on the smallest singular value of λ(min ε(Gamma) under the assumption that ε is an isotropic random vector for which the linear forms satisfy a weak small ball property.
Abstract: Given $X$ a random vector in ${\mathbb{R}}^n$, set $X_1,...,X_N$ to be independent copies of $X$ and let $\Gamma=\frac{1}{\sqrt{N}}\sum_{i=1}^N e_i$ be the matrix whose rows are $\frac{X_1}{\sqrt{N}},\dots, \frac{X_N}{\sqrt{N}}$. We obtain sharp probabilistic lower bounds on the smallest singular value $\lambda_{\min}(\Gamma)$ in a rather general situation, and in particular, under the assumption that $X$ is an isotropic random vector for which $\sup_{t\in S^{n-1}}{\mathbb{E}}| |^{2+\eta} \leq L$ for some $L,\eta>0$. Our results imply that a Bai-Yin type lower bound holds for $\eta>2$, and, up to a log-factor, for $\eta=2$ as well. The bounds hold without any additional assumptions on the Euclidean norm $\|X\|_{\ell_2^n}$. Moreover, we establish a nontrivial lower bound even without any higher moment assumptions (corresponding to the case $\eta=0$), if the linear forms satisfy a weak `small ball' property.

Journal ArticleDOI
TL;DR: This paper shows that, under the average error probability formalism, the third-order term in the normal approximation for the additive white Gaussian noise channel with a maximal or equal power constraint is at least 1 over 2 log n+O(1).
Abstract: This paper shows that, under the average error probability formalism, the third-order term in the normal approximation for the additive white Gaussian noise channel with a maximal or equal power constraint is at least $({1}/{2})\log n\,+\,O(1)$ . This improves on the lower bound by Polyanskiys–Poor–Verdu (2010) and matches the upper bound proved by the same authors.

Journal ArticleDOI
TL;DR: The proposed distributed BP algorithm has a near-optimal delay performance, approaching that of the high-complexity exhaustive search method; the modified BP offers a good delay performance at low communication complexity; both the average degree distribution and the outage upper bound analysis relying on stochastic geometry match well with the authors' Monte-Carlo simulations.
Abstract: Heterogeneous cellular networks (HCNs) with embedded small cells are considered, where multiple mobile users wish to download network content of different popularity. By caching data into the small-cell base stations, we will design distributed caching optimization algorithms via belief propagation (BP) for minimizing the downloading latency. First, we derive the delay-minimization objective function and formulate an optimization problem. Then, we develop a framework for modeling the underlying HCN topology with the aid of a factor graph. Furthermore, a distributed BP algorithm is proposed based on the network's factor graph. Next, we prove that a fixed point of convergence exists for our distributed BP algorithm. In order to reduce the complexity of the BP, we propose a heuristic BP algorithm. Furthermore, we evaluate the average downloading performance of our HCN for different numbers and locations of the base stations and mobile users, with the aid of stochastic geometry theory. By modeling the nodes distributions using a Poisson point process, we develop the expressions of the average factor graph degree distribution, as well as an upper bound of the outage probability for random caching schemes. We also improve the performance of random caching. Our simulations show that 1) the proposed distributed BP algorithm has a near-optimal delay performance, approaching that of the high-complexity exhaustive search method; 2) the modified BP offers a good delay performance at low communication complexity; 3) both the average degree distribution and the outage upper bound analysis relying on stochastic geometry match well with our Monte-Carlo simulations; and 4) the optimization based on the upper bound provides both a better outage and a better delay performance than the benchmarks.

Proceedings Article
07 Dec 2015
TL;DR: In this paper, an upper sample complexity bound of O(S|2|A|H2/∊2 ln 1/δ) was derived for episodic fixed-horizon MDPs.
Abstract: Recently, there has been significant progress in understanding reinforcement learning in discounted infinite-horizon Markov decision processes (MDPs) by deriving tight sample complexity bounds. However, in many real-world applications, an interactive learning agent operates for a fixed or bounded period of time, for example tutoring students for exams or handling customer service requests. Such scenarios can often be better treated as episodic fixed-horizon MDPs, for which only looser bounds on the sample complexity exist. A natural notion of sample complexity in this setting is the number of episodes required to guarantee a certain performance with high probability (PAC guarantee). In this paper, we derive an upper PAC bound O(|S|2|A|H2/∊2 ln 1/δ) and a lower PAC bound Ω(|S||A|H2/∊2 ln 1/δ+c) that match up to log-terms and an additional linear dependency on the number of states |S|. The lower bound is the first of its kind for this setting. Our upper bound leverages Bernstein's inequality to improve on previous bounds for episodic finite-horizon MDPs which have a time-horizon dependency of at least H3.

Journal ArticleDOI
TL;DR: It is shown that the coherence of a noise source can be quantified by the unitarity, which relates to the average change in purity averaged over input pure states and provides a lower bound on the optimal achievable gate infidelity under a given noisy process.
Abstract: Noise mechanisms in quantum systems can be broadly characterized as either coherent (i.e., unitary) or incoherent. For a given fixed average error rate, coherent noise mechanisms will generally lead to a larger worst-case error than incoherent noise. We show that the coherence of a noise source can be quantified by the unitarity, which we relate to the average change in purity averaged over input pure states. We then show that the unitarity can be efficiently estimated using a protocol based on randomized benchmarking that is efficient and robust to state-preparation and measurement errors. We also show that the unitarity provides a lower bound on the optimal achievable gate infidelity under a given noisy process.

Journal ArticleDOI
TL;DR: This paper considers the containment control problem for multi-agent systems with general linear dynamics and multiple leaders whose control inputs are possibly nonzero and time varying and proposes a distributed adaptive continuous controller, which can be designed and implemented by each follower in a fully distributed fashion.
Abstract: Summary This paper considers the containment control problem for multi-agent systems with general linear dynamics and multiple leaders whose control inputs are possibly nonzero and time varying. Based on the relative states of neighboring agents, a distributed static continuous controller is designed, under which the containment error is uniformly ultimately bounded and the upper bound of the containment error can be made arbitrarily small, if the subgraph associated with the followers is undirected and, for each follower, there exists at least one leader that has a directed path to that follower. It is noted that the design of the static controller requires the knowledge of the eigenvalues of the Laplacian matrix and the upper bounds of the leaders’ control inputs. In order to remove these requirements, a distributed adaptive continuous controller is further proposed, which can be designed and implemented by each follower in a fully distributed fashion. Extensions to the case where only local output information is available and to the case of multi-agent systems with matching uncertainties are also discussed. Copyright © 2014 John Wiley & Sons, Ltd.

Posted Content
TL;DR: In this paper, the authors studied the problem of estimating the number of samples required to answer a sequence of adaptive queries about an unknown distribution, as a function of the type of queries and the desired level of accuracy.
Abstract: Adaptivity is an important feature of data analysis---the choice of questions to ask about a dataset often depends on previous interactions with the same dataset. However, statistical validity is typically studied in a nonadaptive model, where all questions are specified before the dataset is drawn. Recent work by Dwork et al. (STOC, 2015) and Hardt and Ullman (FOCS, 2014) initiated the formal study of this problem, and gave the first upper and lower bounds on the achievable generalization error for adaptive data analysis. Specifically, suppose there is an unknown distribution $\mathbf{P}$ and a set of $n$ independent samples $\mathbf{x}$ is drawn from $\mathbf{P}$. We seek an algorithm that, given $\mathbf{x}$ as input, accurately answers a sequence of adaptively chosen queries about the unknown distribution $\mathbf{P}$. How many samples $n$ must we draw from the distribution, as a function of the type of queries, the number of queries, and the desired level of accuracy? In this work we make two new contributions: (i) We give upper bounds on the number of samples $n$ that are needed to answer statistical queries. The bounds improve and simplify the work of Dwork et al. (STOC, 2015), and have been applied in subsequent work by those authors (Science, 2015, NIPS, 2015). (ii) We prove the first upper bounds on the number of samples required to answer more general families of queries. These include arbitrary low-sensitivity queries and an important class of optimization queries. As in Dwork et al., our algorithms are based on a connection with algorithmic stability in the form of differential privacy. We extend their work by giving a quantitatively optimal, more general, and simpler proof of their main theorem that stability implies low generalization error. We also study weaker stability guarantees such as bounded KL divergence and total variation distance.

Proceedings Article
25 Jul 2015
TL;DR: FlyMC as discussed by the authors is a variant of MCMC with auxiliary variables that only queries the likelihoods of a subset of the data at each iteration yet simulates from the exact posterior distribution.
Abstract: Markov chain Monte Carlo (MCMC) is a popular tool for Bayesian inference. However, MCMC cannot be practically applied to large data sets because of the prohibitive cost of evaluating every likelihood term at every iteration. Here we present Firefly Monte Carlo (FlyMC) MCMC algorithm with auxiliary variables that only queries the likelihoods of a subset of the data at each iteration yet simulates from the exact posterior distribution. FlyMC is compatible with modern MCMC algorithms, and only requires a lower bound on the per-datum likelihood factors. In experiments, we find that FlyMC generates samples from the posterior more than an order of magnitude faster than regular MCMC, allowing MCMC methods to tackle larger datasets than were previously considered feasible.

Journal ArticleDOI
TL;DR: This article presents a realistic “meta-GGA made very simple” (MGGA-MVS) for exchange that respects this optimal bound on the exchange energy, which no previous beyond-LSDA approximation satisfies.
Abstract: Because of its useful accuracy and efficiency, density functional theory (DFT) is one of the most widely used electronic structure theories in physics, materials science, and chemistry. Only the exchange-correlation energy is unknown, and needs to be approximated in practice. Exact constraints provide useful information about this functional. The local spin-density approximation (LSDA) was the first constraint-based density functional. The Lieb–Oxford lower bound on the exchange-correlation energy for any density is another constraint that plays an important role in the development of generalized gradient approximations (GGAs) and meta-GGAs. Recently, a strongly and optimally tightened lower bound on the exchange energy was proved for one- and two-electron densities, and conjectured for all densities. In this article, we present a realistic “meta-GGA made very simple” (MGGA-MVS) for exchange that respects this optimal bound, which no previous beyond-LSDA approximation satisfies. This constraint might have been expected to worsen predicted thermochemical properties, but in fact they are improved over those of the Perdew–Burke–Ernzerhof GGA, which has nearly the same correlation part. MVS exchange is however radically different from that of other GGAs and meta-GGAs. Its exchange enhancement factor has a very strong dependence upon the orbital kinetic energy density, which permits accurate energies even with the drastically tightened bound. When this nonempirical MVS meta-GGA is hybridized with 25% of exact exchange, the resulting global hybrid gives excellent predictions for atomization energies, reaction barriers, and weak interactions of molecules.

Proceedings ArticleDOI
21 Jul 2015
TL;DR: This paper presents several new bounds on the time and message complexities of randomized Monte Carlo algorithms for Graph Connectivity and Minimum Spanning Tree in the Congested Clique, and shows that Ω(n2) messages are needed by any algorithm that solves GC, regardless of the number of rounds used.
Abstract: We study two fundamental graph problems, Graph Connectivity (GC) and Minimum Spanning Tree (MST), in the well-studied Congested Clique model, and present several new bounds on the time and message complexities of randomized algorithms for these problems. No non-trivial (i.e., super-constant) time lower bounds are known for either of the aforementioned problems; in particular, an important open question is whether or not constant-round algorithms exist for these problems. We make progress toward answering this question by presenting randomized Monte Carlo algorithms for both problems that run in O(log log log n) rounds (where n is the size of the clique). Our results improve by an exponential factor on the long-standing (deterministic) time bound of O(log log n) rounds for these problems due to Lotker et al. (SICOMP 2005). Our algorithms make use of several algorithmic tools including graph sketching, random sampling, and fast sorting.The second contribution of this paper is to present several almost-tight bounds on the message complexity of these problems. Specifically, we show that Ω(n2) messages are needed by any algorithm (including randomized Monte Carlo algorithms, and regardless of the number of rounds) that solves the GC (and hence also the MST) problem if each machine in the Congested Clique has initial knowledge only of itself (the so-called KT0 model). In contrast, if the machines have initial knowledge of their neighbors' IDs (the so-called KT1 model), we present a randomized Monte Carlo algorithm for MST that uses O(n polylog n) messages and runs in O(polylog n) rounds. To complement this, we also present a lower bound in the KT1 model that shows that Ω(n) messages are required by any algorithm that solves GC, regardless of the number of rounds used. Our results are a step toward understanding the power of randomization in the Congested Clique with respect to both time and message complexity.

Posted Content
TL;DR: In this article, the performance of multiple-input multiple-output underwater wireless optical communication (MIMO UWOC) systems with on-off keying (OOK) modulation is analyzed.
Abstract: In this paper, we analytically study the performance of multiple-input multiple-output underwater wireless optical communication (MIMO UWOC) systems with on-off keying (OOK) modulation. To mitigate turbulence-induced fading, which is amongst the major degrading effects of underwater channels on the propagating optical signal, we use spatial diversity over UWOC links. Furthermore, the effects of absorption and scattering are considered in our analysis. We analytically obtain the exact and an upper bound bit error rate (BER) expressions for both optimal and equal gain combining. In order to more effectively calculate the system BER, we apply Gauss-Hermite quadrature formula as well as approximation to the sum of lognormal random variables. We also apply photon-counting method to evaluate the system BER in the presence of shot noise. Our numerical results indicate an excellent match between the exact and upper bound BER curves. Also {a good match} between {the} analytical results and numerical simulations confirms the accuracy of our derived expressions. Moreover, our results show that spatial diversity can considerably improve the system performance, especially for channels with higher turbulence, e.g., a $3\times1$ MISO transmission in a $25$ {m} coastal water link with log-amplitude variance of $0.16$ can introduce $8$ {dB} performance improvement at the BER of $10^{-9}$.

Journal ArticleDOI
TL;DR: This paper will be able to establish a better bound on $\tau$ making use of Lyapunov functionals and discuss the stabilization not only in the sense of exponential stability but also in other sense---that of $H_\infty$ stability or asymptotic stability.
Abstract: Recently, Mao [Automatica J. IFAC, 49 (2013), pp. 3677--3681] initiated the study the mean-square exponential stabilization of continuous-time hybrid stochastic differential equations by feedback controls based on discrete-time state observations. In the same paper Mao also obtains an upper bound on the duration $\tau$ between two consecutive state observations. However, it is due to the general technique used there that the bound on $\tau$ is not very sharp. In this paper, we will be able to establish a better bound on $\tau$ making use of Lyapunov functionals. We will discuss the stabilization not only in the sense of exponential stability (as Mao does in [Automatica J. IFAC, 49 (2013), pp. 3677--3681]) but also in other sense---that of $H_\infty$ stability or asymptotic stability. We will consider not only the mean square stability but also the almost sure stability.

Journal ArticleDOI
TL;DR: In this paper, the Li-Yau gradient estimate for the heat kernel on graphs was shown to be locally local and can be considered as a new notion of curvature for graphs.
Abstract: We prove the Li-Yau gradient estimate for the heat kernel on graphs. The onlyassumption is a variant of the curvature-dimension inequality, which is purely local,and can be considered as a new notion of curvature for graphs. We compute thiscurvature for lattices and trees and conclude that it behaves more naturally than thealready existing notions of curvature. Moreover, we show that if a graph has non-negative curvature then it has polynomial volume growth.We also derive Harnack inequalities and heat kernel bounds from the gradient esti-mate, and show how it can be used to strengthen the classical Buser inequality relatingthe spectral gap and the Cheeger constant of a graph. 1 Introduction and main ideas In their celebrated work [15] Li and Yau proved an upper bound on the gradient ofpositive solutions of the heat equation. In its simplest form, for an n-dimensional com-pact manifold with non-negative Ricci curvature the Li-Yau gradient estimate statesthat a positive solution uof the heat equation (∆−∂

Journal ArticleDOI
TL;DR: This work introduces a technique to design interval observers for linear systems affected by state and measurement disturbances, based on the Internal Positive Representations of systems, that exploits the order preserving property of positive systems.
Abstract: Interval observers are dynamic systems that provide upper and lower bounds of the true state trajectories of systems. In this work we introduce a technique to design interval observers for linear systems affected by state and measurement disturbances, based on the Internal Positive Representations (IPRs) of systems, that exploits the order preserving property of positive systems. The method can be applied to both continuous and discrete time systems.