scispace - formally typeset
Search or ask a question

Showing papers on "Parametric statistics published in 2012"


Journal ArticleDOI
TL;DR: In this paper, a new signed differential mapping (SDM) method was proposed, called Effect Size SDM (ES-SDM), which enables the combination of statistical parametric maps and peak coordinates and uses well-established statistics.

552 citations


Journal ArticleDOI
TL;DR: In this article, the non-paranormal graphical models are used as a safe replacement of the popular Gaussian graphical models, even when the data are truly Gaussian, for graph recovery and parameter estimation.
Abstract: ing the Spearman’s rho and Kendall’s tau. We prove that the nonparanormal skeptic achieves the optimal parametric rates of convergence for both graph recovery and parameter estimation. This result suggests that the nonparanormal graphical models can be used as a safe replacement of the popular Gaussian graphical models, even when the data are truly Gaussian. Besides theoretical analysis, we also conduct thorough numerical simulations to compare the graph recovery performance of dierent estimators under both ideal and noisy settings. The proposed methods are then applied on a largescale genomic dataset to illustrate their empirical usefulness. The R package huge implementing the proposed methods is available on the Comprehensive R Archive Network: http://cran.r-project.org/.

521 citations


Journal ArticleDOI
TL;DR: A novel nearest neighbor-based feature weighting algorithm, which learns a feature Weighting vector by maximizing the expected leave-one-out classification accuracy with a regularization term, is proposed.
Abstract: Feature selection is of considerable importance in data mining and machine learning, especially for high dimensional data. In this paper, we propose a novel nearest neighbor-based feature weighting algorithm, which learns a feature weighting vector by maximizing the expected leave-one-out classification accuracy with a regularization term. The algorithm makes no parametric assumptions about the distribution of the data and scales naturally to multiclass problems. Experiments conducted on artificial and real data sets demonstrate that the proposed algorithm is largely insensitive to the increase in the number of irrelevant features and performs better than the state-of-the-art methods in most cases.

401 citations


Journal ArticleDOI
TL;DR: Building on over 3 decades of work, Pmetrics adopts a robust, reliable, and mature nonparametric approach to population modeling, which was better than the parametric method at discovering true pharmacokinetic subgroups and an outlier.
Abstract: Introduction:Nonparametric population modeling algorithms have a theoretical superiority over parametric methods to detect pharmacokinetic and pharmacodynamic subgroups and outliers within a study population.Methods:The authors created “Pmetrics,” a new Windows and Unix R software package that updat

398 citations


Journal ArticleDOI
TL;DR: The key idea is to align the complexity level and order of analysis with the reliability and detail level of statistical information on the input parameters to avoid the necessity to assign parametric probability distributions that are not sufficiently supported by limited available data.

350 citations


Journal ArticleDOI
TL;DR: In this paper, a parametric frontier approach is proposed to estimate economy-wide energy efficiency performance from a production efficiency point of view using the Shephard energy distance function to define an energy efficiency index and adopts the stochastic frontier analysis technique to estimate the index.

304 citations


Proceedings ArticleDOI
Xudong Cao1, Yichen Wei1, Fang Wen1, Jian Sun1
16 Jun 2012
TL;DR: This paper presents a very efficient, highly accurate, “Explicit Shape Regression” approach for face alignment that significantly outperforms the state-of-the-art in terms of both accuracy and efficiency.
Abstract: We present a very efficient, highly accurate, “Explicit Shape Regression” approach for face alignment. Unlike previous regression-based approaches, we directly learn a vectorial regression function to infer the whole facial shape (a set of facial landmarks) from the image and explicitly minimize the alignment errors over the training data. The inherent shape constraint is naturally encoded into the regressor in a cascaded learning framework and applied from coarse to fine during the test, without using a fixed parametric shape model as in most previous methods. To make the regression more effective and efficient, we design a two-level boosted regression, shape-indexed features and a correlation-based feature selection method. This combination enables us to learn accurate models from large training data in a short time (20 minutes for 2,000 training images), and run regression extremely fast in test (15 ms for a 87 landmarks shape). Experiments on challenging data show that our approach significantly outperforms the state-of-the-art in terms of both accuracy and efficiency.

303 citations


Journal ArticleDOI
20 Dec 2012-PLOS ONE
TL;DR: The statistical approaches for several tests of hypothesis and power/sample size calculations are detailed and applied to taxonomic abundance distribution and rank abundance distribution data using HMP Jumpstart data on 24 subjects for saliva, subgingival, and supragingival samples.
Abstract: This paper presents new biostatistical methods for the analysis of microbiome data based on a fully parametric approach using all the data. The Dirichlet-multinomial distribution allows the analyst to calculate power and sample sizes for experimental design, perform tests of hypotheses (e.g., compare microbiomes across groups), and to estimate parameters describing microbiome properties. The use of a fully parametric model for these data has the benefit over alternative non-parametric approaches such as bootstrapping and permutation testing, in that this model is able to retain more information contained in the data. This paper details the statistical approaches for several tests of hypothesis and power/sample size calculations, and applies them for illustration to taxonomic abundance distribution and rank abundance distribution data using HMP Jumpstart data on 24 subjects for saliva, subgingival, and supragingival samples. Software for running these analyses is available.

295 citations


Journal ArticleDOI
TL;DR: In this paper, a semiparametric frontier model that combines the DEA-type nonparametric frontier, which satisfies monotonicity and concavity, with the SFA-style stochastic homoskedastic composite error term is proposed.
Abstract: The field of productive efficiency analysis is currently divided between two main paradigms: the deterministic, nonparametric Data Envelopment Analysis (DEA) and the parametric Stochastic Frontier Analysis (SFA). This paper examines an encompassing semiparametric frontier model that combines the DEA-type nonparametric frontier, which satisfies monotonicity and concavity, with the SFA-style stochastic homoskedastic composite error term. To estimate this model, a new two-stage method is proposed, referred to as Stochastic Non-smooth Envelopment of Data (StoNED). The first stage of the StoNED method applies convex nonparametric least squares (CNLS) to estimate the shape of the frontier without any assumptions about its functional form or smoothness. In the second stage, the conditional expectations of inefficiency are estimated based on the CNLS residuals, using the method of moments or pseudolikelihood techniques. Although in a cross-sectional setting distinguishing inefficiency from noise in general requires distributional assumptions, we also show how these can be relaxed in our approach if panel data are available. Performance of the StoNED method is examined using Monte Carlo simulations.

285 citations


Book
22 Sep 2012
TL;DR: The PARADISE Bibliography as discussed by the authors is a collection of case studies in car steering, flight control, and sampled-data control systems, with a focus on robustness analysis by value sets for nonlinear co-efficient functions.
Abstract: * Parametric Plants and Controllers: A Crane Example * Boundary Crossing and Parameter Space Approach * Eigenvalue Specifications * Boundary Mapping in Parameter Space * Frequency Domain Analysis and Design * Case Studies in Car Steering * Case Studies in Flight Control * Robustness Analysis by Value Sets * Values Sets for Non-linear Co-efficient Functions * The Stability Radius * Robustness of Sampled-Data Control Systems A Polynominals and Polynominals Equations, Resultant Method B Introduction to PARADISE Bibliography

268 citations


Journal ArticleDOI
TL;DR: A finite dimensional static time-varying linear state feedback controller is obtained by truncating the predictor based controller and by adopting the parametric Lyapunov equation based controller design approach.

Journal ArticleDOI
TL;DR: A survey of the recent results and future prospects of the tensor-structured numerical methods in applications to multidimensional problems in scientific computing and focuses on the recent quantics-TT tensor approximation method that allows to represent N-d tensors with log-volume complexity.


Journal ArticleDOI
TL;DR: Very‐short‐term probabilistic forecasts, which are essential for an optimal management of wind generation, ought to account for the non‐linear and double‐bounded nature of that stochastic process, in the form of discrete–continuous mixtures of generalized logit–normal distributions and probability masses at the bounds.
Abstract: Summary. Very-short-term probabilistic forecasts, which are essential for an optimal management of wind generation, ought to account for the non-linear and double-bounded nature of that stochastic process. They take here the form of discrete–continuous mixtures of generalized logit–normal distributions and probability masses at the bounds. Both auto-regressive and conditional parametric auto-regressive models are considered for the dynamics of their location and scale parameters. Estimation is performed in a recursive least squares framework with exponential forgetting. The superiority of this proposal over classical assumptions about the shape of predictive densities, e.g. normal and beta, is demonstrated on the basis of 10-min-ahead point and probabilistic forecasting at the Horns Rev wind farm in Denmark.

Journal ArticleDOI
Xiao-Heng Chang1
TL;DR: Two sufficient conditions for the H filter design are proposed in terms of linear matrix inequalities (LMIs) when these LMIs are feasible, and an explicit expression of the desired filter is given.
Abstract: This paper is concerned with the H∞ filtering problem for continuous-time Takagi-Sugeno (T-S) fuzzy systems. Different from existing results for fuzzy H∞ filtering, the proposed ones are toward uncertain fuzzy systems with linear fractional parametric uncertainties. Attention is focused on the design of a fuzzy filter such that the filtering error system preserves a prescribed H∞ performance, where the filter to be designed is assumed to have gain variations. By a descriptor representation approach, two sufficient conditions for the H∞ filter design are proposed in terms of linear matrix inequalities (LMIs). When these LMIs are feasible, an explicit expression of the desired filter is given. A simulation example will be given to show the efficiency of the proposed methods.

Journal ArticleDOI
TL;DR: This work proposes a parametric sparse estimation technique based on finite rate of innovation (FRI) principles for MIMO communications, which is a generalization of conventional spectral estimation methods to multiple input signals with common support.
Abstract: We consider the problem of estimating sparse communication channels in the MIMO context. In small to medium bandwidth communications, as in the current standards for OFDM and CDMA communication systems (with bandwidth up to 20 MHz), such channels are individually sparse and at the same time share a common support set. Since the underlying physical channels are inherently continuous-time, we propose a parametric sparse estimation technique based on finite rate of innovation (FRI) principles. Parametric estimation is especially relevant to MIMO communications as it allows for a robust estimation and concise description of the channels. The core of the algorithm is a generalization of conventional spectral estimation methods to multiple input signals with common support. We show the application of our technique for channel estimation in OFDM (uniformly/contiguous DFT pilots) and CDMA downlink (Walsh-Hadamard coded schemes). In the presence of additive white Gaussian noise, theoretical lower bounds on the estimation of sparse common support (SCS) channel parameters in Rayleigh fading conditions are derived. Finally, an analytical spatial channel model is derived, and simulations on this model in the OFDM setting show the symbol error rate (SER) is reduced by a factor 2 (0 dB of SNR) to 5 (high SNR) compared to standard non-parametric methods - e.g. lowpass interpolation.

Journal ArticleDOI
TL;DR: This paper presents a methodology that combines the structure of mixed effects models for longitudinal and clustered data with the flexibility of tree-based estimation methods, and applies the resulting estimation method to pricing in online transactions, showing that the RE-EM tree is less sensitive to parametric assumptions and provides improved predictive power compared to linear models with random effects and regression trees without random effects.
Abstract: Longitudinal data refer to the situation where repeated observations are available for each sampled object. Clustered data, where observations are nested in a hierarchical structure within objects (without time necessarily being involved) represent a similar type of situation. Methodologies that take this structure into account allow for the possibilities of systematic differences between objects that are not related to attributes and autocorrelation within objects across time periods. A standard methodology in the statistics literature for this type of data is the mixed effects model, where these differences between objects are represented by so-called "random effects" that are estimated from the data (population-level relationships are termed "fixed effects," together resulting in a mixed effects model). This paper presents a methodology that combines the structure of mixed effects models for longitudinal and clustered data with the flexibility of tree-based estimation methods. We apply the resulting estimation method, called the RE-EM tree, to pricing in online transactions, showing that the RE-EM tree is less sensitive to parametric assumptions and provides improved predictive power compared to linear models with random effects and regression trees without random effects. We also apply it to a smaller data set examining accident fatalities, and show that the RE-EM tree strongly outperforms a tree without random effects while performing comparably to a linear model with random effects. We also perform extensive simulation experiments to show that the estimator improves predictive performance relative to regression trees without random effects and is comparable or superior to using linear models with random effects in more general situations.

Journal ArticleDOI
TL;DR: In this paper, a method for inverse identification of Johnson Cook parameters based on the Levenberg-Marquardt search algorithm is presented, using a particular set of Johnson-Cook parameters to describe the material behaviour.

Journal ArticleDOI
TL;DR: The authors examined the impact of random effects distribution misspecification on a variety of inferences, including prediction, inference about covariate effects, prediction of random effect and estimation of variance.
Abstract: Statistical models that include random effects are commonly used to analyze longitudinal and correlated data, often with strong and parametric assumptions about the random effects distribution. There is marked disagreement in the literature as to whether such parametric assumptions are important or innocuous. In the context of generalized linear mixed models used to analyze clustered or longitudinal data, we examine the impact of random effects distribution misspecification on a variety of inferences, including prediction, inference about covariate effects, prediction of random effects and estimation of random effects variances. We describe examples, theoretical calculations and simulations to elucidate situations in which the specification is and is not important. A key conclusion is the large degree of robustness of maximum likelihood for a wide variety of commonly encountered situations.

Journal ArticleDOI
TL;DR: This paper proposes a new approach to using multiple models to cope with transients which depends on the collective outputs of all the models, and can be viewed as a time-varying convex combination of the estimates.
Abstract: The concept of using multiple models to cope with transients which arise in adaptive systems with large parametric uncertainties was introduced in the 1990s. Both switching between multiple fixed models, and switching and tuning between fixed and adaptive models was proposed, and the stability of the resulting schemes was established. In all cases, the number of models needed is generally large (cn where n is the dimension of the parameter vector and c an integer), and the models do not “cooperate” in any real sense. In this paper, a new approach is proposed which represents a significant departure from past methods. First, it requires (n+1) models (in contrast to cn) which is significantly smaller, when “n ” is large. Second, while each of the (n+1) models chosen generates an estimate of the plant parameter vector, the new approach provides an estimate which depends on the collective outputs of all the models, and can be viewed as a time-varying convex combination of the estimates. It is then shown that control based on such an estimate results in a stable overall system. Further, arguments are given as to why such a procedure should result in faster convergence of the estimate to the true value of the plant parameter as compared to conventional adaptive controllers, resulting in better performance. Simulation studies are included to practically verify the arguments presented, and demonstrate the improvement in performance.

Journal ArticleDOI
TL;DR: The aim of this article is to bring together different specifications for copula models with time-varying dependence structure by conducting a simulation study to show the performance for model selection, to compare the model fit for different setups and to study the ability of the models to estimate the (latent) time-VARYing dependence parameter.
Abstract: The aim of this article is to bring together different specifications for copula models with time-varying dependence structure Copula models are widely used in financial econometrics and risk management They are considered to be a competitive alternative to the Gaussian dependence structure The dynamic structure of the dependence between the data can be modeled by allowing either the copula function or the dependence parameter to be time-varying First, we give a brief description of eight different models, among which there are fully parametric, semiparametric, and adaptive methods The purpose of this study is to compare the applicability of each particular model in different cases We conduct a simulation study to show the performance for model selection, to compare the model fit for different setups and to study the ability of the models to estimate the (latent) time-varying dependence parameter Finally, we provide an illustration by applying the competing models on the same financial dataset and

Journal ArticleDOI
TL;DR: An explicit symmetric linear phase-fitted four-step method with a free coefficient as parameter used for the optimization of the method in order to solve efficiently the Schrodinger equation and related oscillatory problems is developed.

Journal ArticleDOI
TL;DR: In this paper, two tests for the equality of covariance matrices between two high-dimensional populations are proposed, one test is on the whole variance-covariance matrix, and the other is on off-diagonal sub-matrices which define the covariance between two non-overlapping segments of the highdimensional random vectors.
Abstract: We propose two tests for the equality of covariance matrices between two high-dimensional populations. One test is on the whole variance-covariance matrices, and the other is on offdiagonal sub-matrices which define the covariance between two non-overlapping segments of the high-dimensional random vectors. The tests are applicable (i) when the data dimension is much larger than the sample sizes, namely the “large p, small n” situations and (ii) without assuming parametric distributions for the two populations. These two aspects surpass the capability of the conventional likelihood ratio test. The proposed tests can be used to test on covariances associated with gene ontology terms.

Journal ArticleDOI
TL;DR: The new parfm package remedies that lack by providing a wide range of parametric frailty models in R by maximising the marginal log-likelihood, with right-censored and possibly left-truncated data.
Abstract: Frailty models are getting more and more popular to account for overdispersion and/or clustering in survival data. When the form of the baseline hazard is somehow known in advance, the parametric estimation approach can be used advantageously. Nonetheless, there is no unified widely available software that deals with the parametric frailty model. The new parfm package remedies that lack by providing a wide range of parametric frailty models in R. The gamma, inverse Gaussian, and positive stable frailty distributions can be specified, together with five different baseline hazards. Parameter estimation is done by maximising the marginal log-likelihood, with right-censored and possibly left-truncated data. In the multivariate setting, the inverse Gaussian may encounter numerical difficulties with a huge number of events in at least one cluster. The positive stable model shows analogous difficulties but an ad-hoc solution is implemented, whereas the gamma model is very resistant due to the simplicity of its Laplace transform.

Journal ArticleDOI
TL;DR: The paper shows the improvement that occurs in verification by considering optimisation phases, the appropriateness of using new techniques of feedback, and the influence of optimisation parameters.
Abstract: This paper presents a new method for volumetric verification of machine tools. Beyond the consideration of a particular machine, a general verification methodology is presented based on the type of machine to verify the number and movement of axes and different techniques that can be used. A scheme and kinematic model with the inclusion of the measurement system depending on the kinematics of the machine are presented. The model describes the geometry and kinematics of a real milling machine based on a parametric synthetic data generator, which generates a test with known geometric errors and noise to enable a study of different optimisation techniques and models. Similarly, different errors identification techniques and volumetric verification models are presented and analysed. The paper shows the improvement that occurs in verification by considering optimisation phases, the appropriateness of using new techniques of feedback, and the influence of optimisation parameters. Chebyshev polynomials and its characteristics are presented, as well as a regression function for the new verification model. The new developed technique allows the characterisation of the different errors in the whole workspace of the machine and in less time than direct verification methods.

Journal ArticleDOI
TL;DR: In this paper, a multi-observer switching control strategy for robust active fault tolerant fuzzy control of variable-speed wind energy conversion systems (WECS) in the presence of wide wind variation, wind disturbance, parametric uncertainties, and sensors faults is proposed.
Abstract: In this paper, we propose a multiobserver switching control strategy for robust active fault tolerant fuzzy control (RAFTFC) of variable-speed wind energy conversion systems (WECS) in the presence of wide wind variation, wind disturbance, parametric uncertainties, and sensors faults. The Takagi-Sugeno (TS) fuzzy model with parametric uncertainties is adopted for modeling the nonlinear WECS and establishing fuzzy state observers. Sufficient conditions are derived for robust stabilization in the sense of Taylor series stability and are formulated in linear matrix inequalities (LMIs). Application to WECS, subject to uncertain parameters and sensor faults illustrate the effectiveness of the proposed controllers.

Journal ArticleDOI
TL;DR: This work applies a PSO approach in order to reconstruct a non-uniform rational B-spline (NURBS) surface of a certain order from a given set of 3D data points, and yields very good results even in presence of problematic features, such as multi-branches, high-genus or self-intersections.

Journal ArticleDOI
TL;DR: In this article, a massively parallel error analysis of the DMD algorithm using synthetic waveforms that are shown to be representative of the canonical instabilities observed in shear flows is presented.
Abstract: Dynamic mode decomposition (DMD) is a new diagnostic technique in fluid mechanics which is growing in popularity. A powerful analysis tool, it has great potential for measuring the spatial and temporal dynamics of coherent structures in experimental fluid flows. To aid interpretation of experimental data, error-bars on the measured growth rates are needed. In this article, we undertake a massively parallel error analysis of the DMD algorithm using synthetic waveforms that are shown to be representative of the canonical instabilities observed in shear flows. We show that the waveform of the instability has a marked impact on the error of the measured growth rate. Sawtooth and square waves may have an order of magnitude larger error than sine waves under the same conditions. We also show that the effects of data quantity and quality are of critical importance in determining the error in the growth or decay rate, and that the effect of the key parametric variables are modulated by the growth rate itself. We further demonstrate methods by which ensemble and orthogonal data may be introduced to improve the noise response. With regard for the important variables, precise measurement of the growth rates of instabilities may be supplemented with an accurately estimated uncertainty. This opens many new possibilities for the measurement of coherent structure in shear flows.

Journal ArticleDOI
TL;DR: Compared to other parametric human modeling approaches, this method contributes to the related field by introducing a novel method for correlating body shape and body sizes and by establishing an improved parameter optimization technique for the model generation process.
Abstract: The objective of this study is the development of a novel parametric human body shape modeling framework for integration into various product design applications. Our modeling framework is comprised of three phases of database construction, statistical analysis, and model generation. During the database construction phase, a 3D whole body scan data of 250 subjects are obtained, and their data structures are processed so as to be suitable for statistical analysis. Using those preprocessed scan data, the characteristics of the human body shape variation and their correlations with several items of body sizes are investigated in the statistical analysis phase. The correlations obtained from such analysis allow us to develop an interactive modeling interface, which takes the body sizes as inputs and returns a corresponding body shape model as an output. Using this interface, we develop a parametric human body shape modeling system and generate body shape models based on the input body sizes. In our experiment, our modeler produced reasonable results having not only a high level of accuracy but also fine visual fidelity. Compared to other parametric human modeling approaches, our method contributes to the related field by introducing a novel method for correlating body shape and body sizes and by establishing an improved parameter optimization technique for the model generation process.

Journal ArticleDOI
TL;DR: In this article, a parametric deterministic formulation of Bayesian inverse problems with an input parameter from infinite-dimensional, separable Banach spaces is presented, and the sparsity of the posterior density in terms of the summability of the input data's coefficient sequence is analyzed.
Abstract: We present a parametric deterministic formulation of Bayesian inverse problems with an input parameter from infinite-dimensional, separable Banach spaces. In this formulation, the forward problems are parametric, deterministic elliptic partial differential equations, and the inverse problem is to determine the unknown, parametric deterministic coefficients from noisy observations comprising linear functionals of the solution. We prove a generalized polynomial chaos representation of the posterior density with respect to the prior measure, given noisy observational data. We analyze the sparsity of the posterior density in terms of the summability of the input data's coefficient sequence. The first step in this process is to estimate the fluctuations in the prior. We exhibit sufficient conditions on the prior model in order for approximations of the posterior density to converge at a given algebraic rate, in terms of the number N of unknowns appearing in the parametric representation of the prior measure. Similar sparsity and approximation results are also exhibited for the solution and covariance of the elliptic partial differential equation under the posterior. These results then form the basis for efficient uncertainty quantification, in the presence of data with noise.