scispace - formally typeset
Search or ask a question

Showing papers on "Statistical learning theory published in 2013"


Journal ArticleDOI
TL;DR: This paper reviews theories developed to understand the properties and behaviors of multi-view learning and gives a taxonomy of approaches according to the machine learning mechanisms involved and the fashions in which multiple views are exploited.
Abstract: Multi-view learning or learning with multiple distinct feature sets is a rapidly growing direction in machine learning with well theoretical underpinnings and great practical success. This paper reviews theories developed to understand the properties and behaviors of multi-view learning and gives a taxonomy of approaches according to the machine learning mechanisms involved and the fashions in which multiple views are exploited. This survey aims to provide an insightful organization of current developments in the field of multi-view learning, identify their limitations, and give suggestions for further research. One feature of this survey is that we attempt to point out specific open problems which can hopefully be useful to promote the research of multi-view machine learning.

782 citations


Journal ArticleDOI
TL;DR: A new cost-sensitive algorithm (CSMLP) is presented to improve the discrimination ability of (two-class) MLPs and it is theoretically demonstrated that the incorporation of prior information via the cost parameter may lead to balanced decision boundaries in the feature space.
Abstract: Traditional learning algorithms applied to complex and highly imbalanced training sets may not give satisfactory results when distinguishing between examples of the classes. The tendency is to yield classification models that are biased towards the overrepresented (majority) class. This paper investigates this class imbalance problem in the context of multilayer perceptron (MLP) neural networks. The consequences of the equal cost (loss) assumption on imbalanced data are formally discussed from a statistical learning theory point of view. A new cost-sensitive algorithm (CSMLP) is presented to improve the discrimination ability of (two-class) MLPs. The CSMLP formulation is based on a joint objective function that uses a single cost parameter to distinguish the importance of class errors. The learning rule extends the Levenberg-Marquadt's rule, ensuring the computational efficiency of the algorithm. In addition, it is theoretically demonstrated that the incorporation of prior information via the cost parameter may lead to balanced decision boundaries in the feature space. Based on the statistical analysis of results on real data, our approach shows a significant improvement of the area under the receiver operating characteristic curve and G-mean measures of regular MLPs.

195 citations


Book
15 Jun 2013
TL;DR: This unique text/reference describes in detail the latest advances in unsupervised process monitoring and fault diagnosis with machine learning methods.
Abstract: This unique text/reference describes in detail the latest advances in unsupervised process monitoring and fault diagnosis with machine learning methods. Abundant case studies throughout the text demonstrate the efficacy of each method in real-world settings. The broad coverage examines such cutting-edge topics as the use of information theory to enhance unsupervised learning in tree-based methods, the extension of kernel methods to multiple kernel learning for feature extraction from data, and the incremental training of multilayer perceptrons to construct deep architectures for enhanced data projections. Topics and features: discusses machine learning frameworks based on artificial neural networks, statistical learning theory and kernel-based methods, and tree-based methods; examines the application of machine learning to steady state and dynamic operations, with a focus on unsupervised learning; describes the use of spectral methods in process fault diagnosis.

104 citations


Journal ArticleDOI
TL;DR: In this article, the generalization performance of online learning algorithms trained on samples coming from a dependent source of data was studied, and it was shown that the regret of any stable online algorithm concentrates around its regret, an easily computable statistic of the online performance.
Abstract: We study the generalization performance of online learning algorithms trained on samples coming from a dependent source of data. We show that the generalization error of any stable online algorithm concentrates around its regret-an easily computable statistic of the online performance of the algorithm-when the underlying ergodic process is β- or φ -mixing. We show high-probability error bounds assuming the loss function is convex, and we also establish sharp convergence rates and deviation bounds for strongly convex losses and several linear prediction problems such as linear and logistic regression, least-squares SVM, and boosting on dependent data. In addition, our results have straightforward applications to stochastic optimization with dependent data, and our analysis requires only martingale convergence arguments; we need not rely on more powerful statistical tools such as empirical process theory.

69 citations


Journal ArticleDOI
TL;DR: A new model integrating the SVR and the ICA for time estimation in NPD projects, in which ICA is used to tune the parameters of the S VR, and results indicate that the presented model achieves high estimation accuracy and leads to effective prediction.
Abstract: Time estimation in new product development (NPD) projects is often a complex problem due to its nonlinearity and the small quantity of data patterns. Support vector regression (SVR) based on statistical learning theory is introduced as a new neural network technique with maximum generalization ability. The SVR has been utilized to solve nonlinear regression problems successfully. However, the applicability of the SVR is highly affected due to the difficulty of selecting the SVR parameters appropriately. The imperialist competitive algorithm (ICA) as a socio-politically inspired optimization strategy is employed to solve the real world engineering problems. This optimization algorithm is inspired by competition mechanism among imperialists and colonies, in contrast to evolutionary algorithms. This paper presents a new model integrating the SVR and the ICA for time estimation in NPD projects, in which ICA is used to tune the parameters of the SVR. A real data set from a case study of an NPD project in a manufacturing industry is presented to demonstrate the performance of the proposed model. In addition, the comparison is provided between the proposed model and conventional techniques, namely nonlinear regression, back-propagation neural networks (BPNN), pure SVR and general regression neural networks (GRNN). The experimental results indicate that the presented model achieves high estimation accuracy and leads to effective prediction. Highlights? Proposing a new support vector model to capture data patterns of time intervals. ? Employing imperialist competitive algorithm to optimize the parameters of SVR. ? Presenting a real case study in a manufacturing industry in the NPD environment. ? Providing a comparison between the proposed model and conventional techniques.

61 citations


Journal ArticleDOI
TL;DR: It is proved that even the very best generalized structure-based model is inherently limited in its accuracy, and protein-specific models are always likely to be better.
Abstract: A major goal in computational chemistry has been to discover the set of rules that can accurately predict the binding affinity of any protein-drug complex, using only a single snapshot of its three-dimensional structure. Despite the continual development of structure-based models, predictive accuracy remains low, and the fundamental factors that inhibit the inference of all-encompassing rules have yet to be fully explored. Using statistical learning theory and information theory, here we prove that even the very best generalized structure-based model is inherently limited in its accuracy, and protein-specific models are always likely to be better. Our results refute the prevailing assumption that large data sets and advanced machine learning techniques will yield accurate, universally applicable models. We anticipate that the results will aid the development of more robust virtual screening strategies and scoring function error estimations.

42 citations


Journal ArticleDOI
Wencong Lu1, Xiaobo Ji1, Minjie Li1, Liang Liu1, Baohua Yue1, Liangmiao Zhang1 
TL;DR: Support vector machine (SVM), including support vector classification (SVC) and support vector regression (SVR) based on the statistical learning theory (SLT) proposed by Vapnik, is introduced as a relatively new data mining method to meet the different tasks of materials design in the lab.
Abstract: Materials design is the most important and fundamental work on the background of materials genome initiative for global competitiveness proposed by the National Science and Technology Council of America. As far as the methodologies of materials design, besides the thermodynamic and kinetic methods combing databases, both deductive approaches so-called the first principle methods and inductive approaches based on data mining methods are gaining great progress because of their successful applications in materials design. In this paper, support vector machine (SVM), including support vector classification (SVC) and support vector regression (SVR) based on the statistical learning theory (SLT) proposed by Vapnik, is introduced as a relatively new data mining method to meet the different tasks of materials design in our lab. The advantage of using SVM for materials design is discussed based on the applications in the formability of perovskite or BaNiO3 structure, the prediction of energy gaps of binary compounds, the prediction of sintered cold modulus of sialon-corundum castable, the optimization of electric resistances of VPTC semiconductors and the thickness control of In2O3 semiconductor film preparation. The results presented indicate that SVM is an effective modeling tool for the small sizes of sample sets with great potential applications in materials design.

41 citations


Journal ArticleDOI
TL;DR: In this article, the authors present a theoretical analysis for prediction algorithms based on association rules and introduce a problem for which rules are particularly natural, called "sequential event prediction." In sequential event prediction, events in a sequence are revealed one by one, and the goal is to determine which event will next be revealed.
Abstract: We present a theoretical analysis for prediction algorithms based on association rules. As part of this analysis, we introduce a problem for which rules are particularly natural, called "sequential event prediction." In sequential event prediction, events in a sequence are revealed one by one, and the goal is to determine which event will next be revealed. The training set is a collection of past sequences of events. An example application is to predict which item will next be placed into a customer's online shopping cart, given his/her past purchases. In the context of this problem, algorithms based on association rules have distinct advantages over classical statistical and machine learning methods: they look at correlations based on subsets of co-occurring past events (items a and b imply item c), they can be applied to the sequential event prediction problem in a natural way, they can potentially handle the "cold start" problem where the training set is small, and they yield interpretable predictions. In this work, we present two algorithms that incorporate association rules. These algorithms can be used both for sequential event prediction and for supervised classification, and they are simple enough that they can possibly be understood by users, customers, patients, managers, etc. We provide generalization guarantees on these algorithms based on algorithmic stability analysis from statistical learning theory. We include a discussion of the strict minimum support threshold often used in association rule mining, and introduce an "adjusted confidence" measure that provides a weaker minimum support condition that has advantages over the strict minimum support. The paper brings together ideas from statistical learning theory, association rule mining and Bayesian analysis.

40 citations


Journal Article
TL;DR: In this article, the authors propose a method to align statistical modeling with decision making by propagating uncertainty in predictive modeling to the uncertainty in operational cost, where operational cost is the amount spent by the practitioner in solving the problem.
Abstract: This work proposes a way to align statistical modeling with decision making. We provide a method that propagates the uncertainty in predictive modeling to the uncertainty in operational cost, where operational cost is the amount spent by the practitioner in solving the problem. The method allows us to explore the range of operational costs associated with the set of reasonable statistical models, so as to provide a useful way for practitioners to understand uncertainty. To do this, the operational cost is cast as a regularization term in a learning algorithm's objective function, allowing either an optimistic or pessimistic view of possible costs, depending on the regularization parameter. From another perspective, if we have prior knowledge about the operational cost, for instance that it should be low, this knowledge can help to restrict the hypothesis space, and can help with generalization. We provide a theoretical generalization bound for this scenario. We also show that learning with operational costs is related to robust optimization.

40 citations


Journal ArticleDOI
TL;DR: In this paper, the PAC-Bayesian approach was used for quantile forecasting of the French GDP and it was shown that the Gibbs estimator actually achieves fast rates of convergence d/n.
Abstract: We establish rates of convergences in statistical learning for time series forecasting. Using the PAC-Bayesian approach, slow rates of convergence p d/n for the Gibbs estimator under the absolute loss were given in a previous work [7], where n is the sample size and d the dimension of the set of predictors. Under the same weak dependence conditions, we extend this result to any convex Lipschitz loss function. We also identify a condition on the parameter space that ensures similar rates for the classical penalized ERM procedure. We apply this method for quantile forecasting of the French GDP. Under additional conditions on the loss functions (satisfied by the quadratic loss function) and for uniformly mixing processes, we prove that the Gibbs estimator actually achieves fast rates of convergence d/n. We discuss the optimality of these dierent rates pointing out references to lower bounds when they are available. In particular, these results bring a generalization the results of [29] on sparse regression estimation to some autoregression.

36 citations


Proceedings ArticleDOI
12 Jun 2013
TL;DR: This article will illustrate recent interest in identification algorithms with the use of subspace methods as well as nuclear norms as proxies to rank constraints and a quite different route to convexity is to use algebraic techniques manipulate the model parameterizations.
Abstract: System Identification is about estimating models of dynamical systems from measured input-output data. Its traditional foundation is basic statistical techniques, such as maximum likelihood estimation and asymptotic analysis of bias and variance and the like. Maximum likelihood estimation relies on minimization of criterion functions that typically are non-convex, and may cause numerical search problems. Recent interest in identification algorithms has focused on techniques that are centered around convex formulations. This is partly the result of developments in machine learning and statistical learning theory. The development concerns issues of regularization for sparsity and for better tuned bias/variance trade-offs. It also involves the use of subspace methods as well as nuclear norms as proxies to rank constraints. A quite different route to convexity is to use algebraic techniques manipulate the model parameterizations. This article will illustrate all this recent development.

Journal ArticleDOI
TL;DR: Support vector regression (SVR) method based on statistical learning theory (SLT) was employed as a supervised learning algorithm to estimate Poisson ratio from conventional well log data and results indicated that SVR predicted Poisson ratios values are in good agreement with measured values.

Journal ArticleDOI
TL;DR: Numerical results on seven real world biomedical datasets support the effectiveness of the proposed approach compared to other commonly-used sparse SVM methods, including L"1-SVM, and recent approximated L"0-S VM approaches.

Posted Content
TL;DR: In this paper, the authors study the generalization capabilities of regularization in the framework of statistical learning theory and show that the upper and lower bounds of learning rates for $l^q$ regularization learning are asymptotically identical for all $0
Abstract: Regularization is a well recognized powerful strategy to improve the performance of a learning machine and $l^q$ regularization schemes with $0

Book ChapterDOI
01 Jan 2013
TL;DR: The finite sample distribution of many nonparametric methods from statistical learning theory is unknown because the distribution P from which the data were generated is unknown and because there often exist only asymptotical results on the behaviour of such methods.
Abstract: The finite sample distribution of many nonparametric methods from statistical learning theory is unknown because the distribution P from which the data were generated is unknown and because there often exist only asymptotical results on the behaviour of such methods.

01 Nov 2013
TL;DR: This work presents two algorithms that incorporate association rules that can be used both for sequential event prediction and for supervised classification, and provides generalization guarantees on these algorithms based on algorithmic stability analysis from statistical learning theory.
Abstract: We present a theoretical analysis for prediction algorithms based on association rules. As part of this analysis, we introduce a problem for which rules are particularly natural, called "sequential event prediction." In sequential event prediction, events in a sequence are revealed one by one, and the goal is to determine which event will next be revealed. The training set is a collection of past sequences of events. An example application is to predict which item will next be placed into a customer's online shopping cart, given his/her past purchases. In the context of this problem, algorithms based on association rules have distinct advantages over classical statistical and machine learning methods: they look at correlations based on subsets of co-occurring past events (items a and b imply item c), they can be applied to the sequential event prediction problem in a natural way, they can potentially handle the "cold start" problem where the training set is small, and they yield interpretable predictions. In this work, we present two algorithms that incorporate association rules. These algorithms can be used both for sequential event prediction and for supervised classification, and they are simple enough that they can possibly be understood by users, customers, patients, managers, etc. We provide generalization guarantees on these algorithms based on algorithmic stability analysis from statistical learning theory. We include a discussion of the strict minimum support threshold often used in association rule mining, and introduce an "adjusted confidence" measure that provides a weaker minimum support condition that has advantages over the strict minimum support. The paper brings together ideas from statistical learning theory, association rule mining and Bayesian analysis.

Book ChapterDOI
01 Jan 2013
TL;DR: Information-theoretically reformulate two measures of capacity from statistical learning theory: empirical VC-entropy and empirical Rademacher complexity and show these capacity measures count the number of hypotheses about a dataset that a learning algorithm falsifies when it finds the classifier in its repertoire minimizing empirical risk.
Abstract: We information-theoretically reformulate two measures of capacity from statistical learning theory: empirical VC-entropy and empirical Rademacher complexity. We show these capacity measures count the number of hypotheses about a dataset that a learning algorithm falsifies when it finds the classifier in its repertoire minimizing empirical risk. It then follows from that the future performance of predictors on unseen data is controlled in part by how many hypotheses the learner falsifies. As a corollary we show that empirical VC-entropy quantifies the message length of the true hypothesis in the optimal code of a particular probability distribution, the so-called actual repertoire.

Book ChapterDOI
01 Jan 2013
TL;DR: In this article, a perturbed version of the operator equation for the estimator is used to solve the Tikhonov regularization problem, where this equation is seen as a perturbation of the ideal estimator.
Abstract: One-parameter regularization methods, such as the Tikhonov regularization, are used to solve the operator equation for the estimator in the statistical learning theory. Recently, there has been a lot of interest in the construction of the so called extrapolating estimators, which approximate the input–output relationship beyond the scope of the empirical data. The standard Tikhonov regularization produces rather poor extrapolating estimators. In this paper, we propose a novel view on the operator equation for the estimator where this equation is seen as a perturbed version of the operator equation for the ideal estimator. This view suggests the dual regularized total least squares (DRTLS) and multi-penalty regularization (MPR), which are multi-parameter regularization methods, as methods of choice for constructing better extrapolating estimators. We propose and test several realizations of DRTLS and MPR for constructing extrapolating estimators. It will be seen that, among the considered realizations, a realization of MPR gives best extrapolating estimators. For this realization, we propose a rule for the choice of the used regularization parameters that allows an automatic selection of the suitable extrapolating estimator.

Proceedings Article
29 Apr 2013
TL;DR: A formalism of localization for online learning problems, which, similarly to statistical learning theory, can be used to obtain fast rates, is introduced and a novel upper bound on regret in terms of classical Rademacher complexity is established.
Abstract: We introduce a formalism of localization for online learning problems, which, similarly to statistical learning theory, can be used to obtain fast rates. In particular, we introduce local sequential Rademacher complexities and other local measures. Based on the idea of relaxations for deriving algorithms, we provide a template method that takes advantage of localization. Furthermore, we build a general adaptive method that can take advantage of the suboptimality of the observed sequence. We illustrate the utility of the introduced concepts on several problems. Among them is a novel upper bound on regret in terms of classical Rademacher complexity when the data are i.i.d.

Journal ArticleDOI
TL;DR: This paper uses tools of statistical learning in order to design a more accurate prediction operator in Harten's framework based on a training sample, resulting in multiresolution decompositions with enhanced sparsity.

Journal Article
TL;DR: This paper develops the deviation inequalities and the symmetrization inequality for the learning process, and develops the risk bounds based on the covering number, and studies the asymptotic convergence and the rate of convergence of thelearning process for Leevy process.
Abstract: Leevy processes refer to a class of stochastic processes, for example, Poisson processes and Brownian motions, and play an important role in stochastic processes and machine learning. Therefore, it is essential to study risk bounds of the learning process for time-dependent samples drawn from a Leevy process (or briefly called learning process for Leevy process). It is noteworthy that samples in this learning process are not independently and identically distributed (i.i.d.). Therefore, results in traditional statistical learning theory are not applicable (or at least cannot be applied directly), because they are obtained under the sample-i.i.d. assumption. In this paper, we study risk bounds of the learning process for time-dependent samples drawn from a Leevy process, and then analyze the asymptotical behavior of the learning process. In particular, we first develop the deviation inequalities and the symmetrization inequality for the learning process. By using the resultant inequalities, we then obtain the risk bounds based on the covering number. Finally, based on the resulting risk bounds, we study the asymptotic convergence and the rate of convergence of the learning process for Leevy process. Meanwhile, we also give a comparison to the related results under the sample-i.i.d. assumption.

Journal ArticleDOI
TL;DR: This paper addresses the elastic-net regularization problem within the framework of statistical learning theory and shows the learning rate to be faster compared with existing results.

Proceedings ArticleDOI
01 Aug 2013
TL;DR: The support vector machine's classification mechanism and its application in mechanical fault diagnosis are introduced and some of the shortcomings of the machine learning algorithm are put forward.
Abstract: Support vector machine is a machine learning algorithm developed by Vapnik from the statistical learning theory for data classification via study from a small sample of fault data. For fault data it can isolate the fault categories accurately even though only has the small sample of data. In the present work, support vector machine's classification mechanism and its application in mechanical fault diagnosis are introduced. Therefore, give an instance the support vector machine makes fault classification for the coal mine scraper conveyor's faults. Last but not the least, put forward some of the shortcomings of the support vector machine and look forward to the direction of development of the support vector machine fault diagnosis in the future.

Posted Content
TL;DR: Experimental results reveal that a prototype system developed using SLT-based methods outperforms seven existing fake website detection systems on a test bed encompassing 900 real and fake websites.
Abstract: Existing fake website detection systems are unable to effectively detect fake websites. In this study, we advocate the development of fake website detection systems that employ classification methods grounded in statistical learning theory (SLT). Experimental results reveal that a prototype system developed using SLT-based methods outperforms seven existing fake website detection systems on a test bed encompassing 900 real and fake websites.

Journal ArticleDOI
TL;DR: A theoretical analysis for prediction algorithms based on association rules, which introduces a problem for which rules are particularly natural, called "sequential...
Abstract: We present a theoretical analysis for prediction algorithms based on association rules. As part of this analysis, we introduce a problem for which rules are particularly natural, called "sequential...

Journal ArticleDOI
TL;DR: The results of this study showed that the GA-SVM approach has the potential to be a practical tool for predicting compression index of soil.
Abstract: The compression index is an important soil property that is essential to many geotechnical designs. As the determination of the compression index from consolidation tests is relatively time-consuming. Support Vector Machine (SVM) is a statistical learning theory based on a structural risk minimization principle that minimizes both error and weight terms. Considering the fact that parameters in SVM model are difficult to be decided, a genetic SVM was presented in which the parameters in SVM method are optimized by Genetic Algorithm (GA). Taking plasticity index, water content, void ration and density of soil as primary influence factors, the prediction model of compression index based on GA-SVM approach was obtained. The results of this study showed that the GA-SVM approach has the potential to be a practical tool for predicting compression index of soil.

Book ChapterDOI
01 Jan 2013
TL;DR: The basics of kernel methods and their position in the generalized data-driven fault diagnostic framework are reviewed, and unsupervised kernel methods, such as kernel principal component analysis, are considered in detail, analogous to the application of linear principal components analysis in multivariate statistical process control.
Abstract: The basics of kernel methods and their position in the generalized data-driven fault diagnostic framework are reviewed. The review starts out with statistical learning theory, covering concepts such as loss functions, overfitting and structural and empirical risk minimization. This is followed by linear margin classifiers, kernels and support vector machines. Transductive support vector machines are discussed and illustrated by way of an example related to multivariate image analysis of coal particles on conveyor belts. Finally, unsupervised kernel methods, such as kernel principal component analysis, are considered in detail, analogous to the application of linear principal component analysis in multivariate statistical process control. Fault diagnosis in a simulated nonlinear system by the use of kernel principal component analysis is included as an example to illustrate the concepts.

Book ChapterDOI
01 Jan 2013
TL;DR: In this article, the authors present an overview of statistical learning theory, and describe key results regarding uniform convergence of empirical means and related sample complexity, and provide an extension of the probability inequalities studied in Chap. 8 to the case when parameterized families of functions are considered, instead of a fixed function.
Abstract: This chapter presents an overview of statistical learning theory, and describes key results regarding uniform convergence of empirical means and related sample complexity. This theory provides a fundamental extension of the probability inequalities studied in Chap. 8 to the case when parameterized families of functions are considered, instead of a fixed function. The chapter formally studies the UCEM (uniform convergence of empirical means) property and the VC dimension in the context of the Vapnik–Chervonenkis theory. Extensions to the Pollard theory for continuous-valued functions are also discussed.

Dissertation
06 Dec 2013
TL;DR: In this paper, the authors give an overview of high-dimensional statistics and statistical learning, under various sparsity assumptions, and provide explicitely the estimators used and optimal oracle inequalities satisfied by these estimators.
Abstract: The aim of this habilitation thesis is to give an overview of my works on high-dimensional statistics and statistical learning, under various sparsity assumptions. In a first part, I will describe the major challenges of high-dimensional statistics in the context of the generic linear regression model. After a brief review of existing results, I will present the theoretical study of aggregated estimators that was done in (Alquier & Lounici 2011). The second part essentially aims at providing extensions of the various theories presented in the first part to the estimation of time series models (Alquier & Doukhan 2011, Alquier & Wintenberger 2013, Alquier & Li 2012, Alquier, Wintenberger & Li 2012). Finally, the third part presents various extensions to nonparametric models, or to specific applications such as quantum statistics (Alquier & Biau 2013, Guedj & Alquier 2013, Alquier, Meziani & Peyre 2013, Alquier, Butucea, Hebiri, Meziani & Morimae 2013, Alquier 2013, Alquier 2008). In each section, we provide explicitely the estimators used and, as much as possible, optimal oracle inequalities satisfied by these estimators.

Proceedings ArticleDOI
16 Nov 2013
TL;DR: A novel analog circuit fault diagnosis method which is called wavelet kernel support vector machine, which is based on statistical learning theory, which has advantages of better classification ability and generalization performance is proposed.
Abstract: Analog circuit fault diagnosis can be regarded as the pattern recognition issue and addressed by machine learning theory. As compared with neural networks, support Vector Machine (SVM) is based on statistical learning theory, which has advantages of better classification ability and generalization performance. The marr wavelet kernel is proposed and the existence is proven by theoretic analysis and demonstration. Based on this, a novel analog circuit fault diagnosis method which is called wavelet kernel support vector machine is proposed in the paper. Using principal component analysis (PCA) as a tool for extracting fault features, the WSVM is then applied to the analog circuit fault diagnosis. The effectiveness of the proposed method is verified by the experimental results.