scispace - formally typeset
Search or ask a question
Topic

Linear discriminant analysis

About: Linear discriminant analysis is a research topic. Over the lifetime, 18361 publications have been published within this topic receiving 603195 citations. The topic is also known as: Linear discriminant analysis & LDA.


Papers
More filters
Proceedings ArticleDOI
02 May 2007
TL;DR: This work proposes a new method called sub-band common spatial pattern (SBCSP), which outperforms the other two approaches and achieves similar result as compared to the best one in the literature which was obtained by a time-consuming fine-tuning process.
Abstract: Brain-computer interface (BCI) is a system to translate humans thoughts into commands. For electroencephalography (EEG) based BCI, motor imagery is considered as one of the most effective ways. Different imagery activities can be classified based on the changes in mu and/or beta rhythms and their spatial distributions. However, the change in these rhythmic patterns varies from one subject to another. This causes an unavoidable time-consuming fine-tuning process in building a BCI for every subject. To address this issue, we propose a new method called sub-band common spatial pattern (SBCSP) to solve the problem. First, we decompose the EEG signals into sub-bands using a filter bank. Subsequently, we apply a discriminative analysis to extract SBCSP features. The SBCSP features are then fed into linear discriminant analyzers (LDA) to obtain scores which reflect the classification capability of each frequency band. Finally, the scores are fused to make decision. We evaluate two fusion methods: recursive band elimination (RBE) and meta-classifier (MC). We assess our approaches on a standard database from BCI Competition III. We also compare our method with two other approaches that address the same issue. The results show that our method outperforms the other two approaches and achieves similar result as compared to the best one in the literature which was obtained by a time-consuming fine-tuning process.

280 citations

Book ChapterDOI
26 Jun 2003
TL;DR: Two extensions of LLE to supervised feature extraction were independently proposed by the authors of this paper and are unified in a common framework and applied to a number of benchmark data sets.
Abstract: Locally linear embedding (LLE) is a recently proposed method for unsupervised nonlinear dimensionality reduction. It has a number of attractive features: it does not require an iterative algorithm, and just a few parameters need to be set. Two extensions of LLE to supervised feature extraction were independently proposed by the authors of this paper. Here, both methods are unified in a common framework and applied to a number of benchmark data sets. Results show that they perform very well on high-dimensional data which exhibits a manifold structure.

280 citations

Journal ArticleDOI
01 Dec 2000
TL;DR: Two related methods for merging classifiers are presented, one of which outperforms other methods for combining classifiers, like Stacked Generalization, and competes well against Boosting at statistically significant confidence levels.
Abstract: Using multiple classifiers for increasing learning accuracy is an active research area. In this paper we present two related methods for merging classifiers. The first method, Cascade Generalization, couples classifiers loosely. It belongs to the family of stacking algorithms. The basic idea of Cascade Generalization is to use sequentially the set of classifiers, at each step performing an extension of the original data by the insertion of new attributes. The new attributes are derived from the probability class distribution given by a base classifier. This constructive step extends the representational language for the high level classifiers, relaxing their bias. The second method exploits tight coupling of classifiers, by applying Cascade Generalization locally. At each iteration of a divide and conquer algorithm, a reconstruction of the instance space occurs by the addition of new attributes. Each new attribute represents the probability that an example belongs to a class given by a base classifier. We have implemented three Local Generalization Algorithms. The first merges a linear discriminant with a decision tree, the second merges a naive Bayes with a decision tree, and the third merges a linear discriminant and a naive Bayes with a decision tree. All the algorithms show an increase of performance, when compared with the corresponding single models. Cascade also outperforms other methods for combining classifiers, like Stacked Generalization, and competes well against Boosting at statistically significant confidence levels.

280 citations

Journal ArticleDOI
TL;DR: Three different techniques are used: Multivariate discriminant analysis, case-based forecasting, and neural network to predict Korean bankrupt and nonbankrupt firms, with good results.
Abstract: Bankruptcy prediction is one of the major business classification problems. 1n this paper, we use three different techniques: (1) Multivariate discriminant analysis, (2) case-based forecasting, and (3) neural network to predict Korean bankrupt and nonbankrupt firms. The average hit ratios of three methods range from 81.5 to 83.8%. Neural network performs better than discriminant analysis and the case-based forecasting system.

280 citations

Journal ArticleDOI
TL;DR: In this paper, the authors exploited environmental and multi-temporal landslide information for an area in Umbria, Italy, to produce four single and two combined landslide susceptibility zonations.

280 citations


Network Information
Related Topics (5)
Regression analysis
31K papers, 1.7M citations
85% related
Artificial neural network
207K papers, 4.5M citations
80% related
Feature extraction
111.8K papers, 2.1M citations
80% related
Cluster analysis
146.5K papers, 2.9M citations
79% related
Image segmentation
79.6K papers, 1.8M citations
79% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20251
20242
2023756
20221,711
2021678
2020815