scispace - formally typeset
Search or ask a question

Showing papers by "Klaus-Robert Müller published in 2014"


Journal ArticleDOI
TL;DR: It is found that conventional representations of the input data, such as the Coulomb matrix, are not suitable for the training of learning machines in the case of periodic solids and proposes a novel crystal structure representation for which learning and competitive prediction accuracies become possible within an unrestricted class of spd systems of arbitrary unit-cell size.
Abstract: High-throughput density functional calculations of solids are highly time-consuming. As an alternative, we propose a machine learning approach for the fast prediction of solid-state properties. To achieve this, local spin-density approximation calculations are used as a training set. We focus on predicting the value of the density of electronic states at the Fermi energy. We find that conventional representations of the input data, such as the Coulomb matrix, are not suitable for the training of learning machines in the case of periodic solids. We propose a novel crystal structure representation for which learning and competitive prediction accuracies become possible within an unrestricted class of $spd$ systems of arbitrary unit-cell size.

415 citations


Journal ArticleDOI
01 Mar 2014
TL;DR: This study systematically compare linear and nonlinear regression techniques for an independent, simultaneous and proportional myoelectric control of wrist movements with two DoF and showed that KRR, a nonparametric statistical learning method, outperformed the other methods.
Abstract: In recent years the number of active controllable joints in electrically powered hand-prostheses has increased significantly. However, the control strategies for these devices in current clinical use are inadequate as they require separate and sequential control of each degree-of-freedom (DoF). In this study we systematically compare linear and nonlinear regression techniques for an independent, simultaneous and proportional myoelectric control of wrist movements with two DoF. These techniques include linear regression, mixture of linear experts (ME), multilayer-perceptron, and kernel ridge regression (KRR). They are investigated offline with electro-myographic signals acquired from ten able-bodied subjects and one person with congenital upper limb deficiency. The control accuracy is reported as a function of the number of electrodes and the amount and diversity of training data providing guidance for the requirements in clinical practice. The results showed that KRR, a nonparametric statistical learning method, outperformed the other methods. However, simple transformations in the feature space could linearize the problem, so that linear models could achieve similar performance as KRR at much lower computational costs. Especially ME, a physiologically inspired extension of linear regression represents a promising candidate for the next generation of prosthetic devices.

320 citations


Journal ArticleDOI
TL;DR: It is shown that the popular common spatial patterns (CSP) algorithm can be formulated as a divergence maximization problem and computed within this framework and unifies many of the recently proposed CSP variants in a principled manner.
Abstract: Controlling a device with a brain-computer interface requires extraction of relevant and robust features from high-dimensional electroencephalographic recordings Spatial filtering is a crucial step in this feature extraction process This paper reviews algorithms for spatial filter computation and introduces a general framework for this task based on divergence maximization We show that the popular common spatial patterns (CSP) algorithm can be formulated as a divergence maximization problem and computed within our framework Our approach easily permits enforcing different invariances and utilizing information from other subjects; thus, it unifies many of the recently proposed CSP variants in a principled manner Furthermore, it allows to design novel spatial filtering algorithms by incorporating regularization schemes into the optimization process or applying other divergences We evaluate the proposed approach using three regularization schemes, investigate the advantages of beta divergence, and show that subject-independent feature spaces can be extracted by jointly optimizing the divergence problems of multiple users We discuss the relations to several CSP variants and investigate the advantages and limitations of our approach with simulations Finally, we provide experimental results on a dataset containing recordings from 80 subjects and interpret the obtained patterns from a neurophysiological perspective

175 citations


Journal ArticleDOI
TL;DR: Taking into account the results of the simulations and real EEG recordings, SPoC represents an adequate approach for the optimal extraction of neuronal components showing coupling of power with continuously changing behaviorally relevant parameters.

118 citations


Journal ArticleDOI
TL;DR: This work proposes a probabilistic framework for BCI applications which exploit event-related potentials (ERPs) and shows how the framework harvests the structure suitable to solve the decoding task by transfer learning, unsupervised adaptation, language model and dynamic stopping.
Abstract: Objective Most BCIs have to undergo a calibration session in which data is recorded to train decoders with machine learning Only recently zero-training methods have become a subject of study This work proposes a probabilistic framework for BCI applications which exploit event-related potentials (ERPs) For the example of a visual P300 speller we show how the framework harvests the structure suitable to solve the decoding task by (a) transfer learning, (b) unsupervised adaptation, (c) language model and (d) dynamic stopping Approach A simulation study compares the proposed probabilistic zero framework (using transfer learning and task structure) to a state-of-the-art supervised model on n = 22 subjects The individual influence of the involved components (a)–(d) are investigated Main results Without any need for a calibration session, the probabilistic zero-training framework with inter-subject transfer learning shows excellent performance—competitive to a state-of-the-art supervised method using calibration Its decoding quality is carried mainly by the effect of transfer learning in combination with continuous unsupervised adaptation Significance A high-performing zero-training BCI is within reach for one of the most popular BCI paradigms: ERP spelling Recording calibration data for a supervised BCI would require valuable time which is lost for spelling The time spent on calibration would allow a novel user to spell 29 symbols with our unsupervised approach It could be of use for various clinical and non-clinical ERP-applications of BCI

98 citations


Journal ArticleDOI
10 Nov 2014-PLOS ONE
TL;DR: Stronger deflections of the ERPs in response to face stimuli are indicated, which are further enhanced when using the self-face images, and thereby improving P300-based spelling performance and leading to a significant reduction of stimulus sequences required for correct character classification.
Abstract: Event-related potential (ERP)-based P300 spellers are commonly used in the field of brain-computer interfaces as an alternative channel of communication for people with severe neuro-muscular diseases. This study introduces a novel P300 based brain-computer interface (BCI) stimulus paradigm using a random set presentation pattern and exploiting the effects of face familiarity. The effect of face familiarity is widely studied in the cognitive neurosciences and has recently been addressed for the purpose of BCI. In this study we compare P300-based BCI performances of a conventional row-column (RC)-based paradigm with our approach that combines a random set presentation paradigm with (non-) self-face stimuli. Our experimental results indicate stronger deflections of the ERPs in response to face stimuli, which are further enhanced when using the self-face images, and thereby improving P300-based spelling performance. This lead to a significant reduction of stimulus sequences required for correct character classification. These findings demonstrate a promising new approach for improving the speed and thus fluency of BCI-enhanced communication with the widely used P300-based BCI setup.

94 citations


Journal ArticleDOI
28 Jul 2014-PLOS ONE
TL;DR: The results demonstrate that the calibration recording can be bypassed by utilizing an unsupervised trained classifier, that is initialized randomly and updated during usage, that performs comparably to a classic supervised model.
Abstract: Despite several approaches to realize subject-to-subject transfer of pre-trained classifiers, the full performance of a Brain-Computer Interface (BCI) for a novel user can only be reached by presenting the BCI system with data from the novel user. In typical state-of-the-art BCI systems with a supervised classifier, the labeled data is collected during a calibration recording, in which the user is asked to perform a specific task. Based on the known labels of this recording, the BCI's classifier can learn to decode the individual's brain signals. Unfortunately, this calibration recording consumes valuable time. Furthermore, it is unproductive with respect to the final BCI application, e.g. text entry. Therefore, the calibration period must be reduced to a minimum, which is especially important for patients with a limited concentration ability. The main contribution of this manuscript is an online study on unsupervised learning in an auditory event-related potential (ERP) paradigm. Our results demonstrate that the calibration recording can be bypassed by utilizing an unsupervised trained classifier, that is initialized randomly and updated during usage. Initially, the unsupervised classifier tends to make decoding mistakes, as the classifier might not have seen enough data to build a reliable model. Using a constant re-analysis of the previously spelled symbols, these initially misspelled symbols can be rectified posthoc when the classifier has learned to decode the signals. We compare the spelling performance of our unsupervised approach and of the unsupervised posthoc approach to the standard supervised calibration-based dogma for n = 10 healthy users. To assess the learning behavior of our approach, it is unsupervised trained from scratch three times per user. Even with the relatively low SNR of an auditory ERP paradigm, the results show that after a limited number of trials (30 trials), the unsupervised approach performs comparably to a classic supervised model.

84 citations


Journal ArticleDOI
27 Aug 2014-PLOS ONE
TL;DR: This work presents a BCI system designed to establish external control for severely motor-impaired patients within a very short time and finds evidence that the BCI could outperform the best assistive technology of the patient in terms of control accuracy, reaction time and information transfer rate.
Abstract: Brain-Computer Interfaces (BCIs) strive to decode brain signals into control commands for severely handicapped people with no means of muscular control. These potential users of noninvasive BCIs display a large range of physical and mental conditions. Prior studies have shown the general applicability of BCI with patients, with the conflict of either using many training sessions or studying only moderately restricted patients. We present a BCI system designed to establish external control for severely motor-impaired patients within a very short time. Within only six experimental sessions, three out of four patients were able to gain significant control over the BCI, which was based on motor imagery or attempted execution. For the most affected patient, we found evidence that the BCI could outperform the best assistive technology (AT) of the patient in terms of control accuracy, reaction time and information transfer rate. We credit this success to the applied user-centered design approach and to a highly flexible technical setup. State-of-the art machine learning methods allowed the exploitation and combination of multiple relevant features contained in the EEG, which rapidly enabled the patients to gain substantial BCI control. Thus, we could show the feasibility of a flexible and tailorable BCI application in severely disabled users. This can be considered a significant success for two reasons: Firstly, the results were obtained within a short period of time, matching the tight clinical requirements. Secondly, the participating patients showed, compared to most other studies, very severe communication deficits. They were dependent on everyday use of AT and two patients were in a locked-in state. For the most affected patient a reliable communication was rarely possible with existing AT.

74 citations


Journal ArticleDOI
14 Feb 2014-PLOS ONE
TL;DR: A novel, fully Bayesian–and thereby probabilistic–framework, called Bayesian Spatio-Spectral Filter Optimization (BSSFO) is discussed and applied to a large data set of 80 non-invasive EEG-based BCI experiments, which allows to analyze which spatio-spectral parameters are common and which ones differ across the subject population.
Abstract: Recently, spatio-temporal filtering to enhance decoding for Brain-Computer-Interfacing (BCI) has become increasingly popular. In this work, we discuss a novel, fully Bayesian–and thereby probabilistic–framework, called Bayesian Spatio-Spectral Filter Optimization (BSSFO) and apply it to a large data set of 80 non-invasive EEG-based BCI experiments. Across the full frequency range, the BSSFO framework allows to analyze which spatio-spectral parameters are common and which ones differ across the subject population. As expected, large variability of brain rhythms is observed between subjects. We have clustered subjects according to similarities in their corresponding spectral characteristics from the BSSFO model, which is found to reflect their BCI performances well. In BCI, a considerable percentage of subjects is unable to use a BCI for communication, due to their missing ability to modulate their brain rhythms–a phenomenon sometimes denoted as BCI-illiteracy or inability. Predicting individual subjects’ performance preceding the actual, time-consuming BCI-experiment enhances the usage of BCIs, e.g., by detecting users with BCI inability. This work additionally contributes by using the novel BSSFO method to predict the BCI-performance using only 2 minutes and 3 channels of resting-state EEG data recorded before the actual BCI-experiment. Specifically, by grouping the individual frequency characteristics we have nicely classified them into the subject ‘prototypes’ (like μ - or β -rhythm type subjects) or users without ability to communicate with a BCI, and then by further building a linear regression model based on the grouping we could predict subjects' performance with the maximum correlation coefficient of 0.581 with the performance later seen in the actual BCI session.

64 citations


Journal ArticleDOI
TL;DR: This work presents a source separation approach using spatial filters which maximize the correlation between the envelopes of brain oscillations recorded with electro-/magnetoencephalography or intracranial multichannel recordings, and demonstrates excellent unsupervised discovery of meaningful power-to-power couplings within as well as across subjects and frequency bands.

55 citations


Journal ArticleDOI
TL;DR: It is demonstrated that increases in intersubject correlations of brain networks can serve as neurophysiological marker for stereoscopic depth and for the strength of the viewing experience.

Journal ArticleDOI
TL;DR: This simulation study demonstrates how the quality of cerebral DOT reconstructions is altered with the choice of the regularization parameter for different methods and proposes a cross-validation procedure which achieves a reconstruction quality close to the optimum.
Abstract: Functional near-infrared spectroscopy (fNIRS) is an optical method for noninvasively determining brain activation by estimating changes in the absorption of near-infrared light. Diffuse optical tomography (DOT) extends fNIRS by applying overlapping “high density” measurements, and thus providing a three-dimensional imaging with an improved spatial resolution. Reconstructing brain activation images with DOT requires solving an underdetermined inverse problem with far more unknowns in the volume than in the surface measurements. All methods of solving this type of inverse problem rely on regularization and the choice of corresponding regularization or convergence criteria. While several regularization methods are available, it is unclear how well suited they are for cerebral functional DOT in a semi-infinite geometry. Furthermore, the regularization parameter is often chosen without an independent evaluation, and it may be tempting to choose the solution that matches a hypothesis and rejects the other. In this simulation study, we start out by demonstrating how the quality of cerebral DOT reconstructions is altered with the choice of the regularization parameter for different methods. To independently select the regularization parameter, we propose a cross-validation procedure which achieves a reconstruction quality close to the optimum. Additionally, we compare the outcome of seven different image reconstruction methods for cerebral functional DOT. The methods selected include reconstruction procedures that are already widely used for cerebral DOT [minimum l2-norm estimate (l2MNE) and truncated singular value decomposition], recently proposed sparse reconstruction algorithms [minimum l1- and a smooth minimum l0-norm estimate (l1MNE, l0MNE, respectively)] and a depth- and noise-weighted minimum norm (wMNE). Furthermore, we expand the range of algorithms for DOT by adapting two EEG-source localization algorithms [sparse basis field expansions and linearly constrained minimum variance (LCMV) beamforming]. Independent of the applied noise level, we find that the LCMV beamformer is best for single spot activations with perfect location and focality of the results, whereas the minimum l1-norm estimate succeeds with multiple targets.

Journal ArticleDOI
TL;DR: LASSO is a promising channel selection method for accurate simultaneous and proportional prosthesis control and will provide a useful guideline to select optimal channel subsets when developing clinical myoelectric prostheses control systems based on continuous movements with multiple DoFs.
Abstract: Objective. Recent studies have shown the possibility of simultaneous and proportional control of electrically powered upper-limb prostheses, but there has been little investigation on optimal channel selection. The objective of this study is to find a robust channel selection method and the channel subsets most suitable for simultaneous and proportional myoelectric prosthesis control of multiple degrees-of-freedom (DoFs). Approach. Ten able-bodied subjects and one person with congenital upper-limb deficiency took part in this study, and performed wrist movements with various combinations of two DoFs (flexion/extension and radial/ulnar deviation). During the experiment, high density electromyographic (EMG) signals and the actual wrist angles were recorded with an 8 × 24 electrode array and a motion tracking system, respectively. The wrist angles were estimated from EMG features with ridge regression using the subsets of channels chosen by three different channel selection methods: (1) least absolute shrinkage and selection operator (LASSO), (2) sequential feature selection (SFS), and (3) uniform selection (UNI). Main results. SFS generally showed higher estimation accuracy than LASSO and UNI, but LASSO always outperformed SFS in terms of robustness, such as noise addition, channel shift and training data reduction. It was also confirmed that about 95% of the original performance obtained using all channels can be retained with only 12 bipolar channels individually selected by LASSO and SFS. Significance. From the analysis results, it can be concluded that LASSO is a promising channel selection method for accurate simultaneous and proportional prosthesis control. We expect that our results will provide a useful guideline to select optimal channel subsets when developing clinical myoelectric prosthesis control systems based on continuous movements with multiple DoFs.

Journal ArticleDOI
TL;DR: This work presents a way to robustify the popular common spatial patterns (CSP) algorithm under a maxmin approach and proposes to robustly compute spatial filters by maximizing the minimum variance ratio within a prefixed set of covariance matrices called the tolerance set.
Abstract: Electroencephalographic signals are known to be nonstationary and easily affected by artifacts; therefore, their analysis requires methods that can deal with noise. In this work, we present a way to robustify the popular common spatial patterns CSP algorithm under a maxmin approach. In contrast to standard CSP that maximizes the variance ratio between two conditions based on a single estimate of the class covariance matrices, we propose to robustly compute spatial filters by maximizing the minimum variance ratio within a prefixed set of covariance matrices called the tolerance set. We show that this kind of maxmin optimization makes CSP robust to outliers and reduces its tendency to overfit. We also present a data-driven approach to construct a tolerance set that captures the variability of the covariance matrices over time and shows its ability to reduce the nonstationarity of the extracted features and significantly improve classification accuracy. We test the spatial filters derived with this approach and compare them to standard CSP and a state-of-the-art method on a real-world brain-computer interface BCI data set in which we expect substantial fluctuations caused by environmental differences. Finally we investigate the advantages and limitations of the maxmin approach with simulations.

Journal ArticleDOI
TL;DR: The findings imply that the long-range correlative properties of the EEG should be studied in source space, in such a way that the SNR is maximized, or at least with spatial decomposition techniques approximating source activities, rather than in sensor space.

Posted Content
TL;DR: In this article, kernel ridge regression is used to approximate the kinetic energy of non-interacting fermions in a one-dimensional box as a functional of their density, and the properties of different kernels and methods of cross-validation are explored.
Abstract: Kernel ridge regression is used to approximate the kinetic energy of non-interacting fermions in a one-dimensional box as a functional of their density. The properties of different kernels and methods of cross-validation are explored, and highly accurate energies are achieved. Accurate {\em constrained optimal densities} are found via a modified Euler-Lagrange constrained minimization of the total energy. A projected gradient descent algorithm is derived using local principal component analysis. Additionally, a sparse grid representation of the density can be used without degrading the performance of the methods. The implications for machine-learned density functional approximations are discussed.

Proceedings ArticleDOI
01 Oct 2014
TL;DR: A novel method is presented for assessing how image quality is processed on a neural level, using Steady-State Visual Evoked Potentials (SSVEPs) as EEG features, and yields a correlation of |r| = 0.93 to MOS on the recorded dataset.
Abstract: Conventionally, the quality of images and related codecs are assessed using subjective tests, such as Degradation Category Rating. These quality assessments consider the behavioral level only. Recently, it has been proposed to complement this approach by investigating how quality is processed in the brain of a user (using electroencephalography, EEG), potentially leading to results that are less biased by subjective factors. In this paper, a novel method is presented for assessing how image quality is processed on a neural level, using Steady-State Visual Evoked Potentials (SSVEPs) as EEG features. We tested our approach in an EEG study with 16 participants who were presented with distorted images of natural textures. Subsequently, we compared our approach analogously to the standardized Degradation Category Rating quality assessment. Remarkably, our novel method yields a correlation of |r| = 0.93 to MOS on the recorded dataset.

Proceedings Article
08 Dec 2014
TL;DR: It is shown that the Sancetta estimator, while being consistent in the high-dimensional limit, suffers from a high bias in finite sample sizes, and an alternative estimator is proposed, which is unbiased, less sensitive to hyperparameter choice and yields superior performance in simulations on toy data and on a real world data set from an EEG-based Brain-Computer-Interfacing experiment.
Abstract: The accurate estimation of covariance matrices is essential for many signal processing and machine learning algorithms. In high dimensional settings the sample covariance is known to perform poorly, hence regularization strategies such as analytic shrinkage of Ledoit/Wolf are applied. In the standard setting, i.i.d. data is assumed, however, in practice, time series typically exhibit strong autocorrelation structure, which introduces a pronounced estimation bias. Recent work by Sancetta has extended the shrinkage framework beyond i.i.d. data. We contribute in this work by showing that the Sancetta estimator, while being consistent in the high-dimensional limit, suffers from a high bias in finite sample sizes. We propose an alternative estimator, which is (1) unbiased, (2) less sensitive to hyperparameter choice and (3) yields superior performance in simulations on toy data and on a real world data set from an EEG-based Brain-Computer-Interfacing experiment.

Journal ArticleDOI
TL;DR: Focusing on the task of label sequence learning, this work defines a general framework that can handle a large class of inference problems based on Hamming-like loss functions and the concept of decomposability for the underlying joint feature map.
Abstract: The task of structured output prediction deals with learning general functional dependencies between arbitrary input and output spaces. In this context, two loss-sensitive formulations for maximum-margin training have been proposed in the literature, which are referred to as margin and slack rescaling, respectively. The latter is believed to be more accurate and easier to handle. Nevertheless, it is not popular due to the lack of known efficient inference algorithms; therefore, margin rescaling - which requires a similar type of inference as normal structured prediction - is the most often used approach. Focusing on the task of label sequence learning, we here define a general framework that can handle a large class of inference problems based on Hamming-like loss functions and the concept of decomposability for the underlying joint feature map. In particular, we present an efficient generic algorithm that can handle both rescaling approaches and is guaranteed to find an optimal solution in polynomial time.

Proceedings ArticleDOI
03 Apr 2014
TL;DR: This study proposes a method of finding optimal threshold of canonical correlation analysis based steady state visual evoked potentials classification for detecting resting state and reducing misclassification and successfully found optimal threshold for the best performance.
Abstract: Brain-machine interfaces (BMIs) are systems that establish a direct connection between the human brain and a machine. These systems are applicable to neuro-rehabilitation. In this study, we propose a method of finding optimal threshold of canonical correlation analysis (CCA) based steady state visual evoked potentials (SSVEPs) classification for detecting resting state and reducing misclassification. As a result, we successfully found optimal threshold for the best performance. This result shows the possibility of SSVEP based exoskeleton online control with a proposed method.

Proceedings ArticleDOI
06 Nov 2014
TL;DR: The results showed that an estimator that shrinks the training model parameters towards the calibration set parameters significantly increased the classifier performance across different testing days, indicating that the proposed methodology can be a practical means for improving robustness in pattern recognition methods for myocontrol.
Abstract: Ensuring robustness of myocontrol algorithms for prosthetic devices is an important challenge. Robustness needs to be maintained under nonstationarities, e.g. due to electrode shifts after donning and doffing, sweating, additional weight or varying arm positions. Such nonstationary behavior changes the signal distributions - a scenario often referred to as covariate shift. This circumstance causes a significant decrease in classification accuracy in daily life applications. Re-training is possible but it is time consuming since it requires a large number of trials. In this paper, we propose to adapt the EMG classifier by a small calibration set only, which is able to capture the relevant aspects of the nonstationarities, but requires re-training data of only very short duration. We tested this strategy on signals acquired across 5 days in able-bodied individuals. The results showed that an estimator that shrinks the training model parameters towards the calibration set parameters significantly increased the classifier performance across different testing days. Even when using only one trial per class as re-training data for each day, the classification accuracy remained > 92% over five days. These results indicate that the proposed methodology can be a practical means for improving robustness in pattern recognition methods for myocontrol.

Proceedings ArticleDOI
04 Jun 2014
TL;DR: This work proposes a novel classification approach which exploits subclass-specific features using mean shrinkage and shows that this novel approach outperforms the standard LDA approach, while being computationally highly efficient.
Abstract: Linear discriminant analysis (LDA) is the most commonly used classification method for single trial data in a brain-computer interface (BCI) framework. The popularity of LDA arises from its robustness, simplicity and high accuracy. However, the standard LDA approach is not capable to exploit sublabel information (such as stimulus identity), which is accessible in data from event related potentials (ERPs): it assumes that the evoked potentials are independent of the stimulus identity and dependent only on the users' attentional state. We question this assumption and investigate several methods which extract subclass-specific features from ERP data. Moreover, we propose a novel classification approach which exploits subclass-specific features using mean shrinkage. Based on a reanalysis of two BCI data sets, we show that our novel approach outperforms the standard LDA approach, while being computationally highly efficient.

Posted Content
TL;DR: In this article, the authors extend the shrinkage concept by allowing simultaneous shrinkage to a set of targets, and derive conditions under which the estimation of the shrinkag intensities yields optimal expected squared error in the limit.
Abstract: Stein showed that the multivariate sample mean is outperformed by "shrinking" to a constant target vector. Ledoit and Wolf extended this approach to the sample covariance matrix and proposed a multiple of the identity as shrinkage target. In a general framework, independent of a specific estimator, we extend the shrinkage concept by allowing simultaneous shrinkage to a set of targets. Application scenarios include settings with (A) additional data sets from potentially similar distributions, (B) non-stationarity, (C) a natural grouping of the data or (D) multiple alternative estimators which could serve as targets. We show that this Multi-Target Shrinkage can be translated into a quadratic program and derive conditions under which the estimation of the shrinkage intensities yields optimal expected squared error in the limit. For the sample mean and the sample covariance as specific instances, we derive conditions under which the optimality of MTS is applicable. We consider two asymptotic settings: the large dimensional limit (LDL), where the dimensionality and the number of observations go to infinity at the same rate, and the finite observations large dimensional limit (FOLDL), where only the dimensionality goes to infinity while the number of observations remains constant. We then show the effectiveness in extensive simulations and on real world data.

Proceedings ArticleDOI
03 Apr 2014
TL;DR: It is shown that the spatial filter computation in BCI can be cast into an information geometric framework based on divergence maximization, which allows to integrate many of the recently proposed CSP algorithms in a principled manner, but also enables to easily develop novel CSP variants with different properties.
Abstract: Algorithms using concepts from information geometry have recently become very popular in machine learning and signal processing. These methods not only have a solid mathematical foundation but they also allow to interpret the optimization process and the solution from a geometric perspective. In this paper we apply information geometry to Brain-Computer Interfacing (BCI). More precisely, we show that the spatial filter computation in BCI can be cast into an information geometric framework based on divergence maximization. This formulation not only allows to integrate many of the recently proposed CSP algorithms in a principled manner, but also enables us to easily develop novel CSP variants with different properties. We evaluate the potentials of our information geometric framework on a data set containing recordings from 80 subjects.

Posted ContentDOI
01 Feb 2014
TL;DR: The first long-term aerosol sampling and chemical characterization results from measurements at the Cape Verde Atmospheric Observatory (CVAO) on the island of São Vicente are presented in this paper.
Abstract: 11 The first long-term aerosol sampling and chemical characterization results from measurements at 12 the Cape Verde Atmospheric Observatory (CVAO) on the island of São Vicente are presented 13 and are discussed with respect to air mass origin and seasonal trends. In total 671 samples were 14 collected using a high volume PM10 sampler on quartz fiber filters from January 2007 to 15 December 2011. The samples were analyzed for their aerosol chemical composition including 16 their ionic and organic constituents. Back trajectory analyses showed that the aerosol at CVAO 17 was strongly influenced by emissions from Europe and Africa with the later often responsible for 18 high mineral dust loading. Sea salt and mineral dust dominated the aerosol mass and made up in 19 total about 80% of the aerosol mass. The 5 year PM10 mean was 47.1 ± 55.5 μg/m3 while the 20 mineral dust and sea salt means were 27.9 ± 48.7 μg/m3 and 11.1 ± 5.5 μg/m3, respectively. Non21 sea-salt (nss) sulfate made up 62 % of the total sulfate and originated from both long range 22 transport from Africa or Europe and marine sources. Strong seasonal variation was observed for 23 the aerosol components. While nitrate showed no clear seasonal variation with an annual mean of 24 1.1 ± 0.6 μg/m3, the aerosol mass, OC and EC, showed strong winter maxima due to strong 25 influence of African air mass inflow. Additionally during summer, elevated concentrations of 26 OM were observed originating from marine emissions. A summer maximum was observed for 27 non-sea-salt sulfate and was connected to periods when air mass inflow was predominantly of 28 2 marine origin indicating that marine biogenic emissions were a significant source. Ammonium 29 showed a distinct maximum in spring and coincided with ocean surface water chlorophyll a 30 concentrations. Good correlations were also observed between nss-sulfate and oxalate during the 31 summer and winter seasons indicating a likely photochemical in-cloud processing of the marine 32 and anthropogenic precursors of these species. High temporal variability was observed in both 33 chloride and bromide depletion differing significantly within the seasons, air mass history and 34 Saharan dust concentration. Chloride (bromide) depletion varied from 8.8 ± 8.5 % (62 ± 42 %) in 35 Saharan dust dominated air mass to 30 ± 12 % (87 ± 11 %) in polluted Europe air masses. During 36 summer, bromide depletion often reached 100 % in marine as well as in polluted continental 37 samples. In addition to the influence of the aerosol acidic components, photochemistry was one 38 of the main drivers of halogenide depletion during the summer while during dust events, 39 displacement reaction with nitric acid was found to be the dominant mechanism. PMF analysis 40 identified three major aerosol sources including sea salt, aged sea salt and long range transport. 41 The ionic budget was dominated by the first two of these factors while the long range transport 42 factor could only account for about 14 % of the total observed ionic mass. 43

Proceedings ArticleDOI
03 Apr 2014
TL;DR: From the analysis results, it is confirmed that the LASSO method can be used to select reasonable electrode subsets for regression based myoelectric prosthesis control.
Abstract: To develop a clinically available prosthesis based on electromyography (EMG) signals, the number of recording electrodes should be as small as possible. In this study, we investigate the possibility of the least absolute shrinkage and selection operator (LASSO) for finding electrode subsets suitable for regression based myoelectric prosthesis control. EMG signals were recorded using 192 electrodes while ten subjects were performing two degree-of-freedom (DoF) wrist movements. Among the whole channels, we selected subsets consisting of 96, 64, 48, 32, 24, 16, 12, and 8 electrodes, respectively, using the LASSO method. As a baseline method, electrode subsets having the same numbers of electrodes were arbitrary selected with regular spacing (uniform selection method). The performance of decoding the movements was estimated using the r-square value. The electrode subsets selected by the LASSO method generally outperformed those chosen by the arbitrary selection method. In particular, the performance of the LASSO method was significantly higher than that of the arbitrary selection method when using the subsets of 8 electrodes. From the analysis results, we could confirm that the LASSO method can be used to select reasonable electrode subsets for regression based myoelectric prosthesis control.

Proceedings ArticleDOI
03 Apr 2014
TL;DR: A novel approach for measuring label noise and removing structured label noise is presented, which shows its usefulness for an EEG data set recorded during a standard d2 test for visual attention and is presented as a remedy to handle this crucial problem.
Abstract: Conventionally, neuroscientific data is analyzed based on the behavioral response of the participant. This approach assumes that behavioral errors of participants are in line with the neural processing. However, this may not be the case, in particular in experiments with time pressure or studies investigating the threshold of perception. In these cases, the error distribution deviates from uniformity due to the heteroscedastic nature of the underlying experimental set-up. This problem of systematic and structured (non-uniform) label noise is ignored when analysis are based on behavioral data, as is being done typically. Thus, we run the risk to arrive at wrong conclusions in our analysis. This paper proposes a remedy to handle this crucial problem: we present a novel approach for a) measuring label noise and b) removing structured label noise. We show its usefulness for an EEG data set recorded during a standard d2 test for visual attention.

Proceedings ArticleDOI
04 Jun 2014
TL;DR: This work presents data-driven analysis strategies that help to obtain interpretable results from multisubject neuroimaging data when complex movie stimuli are used and enable localization and visualization of brain activity using standard statistical parametric maps.
Abstract: A central question in neuroscience is how the brain reacts to real world sensory stimuli. Naturalistic and complex (e.g. movie) stimuli are increasingly used in empirical research but their analysis often relies on considerable human efforts to label or extract stimulus features. Here we present data-driven analysis strategies that help to obtain interpretable results from multisubject neuroimaging data when complex movie stimuli are used. These analyses a) enable localization and visualization of brain activity using standard statistical parametric maps in the subspace of brain activity shared between subjects and b) facilitate interpretation of intersubject correlations. We show experimental results obtained from 50 subjects.

Journal ArticleDOI
TL;DR: Advances in brain–computer interfacing technology suggest future possibilities for direct communication between brains and accurate techniques for decoding brain signals and stimulating brain structures shed light on the feasibility of brain-to-brain coupling.

Posted Content
TL;DR: A novel framework for incorporating algebraic invariance structure into kernels is added by showing that algebraic properties such as sign symmetries in data, phase independence, scaling etc. can be included easily by essentially performing the kernel trick twice.
Abstract: When solving data analysis problems it is important to integrate prior knowledge and/or structural invariances. This paper contributes by a novel framework for incorporating algebraic invariance structure into kernels. In particular, we show that algebraic properties such as sign symmetries in data, phase independence, scaling etc. can be included easily by essentially performing the kernel trick twice. We demonstrate the usefulness of our theory in simulations on selected applications such as sign-invariant spectral clustering and underdetermined ICA.