Open AccessJournal Article
A gentle tutorial of the em algorithm and its application to parameter estimation for Gaussian mixture and hidden Markov models
Reads0
Chats0
TLDR
In this paper, the authors describe the EM algorithm for finding the parameters of a mixture of Gaussian densities and a hidden Markov model (HMM) for both discrete and Gaussian mixture observation models.Abstract:
We describe the maximum-likelihood parameter estimation problem and how the ExpectationMaximization (EM) algorithm can be used for its solution. We first describe the abstract form of the EM algorithm as it is often given in the literature. We then develop the EM parameter estimation procedure for two applications: 1) finding the parameters of a mixture of Gaussian densities, and 2) finding the parameters of a hidden Markov model (HMM) (i.e., the Baum-Welch algorithm) for both discrete and Gaussian mixture observation models. We derive the update equations in fairly explicit detail but we do not prove any convergence properties. We try to emphasize intuition rather than mathematical rigor.read more
Citations
More filters
Journal ArticleDOI
Tonic exploration governs both flexibility and lapses.
TL;DR: The results support the idea that exploratory noise contributes to lapses, affecting rule-based decision-making even when it has no strategic value, and suggest that one key mechanism for regulating exploration may be the depth of rule states.
Journal ArticleDOI
Likelihood-Based Estimation of Multidimensional Langevin Models and Its Application to Biomolecular Dynamics
Illia Horenko,Christof Schütte +1 more
TL;DR: The algorithmic realization combines hidden Markov models (for the unobserved jump process between the metastable states) with likelihood-based estimation of the parameters in the local linear SDEs based on discrete-time observations of the system.
Journal ArticleDOI
Hidden markov models, grammars, and biology: a tutorial
Shibaji Mukherjee,Sushmita Mitra +1 more
TL;DR: The basic algorithms and their application to analyzing biological sequences and modelling structures of bio-molecules like proteins and nucleic acids are discussed and a comparison of the different approaches are discussed.
Book ChapterDOI
Kernel Trick Embedded Gaussian Mixture Model
TL;DR: This paper presents a kernel trick embedded Gaussian Mixture Model (GMM), called kernel GMM, which can solve problems in probabilistic framework and tackle nonlinear problems better than the traditional GMM.
Journal ArticleDOI
Discrimination Between Pathological and Normal Voices Using GMM-SVM Approach
TL;DR: A Gaussian mixture model supervector kernel-support vector machine (GMM-SVM) classifier is compared with GMM classifier for the detection of voice pathology and it is found that a sustain vowel phonation can be classified as normal or pathological with an accuracy of 96.1%.
References
More filters
Journal ArticleDOI
Maximum likelihood from incomplete data via the EM algorithm
Book
The Nature of Statistical Learning Theory
TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Statistical learning theory
TL;DR: Presenting a method for determining the necessary and sufficient conditions for consistency of learning process, the author covers function estimates from small data pools, applying these estimations to real-life problems, and much more.
Book
The Fractal Geometry of Nature
TL;DR: This book is a blend of erudition, popularization, and exposition, and the illustrations include many superb examples of computer graphics that are works of art in their own right.