Open AccessJournal Article
A gentle tutorial of the em algorithm and its application to parameter estimation for Gaussian mixture and hidden Markov models
Reads0
Chats0
TLDR
In this paper, the authors describe the EM algorithm for finding the parameters of a mixture of Gaussian densities and a hidden Markov model (HMM) for both discrete and Gaussian mixture observation models.Abstract:
We describe the maximum-likelihood parameter estimation problem and how the ExpectationMaximization (EM) algorithm can be used for its solution. We first describe the abstract form of the EM algorithm as it is often given in the literature. We then develop the EM parameter estimation procedure for two applications: 1) finding the parameters of a mixture of Gaussian densities, and 2) finding the parameters of a hidden Markov model (HMM) (i.e., the Baum-Welch algorithm) for both discrete and Gaussian mixture observation models. We derive the update equations in fairly explicit detail but we do not prove any convergence properties. We try to emphasize intuition rather than mathematical rigor.read more
Citations
More filters
Journal ArticleDOI
An Integrative Framework of Human Hand Gesture Segmentation for Human–Robot Interaction
TL;DR: A novel framework to segment hand gestures in RGB-depth (RGB-D) images captured by Kinect using humanlike approaches for human–robot interaction is proposed to reduce the error of Kinect sensing and, consequently, to improve the precision of hand gesture segmentation for robot NAO.
Journal ArticleDOI
Voice Activity Detection in Presence of Transient Noise Using Spectral Clustering
Saman Mousazadeh,Israel Cohen +1 more
TL;DR: This paper develops a novel VAD algorithm based on spectral clustering methods which is a supervised learning algorithm that divides the input signal into two separate clusters and demonstrates the advantage of the proposed method compared to conventional statistical model-based VAD algorithms in presence of transient noise.
Journal ArticleDOI
Implementing EM and Viterbi algorithms for Hidden Markov Model in linear memory
TL;DR: A memory sparse version of the Baum-Welch algorithm with modifications to the original probabilistic table topologies to make memory use independent of sequence length (and linearly dependent on state number) and a linear memory implementation of the Viterbi decoding algorithm.
Journal ArticleDOI
Partitive clustering ( K -means family)
TL;DR: A brief overview of clustering is given, well‐known partitional clustering methods are summarized, the major challenges and key issues of these methods are discussed, and simple numerical experiments using toy data sets are carried out to enhance the description of various clustering Methods.
Book ChapterDOI
EM Estimation of Scale Factor for Quantization-Based Audio Watermarking
TL;DR: A preprocessed decoding scheme, which uses an estimated scale factor from the received signal itself without any additional pilot signal, and results show that the proposed scheme provides a reliable decoding from the scaled signal.
References
More filters
Journal ArticleDOI
Maximum likelihood from incomplete data via the EM algorithm
Book
The Nature of Statistical Learning Theory
TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Statistical learning theory
TL;DR: Presenting a method for determining the necessary and sufficient conditions for consistency of learning process, the author covers function estimates from small data pools, applying these estimations to real-life problems, and much more.
Book
The Fractal Geometry of Nature
TL;DR: This book is a blend of erudition, popularization, and exposition, and the illustrations include many superb examples of computer graphics that are works of art in their own right.