scispace - formally typeset
Search or ask a question
Topic

Iterative Viterbi decoding

About: Iterative Viterbi decoding is a research topic. Over the lifetime, 1591 publications have been published within this topic receiving 32738 citations.


Papers
More filters
Journal ArticleDOI
01 Mar 1973
TL;DR: This paper gives a tutorial exposition of the Viterbi algorithm and of how it is implemented and analyzed, and increasing use of the algorithm in a widening variety of areas is foreseen.
Abstract: The Viterbi algorithm (VA) is a recursive optimal solution to the problem of estimating the state sequence of a discrete-time finite-state Markov process observed in memoryless noise. Many problems in areas such as digital communications can be cast in this form. This paper gives a tutorial exposition of the algorithm and of how it is implemented and analyzed. Applications to date are reviewed. Increasing use of the algorithm in a widening variety of areas is foreseen.

5,995 citations

Proceedings ArticleDOI
Michael Collins1
06 Jul 2002
TL;DR: Experimental results on part-of-speech tagging and base noun phrase chunking are given, in both cases showing improvements over results for a maximum-entropy tagger.
Abstract: We describe new algorithms for training tagging models, as an alternative to maximum-entropy models or conditional random fields (CRFs). The algorithms rely on Viterbi decoding of training examples, combined with simple additive updates. We describe theory justifying the algorithms through a modification of the proof of convergence of the perceptron algorithm for classification problems. We give experimental results on part-of-speech tagging and base noun phrase chunking, in both cases showing improvements over results for a maximum-entropy tagger.

2,221 citations

Proceedings ArticleDOI
27 Nov 1989
TL;DR: The Viterbi algorithm is modified to deliver the most likely path sequence in a finite-state Markov chain, as well as either the a posteriori probability for each bit or a reliability value, with the aim of producing soft decisions to be used in the decoding of outer codes.
Abstract: The Viterbi algorithm (VA) is modified to deliver the most likely path sequence in a finite-state Markov chain, as well as either the a posteriori probability for each bit or a reliability value. With this reliability indicator the modified VA produces soft decisions to be used in the decoding of outer codes. The inner software output Viterbi algorithm (SOVA) accepts and delivers soft sample values and can be regraded as a device for improving the signal-to-noise ratio, similar to an FM demodulator. Several applications are investigated to show the gain over the conventional hard-deciding VA, including concatenated convolutional codes, concatenation of trellis-coded modulation with convolutional FEC (forward error correcting) codes, and coded Viterbi equalization. For these applications additional gains of 1-4 dB as compared to the classical hard-deciding algorithms were found. For comparison, the more complex symbol-to-symbol MAP, whose optimal a posteriori probabilities can be transformed into soft outputs, was investigated. >

1,848 citations

01 Jan 1996
TL;DR: It is showed that many iterative decoding algorithms are special cases of two generic algorithms, the min-sum and sum-product algorithms, which also include non-iterative algorithms such as Viterbi decoding.
Abstract: Iterative decoding techniques have become a viable alternative for constructing high performance coding systems. In particular, the recent success of turbo codes indicates that performance close to the Shannon limit may be achieved. In this thesis, it is showed that many iterative decoding algorithms are special cases of two generic algorithms, the min-sum and sum-product algorithms, which also include non-iterative algorithms such as Viterbi decoding. The min-sum and sum-product algorithms are developed and presented as generalized trellis algorithms, where the time axis of the trellis is replaced by an arbitrary graph, the “Tanner graph”. With cycle-free Tanner graphs, the resulting decoding algorithms (e.g., Viterbi decoding) are maximum-likelihood but suffer from an exponentially increasing complexity. Iterative decoding occurs when the Tanner graph has cycles (e.g., turbo codes); the resulting algorithms are in general suboptimal, but significant complexity reductions are possible compared to the cycle-free case. Several performance estimates for iterative decoding are developed, including a generalization of the union bound used with Viterbi decoding and a characterization of errors that are uncorrectable after infinitely many decoding iterations.

1,044 citations

Journal ArticleDOI
TL;DR: A series of database filtering steps, HMMERHEAD, that are applied prior to the scoring algorithms, as implemented in the HMMer package, in an effort to reduce search time, significantly reduces the time needed to score a profile-HMM against large sequence databases.
Abstract: Profile hidden Markov models (profile-HMMs) are sensitive tools for remote protein homology detection, but the main scoring algorithms, Viterbi or Forward, require considerable time to search large sequence databases. We have designed a series of database filtering steps, HMMERHEAD, that are applied prior to the scoring algorithms, as implemented in the HMMER package, in an effort to reduce search time. Using this heuristic, we obtain a 20-fold decrease in Forward and a 6-fold decrease in Viterbi search time with a minimal loss in sensitivity relative to the unfiltered approaches. We then implemented an iterative profile-HMM search method, JackHMMER, which employs the HMMERHEAD heuristic. Due to our search heuristic, we eliminated the subdatabase creation that is common in current iterative profile-HMM approaches. On our benchmark, JackHMMER detects 14% more remote protein homologs than SAM's iterative method T2K. Our search heuristic, HMMERHEAD, significantly reduces the time needed to score a profile-HMM against large sequence databases. This search heuristic allowed us to implement an iterative profile-HMM search method, JackHMMER, which detects significantly more remote protein homologs than SAM's T2K and NCBI's PSI-BLAST.

890 citations


Network Information
Related Topics (5)
Wireless
133.4K papers, 1.9M citations
79% related
Network packet
159.7K papers, 2.2M citations
78% related
Wireless network
122.5K papers, 2.1M citations
78% related
Wireless ad hoc network
49K papers, 1.1M citations
78% related
Key distribution in wireless sensor networks
59.2K papers, 1.2M citations
76% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20233
202210
20182
201721
201634
201543