Bio: Simon Haykin is an academic researcher. The author has contributed to research in topics: Communications system & Communication theory. The author has an hindex of 1, co-authored 1 publications receiving 3054 citations.
01 Jan 1978
TL;DR: This best-selling, easy to read book offers the most complete discussion on the theories and principles behind today's most advanced communications systems.
Abstract: This best-selling, easy to read book offers the most complete discussion on the theories and principles behind today's most advanced communications systems. Throughout, Haykin emphasizes the statistical underpinnings of communication theory in a complete and detailed manner. Readers are guided though topics ranging from pulse modulation and passband digital transmission to random processes and error-control coding. The fifth edition has also been revised to include an extensive treatment of digital communications.
TL;DR: Following the discussion of interference temperature as a new metric for the quantification and management of interference, the paper addresses three fundamental cognitive tasks: radio-scene analysis, channel-state estimation and predictive modeling, and the emergent behavior of cognitive radio.
Abstract: Cognitive radio is viewed as a novel approach for improving the utilization of a precious natural resource: the radio electromagnetic spectrum. The cognitive radio, built on a software-defined radio, is defined as an intelligent wireless communication system that is aware of its environment and uses the methodology of understanding-by-building to learn from the environment and adapt to statistical variations in the input stimuli, with two primary objectives in mind: /spl middot/ highly reliable communication whenever and wherever needed; /spl middot/ efficient utilization of the radio spectrum. Following the discussion of interference temperature as a new metric for the quantification and management of interference, the paper addresses three fundamental cognitive tasks. 1) Radio-scene analysis. 2) Channel-state estimation and predictive modeling. 3) Transmit-power control and dynamic spectrum management. This work also discusses the emergent behavior of cognitive radio.
TL;DR: In this article, the capacity limit of fiber-optic communication systems (or fiber channels?) is estimated based on information theory and the relationship between the commonly used signal to noise ratio and the optical signal-to-noise ratio is discussed.
Abstract: We describe a method to estimate the capacity limit of fiber-optic communication systems (or ?fiber channels?) based on information theory. This paper is divided into two parts. Part 1 reviews fundamental concepts of digital communications and information theory. We treat digitization and modulation followed by information theory for channels both without and with memory. We provide explicit relationships between the commonly used signal-to-noise ratio and the optical signal-to-noise ratio. We further evaluate the performance of modulation constellations such as quadrature-amplitude modulation, combinations of amplitude-shift keying and phase-shift keying, exotic constellations, and concentric rings for an additive white Gaussian noise channel using coherent detection. Part 2 is devoted specifically to the "fiber channel.'' We review the physical phenomena present in transmission over optical fiber networks, including sources of noise, the need for optical filtering in optically-routed networks, and, most critically, the presence of fiber Kerr nonlinearity. We describe various transmission scenarios and impairment mitigation techniques, and define a fiber channel deemed to be the most relevant for communication over optically-routed networks. We proceed to evaluate a capacity limit estimate for this fiber channel using ring constellations. Several scenarios are considered, including uniform and optimized ring constellations, different fiber dispersion maps, and varying transmission distances. We further present evidences that point to the physical origin of the fiber capacity limitations and provide a comparison of recent record experiments with our capacity limit estimation.
TL;DR: The concept of matched filter detection of signals is used to detect piecewise linear segments of blood vessels in these images and the results are compared to those obtained with other methods.
Abstract: Blood vessels usually have poor local contrast, and the application of existing edge detection algorithms yield results which are not satisfactory. An operator for feature extraction based on the optical and spatial properties of objects to be recognized is introduced. The gray-level profile of the cross section of a blood vessel is approximated by a Gaussian-shaped curve. The concept of matched filter detection of signals is used to detect piecewise linear segments of blood vessels in these images. Twelve different templates that are used to search for vessel segments along all possible directions are constructed. Various issues related to the implementation of these matched filters are discussed. The results are compared to those obtained with other methods. >
TL;DR: It is shown that the performance of the new approaches to combining equalization based on linear filtering, with decoding is similar to the trellis-based receiver, while providing large savings in computational complexity.
Abstract: We study the turbo equalization approach to coded data transmission over channels with intersymbol interference. In the original system invented by Douillard et al. (1995), the data are protected by a convolutional code and the receiver consists of two trellis-based detectors, one for the channel (the equalizer) and one for the code (the decoder). It has been shown that iterating equalization and decoding tasks can yield tremendous improvements in bit error rate. We introduce new approaches to combining equalization based on linear filtering, with decoding.. Through simulation and analytical results, we show that the performance of the new approaches is similar to the trellis-based receiver, while providing large savings in computational complexity. Moreover, this paper provides an overview of the design alternatives for turbo equalization with given system parameters, such as the channel response or the signal-to-noise ratio.
TL;DR: This work explores a number of low-complexity soft-input/soft-output (SISO) equalization algorithms based on the minimum mean square error (MMSE) criterion and shows that for the turbo equalization application, the MMSE-based SISO equalizers perform well compared with a MAP equalizer while providing a tremendous complexity reduction.
Abstract: A number of important advances have been made in the area of joint equalization and decoding of data transmitted over intersymbol interference (ISI) channels. Turbo equalization is an iterative approach to this problem, in which a maximum a posteriori probability (MAP) equalizer and a MAP decoder exchange soft information in the form of prior probabilities over the transmitted symbols. A number of reduced-complexity methods for turbo equalization have been introduced in which MAP equalization is replaced with suboptimal, low-complexity approaches. We explore a number of low-complexity soft-input/soft-output (SISO) equalization algorithms based on the minimum mean square error (MMSE) criterion. This includes the extension of existing approaches to general signal constellations and the derivation of a novel approach requiring less complexity than the MMSE-optimal solution. All approaches are qualitatively analyzed by observing the mean-square error averaged over a sequence of equalized data. We show that for the turbo equalization application, the MMSE-based SISO equalizers perform well compared with a MAP equalizer while providing a tremendous complexity reduction.