scispace - formally typeset
Search or ask a question

Showing papers on "Image processing published in 2002"


Journal ArticleDOI
TL;DR: Although the new index is mathematically defined and no human visual system model is explicitly employed, experiments on various image distortion types indicate that it performs significantly better than the widely used distortion metric mean squared error.
Abstract: We propose a new universal objective image quality index, which is easy to calculate and applicable to various image processing applications. Instead of using traditional error summation methods, the proposed index is designed by modeling any image distortion as a combination of three factors: loss of correlation, luminance distortion, and contrast distortion. Although the new index is mathematically defined and no human visual system model is explicitly employed, our experiments on various image distortion types indicate that it performs significantly better than the widely used distortion metric mean squared error. Demonstrative images and an efficient MATLAB implementation of the algorithm are available online at http://anchovy.ece.utexas.edu//spl sim/zwang/research/quality_index/demo.html.

5,285 citations


Journal ArticleDOI
TL;DR: This work built on another training-based super- resolution algorithm and developed a faster and simpler algorithm for one-pass super-resolution that requires only a nearest-neighbor search in the training set for a vector derived from each patch of local image data.
Abstract: We call methods for achieving high-resolution enlargements of pixel-based images super-resolution algorithms. Many applications in graphics or image processing could benefit from such resolution independence, including image-based rendering (IBR), texture mapping, enlarging consumer photographs, and converting NTSC video content to high-definition television. We built on another training-based super-resolution algorithm and developed a faster and simpler algorithm for one-pass super-resolution. Our algorithm requires only a nearest-neighbor search in the training set for a vector derived from each patch of local image data. This one-pass super-resolution algorithm is a step toward achieving resolution independence in image-based representations. We don't expect perfect resolution independence-even the polygon representation doesn't have that-but increasing the resolution independence of pixel-based representations is an important task for IBR.

2,576 citations


Journal ArticleDOI
TL;DR: A new technique for the display of high-dynamic-range images, which reduces the contrast while preserving detail, is presented, based on a two-scale decomposition of the image into a base layer.
Abstract: We present a new technique for the display of high-dynamic-range images, which reduces the contrast while preserving detail. It is based on a two-scale decomposition of the image into a base layer,...

1,715 citations


Proceedings ArticleDOI
01 Jul 2002
TL;DR: A new technique for the display of high-dynamic-range images, which reduces the contrast while preserving detail, is presented, based on a two-scale decomposition of the image into a base layer, encoding large-scale variations, and a detail layer.
Abstract: We present a new technique for the display of high-dynamic-range images, which reduces the contrast while preserving detail. It is based on a two-scale decomposition of the image into a base layer, encoding large-scale variations, and a detail layer. Only the base layer has its contrast reduced, thereby preserving detail. The base layer is obtained using an edge-preserving filter called the bilateral filter. This is a non-linear filter, where the weight of each pixel is computed using a Gaussian in the spatial domain multiplied by an influence function in the intensity domain that decreases the weight of pixels with large intensity differences. We express bilateral filtering in the framework of robust statistics and show how it relates to anisotropic diffusion. We then accelerate bilateral filtering by using a piecewise-linear approximation in the intensity domain and appropriate subsampling. This results in a speed-up of two orders of magnitude. The method is fast and requires no parameter setting.

1,612 citations


Book
01 Sep 2002
TL;DR: This volume unifies and extends the theories of adaptive blind signal and image processing and provides practical and efficient algorithms for blind source separation, Independent, Principal, Minor Component Analysis, and Multichannel Blind Deconvolution (MBD) and Equalization.
Abstract: From the Publisher: With solid theoretical foundations and numerous potential applications, Blind Signal Processing (BSP) is one of the hottest emerging areas in Signal Processing This volume unifies and extends the theories of adaptive blind signal and image processing and provides practical and efficient algorithms for blind source separation, Independent, Principal, Minor Component Analysis, and Multichannel Blind Deconvolution (MBD) and Equalization Containing over 1400 references and mathematical expressions Adaptive Blind Signal and Image Processing delivers an unprecedented collection of useful techniques for adaptive blind signal/image separation, extraction, decomposition and filtering of multi-variable signals and data Offers a broad coverage of blind signal processing techniques and algorithms both from a theoretical and practical point of viewPresents more than 50 simple algorithms that can be easily modified to suit the reader's specific real world problemsProvides a guide to fundamental mathematics of multi-input, multi-output and multi-sensory systemsIncludes illustrative worked examples, computer simulations, tables, detailed graphs and conceptual models within self contained chapters to assist self studyAccompanying CD-ROM features an electronic, interactive version of the book with fully coloured figures and text C and MATLAB user-friendly software packages are also provided MATLAB is a registered trademark of The MathWorks, Inc By providing a detailed introduction to BSP, as well as presenting new results and recent developments, this informative and inspiring work will appeal to researchers, postgraduate students, engineers and scientists working in biomedical engineering, communications, electronics, computer science, optimisations, finance, geophysics and neural networks

1,578 citations


Journal ArticleDOI
TL;DR: Results indicating that querying for images using Blobworld produces higher precision than does querying using color and texture histograms of the entire image in cases where the image contains distinctive objects are presented.
Abstract: Retrieving images from large and varied collections using image content as a key is a challenging and important problem We present a new image representation that provides a transformation from the raw pixel data to a small set of image regions that are coherent in color and texture This "Blobworld" representation is created by clustering pixels in a joint color-texture-position feature space The segmentation algorithm is fully automatic and has been run on a collection of 10,000 natural images We describe a system that uses the Blobworld representation to retrieve images from this collection An important aspect of the system is that the user is allowed to view the internal representation of the submitted image and the query results Similar systems do not offer the user this view into the workings of the system; consequently, query results from these systems can be inexplicable, despite the availability of knobs for adjusting the similarity metrics By finding image regions that roughly correspond to objects, we allow querying at the level of objects rather than global image properties We present results indicating that querying for images using Blobworld produces higher precision than does querying using color and texture histograms of the entire image in cases where the image contains distinctive objects

1,574 citations


Proceedings ArticleDOI
01 Jul 2002
TL;DR: The results demonstrate that the method is capable of drastic dynamic range compression, while preserving fine details and avoiding common artifacts, such as halos, gradient reversals, or loss of local contrast.
Abstract: We present a new method for rendering high dynamic range images on conventional displays. Our method is conceptually simple, computationally efficient, robust, and easy to use. We manipulate the gradient field of the luminance image by attenuating the magnitudes of large gradients. A new, low dynamic range image is then obtained by solving a Poisson equation on the modified gradient field. Our results demonstrate that the method is capable of drastic dynamic range compression, while preserving fine details and avoiding common artifacts, such as halos, gradient reversals, or loss of local contrast. The method is also able to significantly enhance ordinary images by bringing out detail in dark regions.

1,441 citations


Book
01 Jan 2002
TL;DR: Find the secret to improve the quality of life by reading this adaptive blind signal and image processing and make the words as your good value to your life.
Abstract: Find the secret to improve the quality of life by reading this adaptive blind signal and image processing. This is a kind of book that you need now. Besides, it can be your favorite book to read after having this book. Do you ask why? Well, this is a book that has different characteristic with others. You may not need to know who the author is, how well-known the work is. As wise word, never judge the words from who speaks, but make the words as your good value to your life.

1,425 citations


Journal ArticleDOI
TL;DR: This work focuses on detection algorithms that assume multivariate normal distribution models for HSI data and presents some results which illustrate the performance of some detection algorithms using real hyperspectral imaging (HSI) data.
Abstract: We introduce key concepts and issues including the effects of atmospheric propagation upon the data, spectral variability, mixed pixels, and the distinction between classification and detection algorithms. Detection algorithms for full pixel targets are developed using the likelihood ratio approach. Subpixel target detection, which is more challenging due to background interference, is pursued using both statistical and subspace models for the description of spectral variability. Finally, we provide some results which illustrate the performance of some detection algorithms using real hyperspectral imaging (HSI) data. Furthermore, we illustrate the potential deviation of HSI data from normality and point to some distributions that may serve in the development of algorithms with better or more robust performance. We therefore focus on detection algorithms that assume multivariate normal distribution models for HSI data.

1,170 citations


Journal ArticleDOI
TL;DR: A new approach is presented for elastic registration of medical images, and is applied to magnetic resonance images of the brain, where it results in accurate superposition of image data from individuals with significant anatomical differences.
Abstract: A new approach is presented for elastic registration of medical images, and is applied to magnetic resonance images of the brain. Experimental results demonstrate very high accuracy in superposition of images from different subjects. There are two major novelties in the proposed algorithm. First, it uses an attribute vector, i.e., a set of geometric moment invariants (GMIs) that are defined on each voxel in an image and are calculated from the tissue maps, to reflect the underlying anatomy at different scales. The attribute vector, if rich enough, can distinguish between different parts of an image, which helps establish anatomical correspondences in the deformation procedure; it also helps reduce local minima, by reducing ambiguity in potential matches. This is a fundamental deviation of our method, referred to as the hierarchical attribute matching mechanism for elastic registration (HAMMER), from other volumetric deformation methods, which are typically based on maximizing image similarity. Second, in order to avoid being trapped by local minima, i.e., suboptimal poor matches, HAMMER uses a successive approximation of the energy function being optimized by lower dimensional smooth energy functions, which are constructed to have significantly fewer local minima. This is achieved by hierarchically selecting the driving features that have distinct attribute vectors, thus, drastically reducing ambiguity in finding correspondence. A number of experiments demonstrate that the proposed algorithm results in accurate superposition of image data from individuals with significant anatomical differences.

1,134 citations


Journal ArticleDOI
TL;DR: The various applications of neural networks in image processing are categorised into a novel two-dimensional taxonomy for image processing algorithms and their specific conditions are discussed in detail.

Book
18 Feb 2002
TL;DR: The new edition of Feature Extraction and Image Processing provides an essential guide to the implementation of image processing and computer vision techniques, explaining techniques and fundamentals in a clear and concise manner, and features a companion website that includes worksheets, links to free software, Matlab files, solutions and new demonstrations.
Abstract: Image processing and computer vision are currently hot topics with undergraduates and professionals alike. "Feature Extraction and Image Processing" provides an essential guide to the implementation of image processing and computer vision techniques, explaining techniques and fundamentals in a clear and concise manner. Readers can develop working techniques, with usable code provided throughout and working Matlab and Mathcad files on the web. Focusing on feature extraction while also covering issues and techniques such as image acquisition, sampling theory, point operations and low-level feature extraction, the authors have a clear and coherent approach that will appeal to a wide range of students and professionals.The new edition includes: a new coverage of curvature in low-level feature extraction (SIFT and saliency) and features (phase congruency); geometric active contours; morphology; and camera models and an updated coverage of image smoothing (anistropic diffusion); skeletonization; edge detection; curvature; and shape descriptions (moments). It is an essential reading for engineers and students working in this cutting edge field. It is an ideal module text and background reference for courses in image processing and computer vision. It features a companion website that includes worksheets, links to free software, Matlab files, solutions and new demonstrations.

Proceedings ArticleDOI
13 May 2002
TL;DR: In this paper, insights on why image quality assessment is so difficult are provided by pointing out the weaknesses of the error sensitivity based framework and a new philosophy in designing image quality metrics is proposed.
Abstract: Image quality assessment plays an important role in various image processing applications. A great deal of effort has been made in recent years to develop objective image quality metrics that correlate with perceived quality measurement. Unfortunately, only limited success has been achieved. In this paper, we provide some insights on why image quality assessment is so difficult by pointing out the weaknesses of the error sensitivity based framework, which has been used by most image quality assessment approaches in the literature. Furthermore, we propose a new philosophy in designing image quality metrics: The main function of the human eyes is to extract structural information from the viewing field, and the human visual system is highly adapted for this purpose. Therefore, a measurement of structural distortion should be a good approximation of perceived image distortion. Based on the new philosophy, we implemented a simple but effective image quality indexing algorithm, which is very promising as shown by our current results.

Journal ArticleDOI
TL;DR: Algorithm for vision-based detection and classification of vehicles in monocular image sequences of traffic scenes recorded by a stationary camera based on the establishment of correspondences between regions and vehicles, as the vehicles move through the image sequence is presented.
Abstract: This paper presents algorithms for vision-based detection and classification of vehicles in monocular image sequences of traffic scenes recorded by a stationary camera. Processing is done at three levels: raw images, region level, and vehicle level. Vehicles are modeled as rectangular patches with certain dynamic behavior. The proposed method is based on the establishment of correspondences between regions and vehicles, as the vehicles move through the image sequence. Experimental results from highway scenes are provided which demonstrate the effectiveness of the method. We also briefly describe an interactive camera calibration tool that we have developed for recovering the camera parameters using features in the image selected by the user.

Journal ArticleDOI
TL;DR: A new magnetic resonance (MR) image analysis tool that produces cortical surface representations with spherical topology from MR images of the human brain is described, which is designed to require minimal user interaction to produce cortical representations.

Journal ArticleDOI
TL;DR: In this paper, an iterative method for particle image velocimetry (PIV) was proposed to enhance the precision and spatial resolution of numerical interrogation schemes by taking into account the local velocity derivatives in order to increase the interrogation spatial density and a refinement of the local interrogation window size.
Abstract: Image processing methods for particle image velocimetry (PIV) are reviewed. The discussion focuses on iterative methods aimed at enhancing the precision and spatial resolution of numerical interrogation schemes. Emphasis is placed on the efforts made to overcome the limitations of the correlation interrogation method with respect to typical problems such as in-plane loss of pairs, velocity gradient compensation and correlation peak locking. The discussion shows how the correlation signal benefits from simple operations such as the window-offset technique, or the continuous window deformation, which compensates for the in-plane velocity gradient. The image interrogation process is presented within the discussion of the image matching problem and several algorithms and implementations currently in use are classified depending on the choice made about the particle image pattern matching scheme. Several methods that differ in their implementations are found to be substantially similar. Iterative image deformation methods that account for the continuous particle image pattern transformation are analysed and the effect of crucial choices such as image interpolation method, displacement prediction correction and correlation peak fit scheme are discussed. The quantitative performance assessment made through synthetic PIV images yields order-of-magnitude improvement on the precision of the particle image displacement at a sub-pixel level when the image deformation is applied. Moreover, the issue of spatial resolution is addressed and the limiting factors of the specific interrogation methods are discussed. Finally, an attempt for a flow-adaptive spatial resolution method is proposed. The method takes into account the local velocity derivatives in order to perform a local increase of the interrogation spatial density and a refinement of the local interrogation window size. The resulting spatial resolution is selectively enhanced. The method's performance is analysed and compared with some precursor techniques, namely the conventional cross-correlation analysis with and without the effect of a window discrete offset and deformation. The suitability of the method for the measurement in turbulent flows is illustrated with the application to a turbulent backward facing step flow.

Journal ArticleDOI
TL;DR: A fully automatic "pipeline" image analysis framework that enhances the ability to detect small treatment effects not readily detectable through conventional analysis techniques and holds widespread potential for applications in other neurological disorders, as well as for the study of neurobiology in general.
Abstract: The quantitative analysis of magnetic resonance imaging (MRI) data has become increasingly important in both research and clinical studies aiming at human brain development, function, and pathology. Inevitably, the role of quantitative image analysis in the evaluation of drug therapy will increase, driven in part by requirements imposed by regulatory agencies. However, the prohibitive length of time involved and the significant intra- and inter-rater variability of the measurements obtained from manual analysis of large MRI databases represent major obstacles to the wider application of quantitative MRI analysis. We have developed a fully automatic "pipeline" image analysis framework and have successfully applied it to a number of large-scale, multi-center studies (more than 1000 MRI scans). This pipeline system is based on robust image processing algorithms, executed in a parallel, distributed fashion. This paper describes the application of this system to the automatic quantification of multiple sclerosis lesion load in MRI, in the context of a phase III clinical trial. The pipeline results were evaluated through an extensive validation study, revealing that the obtained lesion measurements are statistically indistinguishable from those obtained by trained human observers. Given that intra- and inter-rater measurement variability is eliminated by automatic analysis, this system enhances the ability to detect small treatment effects not readily detectable through conventional analysis techniques. While useful for clinical trial analysis in multiple sclerosis, this system holds widespread potential for applications in other neurological disorders, as well as for the study of neurobiology in general.

Journal ArticleDOI
07 Aug 2002
TL;DR: A chronological overview of the developments in interpolation theory, from the earliest times to the present date, brings out the connections between the results obtained in different ages, thereby putting the techniques currently used in signal and image processing into historical perspective.
Abstract: This paper presents a chronological overview of the developments in interpolation theory, from the earliest times to the present date. It brings out the connections between the results obtained in different ages, thereby putting the techniques currently used in signal and image processing into historical perspective. A summary of the insights and recommendations that follow from relatively recent theoretical as well as experimental studies concludes the presentation.

Journal ArticleDOI
TL;DR: In this paper, the authors compare two general and formal solutions to the problem of fusion of multispectral images with high-resolution panchromatic observations, and compare the results on SPOT data.
Abstract: This paper compares two general and formal solutions to the problem of fusion of multispectral images with high-resolution panchromatic observations. The former exploits the undecimated discrete wavelet transform, which is an octave bandpass representation achieved from a conventional discrete wavelet transform by omitting all decimators and upsampling the wavelet filter bank. The latter relies on the generalized Laplacian pyramid, which is another oversampled structure obtained by recursively subtracting from an image an expanded decimated lowpass version. Both the methods selectively perform spatial-frequencies spectrum substitution from an image to another. In both schemes, context dependency is exploited by thresholding the local correlation coefficient between the images to be merged, to avoid injection of spatial details that are not likely to occur in the target image. Unlike other multiscale fusion schemes, both the present decompositions are not critically subsampled, thus avoiding possible impairments in the fused images, due to missing cancellation of aliasing terms. Results are presented and discussed on SPOT data.

Proceedings ArticleDOI
10 Dec 2002
TL;DR: A no-reference blur metric based on the analysis of the spread of the edges in an image is presented, which is shown to perform well over a range of image content.
Abstract: We present a no-reference blur metric for images and video. The blur metric is based on the analysis of the spread of the edges in an image. Its perceptual significance is validated through subjective experiments. The novel metric is near real-time, has low computational complexity and is shown to perform well over a range of image content. Potential applications include optimization of source coding, network resource management and autofocus of an image capturing device.

Journal ArticleDOI
TL;DR: A new automated method that performs unsupervised pixel purity determination and endmember extraction from multidimensional datasets; this is achieved by using both spatial and spectral information in a combined manner.
Abstract: Spectral mixture analysis provides an efficient mechanism for the interpretation and classification of remotely sensed multidimensional imagery. It aims to identify a set of reference signatures (also known as endmembers) that can be used to model the reflectance spectrum at each pixel of the original image. Thus, the modeling is carried out as a linear combination of a finite number of ground components. Although spectral mixture models have proved to be appropriate for the purpose of large hyperspectral dataset subpixel analysis, few methods are available in the literature for the extraction of appropriate endmembers in spectral unmixing. Most approaches have been designed from a spectroscopic viewpoint and, thus, tend to neglect the existing spatial correlation between pixels. This paper presents a new automated method that performs unsupervised pixel purity determination and endmember extraction from multidimensional datasets; this is achieved by using both spatial and spectral information in a combined manner. The method is based on mathematical morphology, a classic image processing technique that can be applied to the spectral domain while being able to keep its spatial characteristics. The proposed methodology is evaluated through a specifically designed framework that uses both simulated and real hyperspectral data.

Journal ArticleDOI
TL;DR: A generic Fourier descriptor (GFD) is proposed to overcome the drawbacks of existing shape representation techniques by applying two-dimensional Fourier transform on a polar-raster sampled shape image.
Abstract: Shape description is one of the key parts of image content description for image retrieval. Most of the existing shape descriptors are usually either application dependent or non-robust, making them undesirable for generic shape description. In this paper, a generic Fourier descriptor (GFD) is proposed to overcome the drawbacks of existing shape representation techniques. The proposed shape descriptor is derived by applying two-dimensional Fourier transform on a polar-raster sampled shape image. The acquired shape descriptor is application independent and robust. Experimental results show that the proposed GFD outperforms common contour-based and region-based shape descriptors.

Journal ArticleDOI
07 Nov 2002
TL;DR: The "Camera Mouse" system tracks the computer user's movements with a video camera and translates them into the movements of the mouse pointer on the screen, and body features such as the tip of the user's nose or finger can be tracked.
Abstract: The "Camera Mouse" system has been developed to provide computer access for people with severe disabilities. The system tracks the computer user's movements with a video camera and translates them into the movements of the mouse pointer on the screen. Body features such as the tip of the user's nose or finger can be tracked. The visual tracking algorithm is based on cropping an online template of the tracked feature from the current image frame and testing where this template correlates in the subsequent frame. The location of the highest correlation is interpreted as the new location of the feature in the subsequent frame. Various body features are examined for tracking robustness and user convenience. A group of 20 people without disabilities tested the Camera Mouse and quickly learned how to use it to spell out messages or play games. Twelve people with severe cerebral palsy or traumatic brain injury have tried the system, nine of whom have shown success. They interacted with their environment by spelling out messages and exploring the Internet.

Journal ArticleDOI
TL;DR: New addressing, segmentation, and background correction methods for extracting information from microarray scanned images are proposed and it is suggested that seeded region growing segmentation with morphological background correction provides precise and accurate estimates of foreground and background intensities.
Abstract: Microarrays are part of a new class of biotechnologies which allow the monitoring of expression levels for thousands of genes simultaneously. Image analysis is an important aspect of microarray experiments, one that can have a potentially large impact on subsequent analyses such as clustering or the identification of differentially expressed genes. This article reviews a number of existing image analysis approaches for cDNA microarray experiments and proposes new addressing, segmentation, and background correction methods for extracting information from microarray scanned images. The segmentation component uses a seeded region growing algorithm which makes provision for spots of different shapes and sizes. The background estimation approach is based on an image analysis technique known as morphological opening. These new image analysis procedures are implemented in a software package named Spot, built on the R environment for statistical computing. The statistical properties of the different segmentation ...

Journal ArticleDOI
01 Aug 2002
TL;DR: This work discusses the various features of this operator that make it the filter of choice in the area of edge detection, and reviews several linear and nonlinear Gaussian-based edge detection methods.
Abstract: The Gaussian filter has been used extensively in image processing and computer vision for many years. We discuss the various features of this operator that make it the filter of choice in the area of edge detection. Despite these desirable features of the Gaussian filter, edge detection algorithms which use it suffer from many problems. We review several linear and nonlinear Gaussian-based edge detection methods.

Journal ArticleDOI
TL;DR: A new approach for watermarking of digital images providing robustness to geometrical distortions by proposing an embedding and detection scheme where the mark is bound with a content descriptor defined by salient points.
Abstract: This paper presents a new approach for watermarking of digital images providing robustness to geometrical distortions. The weaknesses of classical watermarking methods to geometrical distortions are outlined first. Geometrical distortions can be decomposed into two classes: global transformations such as rotations and translations and local transformations such as the StirMark attack. An overview of existing self-synchronizing schemes is then presented. Theses schemes can use periodical properties of the mark, invariant properties of transforms, template insertion, or information provided by the original image to counter geometrical distortions. Thereafter, a new class of watermarking schemes using the image content is presented. We propose an embedding and detection scheme where the mark is bound with a content descriptor defined by salient points. Three different types of feature points are studied and their robustness to geometrical transformations is evaluated to develop an enhanced detector. The embedding of the signature is done by extracting feature points of the image and performing a Delaunay tessellation on the set of points. The mark is embedded using a classical additive scheme inside each triangle of the tessellation. The detection is done using correlation properties on the different triangles. The performance of the presented scheme is evaluated after JPEG compression, geometrical attack and transformations. Results show that the fact that the scheme is robust to these different manipulations. Finally, in our concluding remarks, we analyze the different perspectives of such content-based watermarking scheme.

Journal ArticleDOI
TL;DR: In this paper, the systematic errors that arise from the use of undermatched shape functions, i.e., shape functions of lower order than the actual displacement field, are analyzed, under certain conditions, the shape functions used can be approximated by a Savitzky-Golay low-pass filter applied to the displacement functions, permitting a convenient error analysis.
Abstract: Digital image correlation techniques are commonly used to measure specimen displacements by finding correspondences between an image of the specimen in an undeformed or reference configuration and a second image under load. To establish correspondences between the two images, numerical techniques are used to locate an initially square image subset in a reference image within an image taken under load. During this process, shape functions of varying order can be applied to the initially square subset. Zero order shape functions permit the subset to translate rigidly, while first-order shape functions represent an affine transform of the subset that permits a combination of translation, rotation, shear and normal strains. In this article, the systematic errors that arise from the use of undermatched shape function, i.e., shape functions of lower order than the actual displacement field, are analyzed. It is shown that, under certain conditions, the shape functions used can be approximated by a Savitzky-Golay low-pass filter applied to the displacement functions, permitting a convenient error analysis. Furthermore, this analysis is not limited to the displacements, but naturally extends to the higher-order terms included in the shape functions. This permits a direct analysis of the systematic strain errors associated with an undermatched shape function. Detailed numerical studies are presented for the case of a second-order displacement field and first- and second-order shape functions. Finally, the relation of this work to previously published studies is discussed.

Journal ArticleDOI
Rainer Lienhart1, A. Wernicke
TL;DR: This work proposes a novel method for localizing and segmenting text in complex images and videos that is not only able to locate and segment text occurrences into large binary images, but is also able to track each text line with sub-pixel accuracy over the entire occurrence in a video.
Abstract: Many images, especially those used for page design on Web pages, as well as videos contain visible text. If these text occurrences could be detected, segmented, and recognized automatically, they would be a valuable source of high-level semantics for indexing and retrieval. We propose a novel method for localizing and segmenting text in complex images and videos. Text lines are identified by using a complex-valued multilayer feed-forward network trained to detect text at a fixed scale and position. The network's output at all scales and positions is integrated into a single text-saliency map, serving as a starting point for candidate text lines. In the case of video, these candidate text lines are refined by exploiting the temporal redundancy of text in video. Localized text lines are then scaled to a fixed height of 100 pixels and segmented into a binary image with black characters on white background. For videos, temporal redundancy is exploited to improve segmentation performance. Input images and videos can be of any size due to a true multiresolution approach. Moreover, the system is not only able to locate and segment text occurrences into large binary images, but is also able to track each text line with sub-pixel accuracy over the entire occurrence in a video, so that one text bitmap is created for all instances of that text line. Therefore, our text segmentation results can also be used for object-based video encoding such as that enabled by MPEG-4.

Journal ArticleDOI
TL;DR: Methods for a geometrical and structural analysis of vessel systems have been evaluated in the clinical environment and have been used in more than 170 cases so far to plan interventions and transplantations.
Abstract: For liver surgical planning, the structure and morphology of the hepatic vessels and their relationship to tumors are of major interest. To achieve a fast and robust assistance with optimal quantitative and visual information, we present methods for a geometrical and structural analysis of vessel systems. Starting from the raw image data a sequence of image processing steps has to be carried out until a three-dimensional representation of the relevant anatomic and pathologic structures is generated. Based on computed tomography (CT) scans, the following steps are performed. 1) The volume data is preprocessed and the vessels are segmented. 2) The skeleton of the vessels is determined and transformed into a graph enabling a geometrical and structural shape analysis. Using this information the different intrahepatic vessel systems are identified automatically. 3) Based on the structural analysis of the branches of the portal vein, their vascular territories are approximated with different methods. These methods are compared and validated anatomically by means of corrosion casts of human livers. 4) Vessels are visualized with graphics primitives fitted to the skeleton to provide smooth visualizations without aliasing artifacts. The image analysis techniques have been evaluated in the clinical environment and have been used in more than 170 cases so far to plan interventions and transplantations.

Journal ArticleDOI
TL;DR: An important function of hyperspectral signal processing is to eliminate the redundancy in the spectral and spatial sample data while preserving the high-quality features needed for detection, discrimination, and classification.
Abstract: Electro-optical remote sensing involves the acquisition of information about an object or scene without coming into physical contact with it. This is achieved by exploiting the fact that the materials comprising the various objects in a scene reflect, absorb, and emit electromagnetic radiation in ways characteristic of their molecular composition and shape. If the radiation arriving at the sensor is measured at each wavelength over a sufficiently broad spectral band, the resulting spectral signature, or simply spectrum, can be used (in principle) to uniquely characterize and identify any given material. An important function of hyperspectral signal processing is to eliminate the redundancy in the spectral and spatial sample data while preserving the high-quality features needed for detection, discrimination, and classification. This dimensionality reduction is implemented in a scene-dependent (adaptive) manner and may be implemented as a distinct step in the processing or as an integral part of the overall algorithm. The most widely used algorithm for dimensionality reduction is principal component analysis (PCA) or, equivalently, Karhunen-Loeve transformation.