scispace - formally typeset
Search or ask a question

Showing papers on "Image processing published in 1997"


Journal ArticleDOI
TL;DR: It is argued that insertion of a watermark under this regime makes the watermark robust to signal processing operations and common geometric transformations provided that the original image is available and that it can be successfully registered against the transformed watermarked image.
Abstract: This paper presents a secure (tamper-resistant) algorithm for watermarking images, and a methodology for digital watermarking that may be generalized to audio, video, and multimedia data. We advocate that a watermark should be constructed as an independent and identically distributed (i.i.d.) Gaussian random vector that is imperceptibly inserted in a spread-spectrum-like fashion into the perceptually most significant spectral components of the data. We argue that insertion of a watermark under this regime makes the watermark robust to signal processing operations (such as lossy compression, filtering, digital-analog and analog-digital conversion, requantization, etc.), and common geometric transformations (such as cropping, scaling, translation, and rotation) provided that the original image is available and that it can be successfully registered against the transformed watermarked image. In these cases, the watermark detector unambiguously identifies the owner. Further, the use of Gaussian noise, ensures strong resilience to multiple-document, or collusional, attacks. Experimental results are provided to support these claims, along with an exposition of pending open problems.

6,194 citations


Journal ArticleDOI
TL;DR: In this article, a geodesic approach based on active contours evolving in time according to intrinsic geometric measures of the image is presented. But this approach is not suitable for 3D object segmentation.
Abstract: A novel scheme for the detection of object boundaries is presented. The technique is based on active contours evolving in time according to intrinsic geometric measures of the image. The evolving contours naturally split and merge, allowing the simultaneous detection of several objects and both interior and exterior boundaries. The proposed approach is based on the relation between active contours and the computation of geodesics or minimal distance curves. The minimal distance curve lays in a Riemannian space whose metric is defined by the image content. This geodesic approach for object segmentation allows to connect classical “snakes” based on energy minimization and geometric active contours based on the theory of curve evolution. Previous models of geometric active contours are improved, allowing stable boundary detection when their gradients suffer from large variations, including gaps. Formal results concerning existence, uniqueness, stability, and correctness of the evolution are presented as well. The scheme was implemented using an efficient algorithm for curve evolution. Experimental results of applying the scheme to real images including objects with holes and medical data imagery demonstrate its power. The results may be extended to 3D object segmentation as well.

4,967 citations


Journal ArticleDOI
TL;DR: This paper describes a new approach to low level image processing; in particular, edge and corner detection and structure preserving noise reduction and the resulting methods are accurate, noise resistant and fast.
Abstract: This paper describes a new approach to low level image processing; in particular, edge and corner detection and structure preserving noise reduction. Non-linear filtering is used to define which parts of the image are closely related to each individual pixel; each pixel has associated with it a local image region which is of similar brightness to that pixel. The new feature detectors are based on the minimization of this local image region, and the noise reduction method uses this region as the smoothing neighbourhood. The resulting methods are accurate, noise resistant and fast. Details of the new feature detectors and of the new noise reduction method are described, along with test results.

3,669 citations


Journal ArticleDOI
TL;DR: A new information-theoretic approach is presented for finding the pose of an object in an image that works well in domains where edge or gradient-magnitude based methods have difficulty, yet it is more robust than traditional correlation.
Abstract: A new information-theoretic approach is presented for finding the pose of an object in an image. The technique does not require information about the surface properties of the object, besides its shape, and is robust with respect to variations of illumination. In our derivation few assumptions are made about the nature of the imaging process. As a result the algorithms are quite general and may foreseeably be used in a wide variety of imaging situations. Experiments are presented that demonstrate the approach registering magnetic resonance (MR) images, aligning a complex 3D object model to real scenes including clutter and occlusion, tracking a human head in a video sequence and aligning a view-based 2D object model to real images. The method is based on a formulation of the mutual information between the model and the image. As applied here the technique is intensity-based, rather than feature-based. It works well in domains where edge or gradient-magnitude based methods have difficulty, yet it is more robust than traditional correlation. Additionally, it has an efficient implementation that is based on stochastic approximation.

3,584 citations


Proceedings ArticleDOI
03 Aug 1997
TL;DR: This work discusses how this work is applicable in many areas of computer graphics involving digitized photographs, including image-based modeling, image compositing, and image processing, and demonstrates a few applications of having high dynamic range radiance maps.
Abstract: We present a method of recovering high dynamic range radiance maps from photographs taken with conventional imaging equipment. In our method, multiple photographs of the scene are taken with different amounts of exposure. Our algorithm uses these differently exposed photographs to recover the response function of the imaging process, up to factor of scale, using the assumption of reciprocity. With the known response function, the algorithm can fuse the multiple photographs into a single, high dynamic range radiance map whose pixel values are proportional to the true radiance values in the scene. We demonstrate our method on images acquired with both photochemical and digital imaging processes. We discuss how this work is applicable in many areas of computer graphics involving digitized photographs, including image-based modeling, image compositing, and image processing. Lastly, we demonstrate a few applications of having high dynamic range radiance maps, such as synthesizing realistic motion blur and simulating the response of the human visual system.

2,967 citations


Book
01 Jan 1997
TL;DR: The Nature of Remote Sensing: Introduction, Sensor Characteristics and Spectral Stastistics, and Spatial Transforms: Introduction.
Abstract: The Nature of Remote Sensing: Introduction. Remote Sensing. Information Extraction from Remote-Sensing Images. Spectral Factors in Remote Sensing. Spectral Signatures. Remote-Sensing Systems. Optical Sensors. Temporal Characteristics. Image Display Systems. Data Systems. Summary. Exercises. References. Optical Radiation Models: Introduction. Visible to Short Wave Infrared Region. Solar Radiation. Radiation Components. Surface-Reflected. Unscattered Component. Surface-Reflected. Atmosphere-Scattered Component. Path-Scattered Component. Total At-Sensor. Solar Radiance. Image Examples in the Solar Region. Terrain Shading. Shadowing. Atmospheric Correction. Midwave to Thermal Infrared Region. Thermal Radiation. Radiation Components. Surface-Emitted Component. Surface-Reflected. Atmosphere-Emitted Component. Path-Emitted Component. Total At-Sensor. Emitted Radiance. Total Solar and Thermal Upwelling Radiance. Image Examples in the Thermal Region. Summary. Exercises. References. Sensor Models: Introduction. Overall Sensor Model. Resolution. The Instrument Response. Spatial Resolution. Spectral Resolution. Spectral Response. Spatial Response. Optical PSFopt. Image Motion PSFIM. Detector PSFdet. Electronics PSFel. Net PSFnet. Comparison of Sensor PSFs. PSF Summary for TM. Imaging System Simulation. Amplification. Sampling and Quantization. Simplified Sensor Model. Geometric Distortion. Orbit Models. Platform Attitude Models. Scanner Models. Earth Model. Line and Whiskbroom ScanGeometry. Pushbroom Scan Geometry. Topographic Distortion. Summary. Exercises. References. Data Models: Introduction. A Word on Notation. Univariate Image Statistics. Histogram. Normal Distribution. Cumulative Histogram. Statistical Parameters. Multivariate Image Statistics. Reduction to Univariate Statistics. Noise Models. Statistical Measures of Image Quality. Contrast. Modulation. Signal-to-Noise Ratio (SNR). Noise Equivalent Signal. Spatial Statistics. Visualization of Spatial Covariance. Covariance with Semivariogram. Separability and Anisotropy. Power Spectral Density. Co-occurrence Matrix. Fractal Geometry. Topographic and Sensor Effects. Topography and Spectral Statistics. Sensor Characteristics and Spectral Stastistics. Sensor Characteristics and Spectral Scattergrams. Summary. Exercises. References. Spectral Transforms: Introduction. Feature Space. Multispectral Ratios. Vegetation Indexes. Image Examples. Principal Components. Standardized Principal Components (SPC) Transform. Maximum Noise Fraction (MNF) Transform. Tasseled Cap Tranformation. Contrast Enhancement. Transformations Based on Global Statistics. Linear Transformations. Nonlinear Transformations. Normalization Stretch. Reference Stretch. Thresholding. Adaptive Transformation. Color Image Contrast Enhancement. Min-max Stretch. Normalization Stretch. Decorrelation Stretch. Color Spacer Transformations. Summary. Exercises. References. Spatial Transforms: Introduction. An Image Model for Spatial Filtering. Convolution Filters. Low Pass and High Pass Filters. High Boost Filters. Directional Filters. The Border Region. Characterization of Filtered Images. The Box Filter Algorithm. Cascaded Linear Filters. Statistical Filters. Gradient Filters. Fourier Synthesis. Discrete Fourier Transforms in 2-D. The Fourier Components. Filtering with the Fourier Transform. Transfer Functions. The Power Spectrum. Scale Space Transforms. Image Resolution Pyramids. Zero-Crossing Filters. Laplacian-of-Gaussian (LoG) Filters. Difference-of-Gaussians (DoG) Filters.Wavelet Transforms. Summary. Exercises. References. Correction and Calibration: Introduction. Noise Correction. Global Noise. Sigma Filter. Nagao-Matsuyama Filter. Local Noise. Periodic Noise. Distriping 359. Global,Linear Detector Matching. Nonlinear Detector Matching. Statistical Modification to Linear and Nonlinear Detector. Matching. Spatial Filtering Approaches. Radiometric Calibration. Sensor Calibration. Atmospheric Correction. Solar and Topographic Correction. Image Examples. Calibration and Normalization of Hyperspectral Imagery. AVIRIS Examples. Distortion Correction. Polynomial Distortion Models. Ground Control Points (GCPs). Coordinate Transformation. Map Projections. Resampling. Summary. Exercises References. Registration and Image Fusion: Introduction. What is Registration? Automated GCP Location. Area Correlation. Other Spatial Features. Orthrectification. Low-Resolution DEM. High-Resolution DEM. Hierarchical Warp Stereo. Multi-Image Fusion. Spatial Domain Fusion. High Frequency Modulation. Spectral Domain Fusion. Fusion Image Examples. Summary. Exercises. References. Thematic Classification: Introduction. The Importance of Image Scale. The Notion of Similarity. Hard Versus Soft Classification. Training the Classifier. Supervised Training. Unsupervised Training. K-Means Clustering Algorithm. Clustering Examples. Hybrid Supervised/Unsupervised Training. Non-Parametric Classification Algorithms. Level-Slice. Nearest-Mean. Artificial Neural Networks (ANNs). Back-Propagation Algorithm. Nonparametric Classification Examples. Parametric Classification Algorithms. Estimation of Model-Parameters. Discriminant Functions. The Normal Distribution Model. Relation to the Nearest-Mean Classifier. Supervised Classification Examples and Comparison to Nonparametric Classifiers. Segmentation. Region Growing. Region Labeling. Sub-Pixel Classification. The Linear Mixing Model. Unmixing Model. Hyperspectral Image Analysis. Visualization of the Image Cube. Feature Extraction. Image Residuals. Pre-Classification Processing and Feature Extraction. Classification Algorithms. Exercises. Error Analysis. Multitemporal Images. Summary. References. Index.

2,290 citations


Journal ArticleDOI
TL;DR: A new method is proposed in which the distribution of complex amplitude at a plane is measured by phase-shifting interferometry and then Fresnel transformed by a digital computer, which can reconstruct an arbitrary cross section of a three-dimensional object with higher image quality and a wider viewing angle than from conventional digital holography using an off-axis configuration.
Abstract: A new method for three-dimensional image formation is proposed in which the distribution of complex amplitude at a plane is measured by phase-shifting interferometry and then Fresnel transformed by a digital computer. The method can reconstruct an arbitrary cross section of a three-dimensional object with higher image quality and a wider viewing angle than from conventional digital holography using an off-axis configuration. Basic principles and experimental verification are described.

1,813 citations


Journal ArticleDOI
Yeong-Taeg Kim1
TL;DR: It is shown mathematically that the proposed algorithm preserves the mean brightness of a given image significantly well compared to typical histogram equalization while enhancing the contrast and, thus, provides a natural enhancement that can be utilized in consumer electronic products.
Abstract: Histogram equalization is widely used for contrast enhancement in a variety of applications due to its simple function and effectiveness. Examples include medical image processing and radar signal processing. One drawback of the histogram equalization can be found on the fact that the brightness of an image can be changed after the histogram equalization, which is mainly due to the flattening property of the histogram equalization. Thus, it is rarely utilized in consumer electronic products such as TV where preserving the original input brightness may be necessary in order not to introduce unnecessary visual deterioration. This paper proposes a novel extension of histogram equalization to overcome such a drawback of histogram equalization. The essence of the proposed algorithm is to utilize independent histogram equalizations separately over two subimages obtained by decomposing the input image based on its mean with a constraint that the resulting equalized subimages are bounded by each other around the input mean. It is shown mathematically that the proposed algorithm preserves the mean brightness of a given image significantly well compared to typical histogram equalization while enhancing the contrast and, thus, provides a natural enhancement that can be utilized in consumer electronic products.

1,562 citations


Journal ArticleDOI
TL;DR: This paper proposes a deterministic strategy, based on alternate minimizations on the image and the auxiliary variable, which leads to the definition of an original reconstruction algorithm, called ARTUR, which can be applied in a large number of applications in image processing.
Abstract: Many image processing problems are ill-posed and must be regularized. Usually, a roughness penalty is imposed on the solution. The difficulty is to avoid the smoothing of edges, which are very important attributes of the image. In this paper, we first give conditions for the design of such an edge-preserving regularization. Under these conditions, we show that it is possible to introduce an auxiliary variable whose role is twofold. First, it marks the discontinuities and ensures their preservation from smoothing. Second, it makes the criterion half-quadratic. The optimization is then easier. We propose a deterministic strategy, based on alternate minimizations on the image and the auxiliary variable. This leads to the definition of an original reconstruction algorithm, called ARTUR. Some theoretical properties of ARTUR are discussed. Experimental results illustrate the behavior of the algorithm. These results are shown in the field of 2D single photon emission tomography, but this method can be applied in a large number of applications in image processing.

1,360 citations


Journal ArticleDOI
TL;DR: P positron-emission tomography (PET) has inherent advantages that avoid the shortcomings of other nuclear medicine imaging methods, and its image reconstruction methods with origins in signal and image processing are discussed.
Abstract: We review positron-emission tomography (PET), which has inherent advantages that avoid the shortcomings of other nuclear medicine imaging methods. PET image reconstruction methods with origins in signal and image processing are discussed, including the potential problems of these methods. A summary of statistical image reconstruction methods, which can yield improved image quality, is also presented.

1,257 citations


Journal ArticleDOI
TL;DR: A hybrid method combining the simplicity of theML and the incorporation of nonellipsoid constraints is presented, giving improved restoration performance, compared with the ML and the POCS approaches.
Abstract: The three main tools in the single image restoration theory are the maximum likelihood (ML) estimator, the maximum a posteriori probability (MAP) estimator, and the set theoretic approach using projection onto convex sets (POCS). This paper utilizes the above known tools to propose a unified methodology toward the more complicated problem of superresolution restoration. In the superresolution restoration problem, an improved resolution image is restored from several geometrically warped, blurred, noisy and downsampled measured images. The superresolution restoration problem is modeled and analyzed from the ML, the MAP, and POCS points of view, yielding a generalization of the known superresolution restoration methods. The proposed restoration approach is general but assumes explicit knowledge of the linear space- and time-variant blur, the (additive Gaussian) noise, the different measured resolutions, and the (smooth) motion characteristics. A hybrid method combining the simplicity of the ML and the incorporation of nonellipsoid constraints is presented, giving improved restoration performance, compared with the ML and the POCS approaches. The hybrid method is shown to converge to the unique optimal solution of a new definition of the optimization problem. Superresolution restoration from motionless measurements is also discussed. Simulations demonstrate the power of the proposed methodology.

Proceedings ArticleDOI
03 Aug 1997
TL;DR: This paper presents a novel approach to creating full view panoramic mosaics from image sequences that does not require any controlled motions or constraints on how the images are taken (as long as there is no strong motion parallax).
Abstract: This paper presents a novel approach to creating full view panoramic mosaics from image sequences. Unlike current panoramic stitching methods, which usually require pure horizontal camera panning, our system does not require any controlled motions or constraints on how the images are taken (as long as there is no strong motion parallax). For example, images taken from a hand-held digital camera can be stitched seamlessly into panoramic mosaics. Because we represent our image mosaics using a set of transforms, there are no singularity problems such as those existing at the top and bottom of cylindrical or spherical maps. Our algorithm is fast and robust because it directly recovers 3D rotations instead of general 8 parameter planar perspective transforms. Methods to recover camera focal length are also presented. We also present an algorithm for efficiently extracting environment maps from our image mosaics. By mapping the mosaic onto an artibrary texture-mapped polyhedron surrounding the origin, we can explore the virtual environment using standard 3D graphics viewers and hardware without requiring special-purpose players. CR

Journal ArticleDOI
TL;DR: The basic idea is that instead of matching two images directly, one performs intermediate within-modality registrations to two template images that are already in register, and a least-squares minimization is used to determine the affine transformations that map between the templates and the images.

Journal ArticleDOI
TL;DR: A maximum a posteriori (MAP) framework for jointly estimating image registration parameters and the high-resolution image is presented and experimental results are provided to illustrate the performance of the proposed MAP algorithm using both visible and infrared images.
Abstract: In many imaging systems, the detector array is not sufficiently dense to adequately sample the scene with the desired field of view. This is particularly true for many infrared focal plane arrays. Thus, the resulting images may be severely aliased. This paper examines a technique for estimating a high-resolution image, with reduced aliasing, from a sequence of undersampled frames. Several approaches to this problem have been investigated previously. However, in this paper a maximum a posteriori (MAP) framework for jointly estimating image registration parameters and the high-resolution image is presented. Several previous approaches have relied on knowing the registration parameters a priori or have utilized registration techniques not specifically designed to treat severely aliased images. In the proposed method, the registration parameters are iteratively updated along with the high-resolution image in a cyclic coordinate-descent optimization procedure. Experimental results are provided to illustrate the performance of the proposed MAP algorithm using both visible and infrared images. Quantitative error analysis is provided and several images are shown for subjective evaluation.

Proceedings ArticleDOI
17 Jun 1997
TL;DR: A new external force for active contours is developed, which is computed as a diffusion of the gradient vectors of a gray-level or binary edge map derived from the image, which has a large capture range and forces active contour regions into concave regions.
Abstract: Snakes, or active contours, are used extensively in computer vision and image processing applications, particularly to locate object boundaries. Problems associated with initialization and poor convergence to concave boundaries, however, have limited their utility. This paper develops a new external force for active contours, largely solving both problems. This external force, which we call gradient vector flow (GVF) is computed as a diffusion of the gradient vectors of a gray-level or binary edge map derived from the image. The resultant field has a large capture range and forces active contours into concave regions. Examples on simulated images and one real image are presented.

Journal ArticleDOI
TL;DR: The main shortcomings of conventional mapping-namely, prolonged x-ray exposure, low spatial resolution, and the inability to accurately navigate to a predefined site-can all be overcome with this new method.
Abstract: Background Cardiac mapping is essential for understanding the mechanisms of arrhythmias and for directing curative procedures. A major limitation of the current methods is the inability to accurately relate local electrograms to their spatial orientation. The objective of this study was to present and test the accuracy of a new method for nonfluoroscopic, catheter-based, endocardial mapping. Methods and Results The method is based on using a new locatable catheter connected to an endocardial mapping and navigating system. The system uses magnetic technology to accurately determine the location and orientation of the catheter and simultaneously records the local electrogram from its tip. By sampling a plurality of endocardial sites, the system reconstructs the three-dimensional geometry of the chamber, with the electrophysiological information color-coded and superimposed on the anatomy. The accuracy of the system was tested in both in vitro and in vivo studies and was found to be highly reproducible (SD, ...

Proceedings ArticleDOI
26 Oct 1997
TL;DR: Experimental results show that the image retrieval precision increases considerably by using the proposed integration approach, and the relevance feedback technique from the IR domain is used in content-based image retrieval to demonstrate the effectiveness of this conversion.
Abstract: Technology advances in the areas of image processing (IP) and information retrieval (IR) have evolved separately for a long time. However, successful content-based image retrieval systems require the integration of the two. There is an urgent need to develop integration mechanisms to link the image retrieval model to text retrieval model, such that the well established text retrieval techniques can be utilized. Approaches of converting image feature vectors (IF domain) to weighted-term vectors (IR domain) are proposed in this paper. Furthermore, the relevance feedback technique from the IR domain is used in content-based image retrieval to demonstrate the effectiveness of this conversion. Experimental results show that the image retrieval precision increases considerably by using the proposed integration approach.

Journal ArticleDOI
TL;DR: A tone reproduction operator is presented that preserves visibility in high dynamic range scenes and introduces a new histogram adjustment technique, based on the population of local adaptation luminances in a scene, that incorporates models for human contrast sensitivity, glare, spatial acuity, and color sensitivity.
Abstract: We present a tone reproduction operator that preserves visibility in high dynamic range scenes. Our method introduces a new histogram adjustment technique, based on the population of local adaptation luminances in a scene. To match subjective viewing experience, the method incorporates models for human contrast sensitivity, glare, spatial acuity, and color sensitivity. We compare our results to previous work and present examples of our techniques applied to lighting simulation and electronic photography.

Proceedings ArticleDOI
17 Jun 1997
TL;DR: A new camera with a hemispherical field of view is presented and results are presented on the software generation of pure perspective images from an omnidirectional image, given any user-selected viewing direction and magnification.
Abstract: Conventional video cameras have limited fields of view that make them restrictive in a variety of vision applications. There are several ways to enhance the field of view of an imaging system. However, the entire imaging system must have a single effective viewpoint to enable the generation of pure perspective images from a sensed image. A new camera with a hemispherical field of view is presented. Two such cameras can be placed back-to-back, without violating the single viewpoint constraint, to arrive at a truly omnidirectional sensor. Results are presented on the software generation of pure perspective images from an omnidirectional image, given any user-selected viewing direction and magnification. The paper concludes with a discussion on the spatial resolution of the proposed camera.

Journal ArticleDOI
TL;DR: The proposed scheme for segmentation is based on the iterative conditional modes (ICM) algorithm in which measurement model parameters are estimated using local information at each site, and the prior model parametersare estimated using the segmentation after each cycle of iterations.
Abstract: A statistical model is presented that represents the distributions of major tissue classes in single-channel magnetic resonance (MR) cerebral images. Using the model, cerebral images are segmented into gray matter, white matter, and cerebrospinal fluid (CSF). The model accounts for random noise, magnetic field inhomogeneities, and biological variations of the tissues. Intensity measurements are modeled by a finite Gaussian mixture. Smoothness and piecewise contiguous nature of the tissue regions are modeled by a three-dimensional (3-D) Markov random field (MRF). A segmentation algorithm, based on the statistical model, approximately finds the maximum a posteriori (MAP) estimation of the segmentation and estimates the model parameters from the image data. The proposed scheme for segmentation is based on the iterative conditional modes (ICM) algorithm in which measurement model parameters are estimated using local information at each site, and the prior model parameters are estimated using the segmentation after each cycle of iterations. Application of the algorithm to a sample of clinical MR brain scans, comparisons of the algorithm with other statistical methods, and a validation study with a phantom are presented. The algorithm constitutes a significant step toward a complete data driven unsupervised approach to segmentation of MR images in the presence of the random noise and intensity inhomogeneities.

Journal ArticleDOI
TL;DR: Findings indicate that this extrinsic-point-based, interactive image-guided neurosurgical system designed at Vanderbilt University is an accurate navigational aid that can provide real-time feedback to the surgeon about anatomical structures encountered in the surgical field.
Abstract: Describes an extrinsic-point-based, interactive image-guided neurosurgical system designed at Vanderbilt University, Nashville, TN, as part of a collaborative effort among the Departments of Neurological Surgery, Computer Science, and Biomedical Engineering. Multimodal image-to-image (II) and image-to-physical (IP) registration is accomplished using implantable markers. Physical space tracking is accomplished with optical triangulation. The authors investigate the theoretical accuracy of point-based registration using numerical simulations, the experimental accuracy of their system using data obtained with a phantom, and the clinical accuracy of their system using data acquired in a prospective clinical trial by 6 neurosurgeons at 4 medical centers from 158 patients undergoing craniotomies to respect cerebral lesions. The authors can determine the position of their markers with an error of approximately 0.4 mm in X-ray computed tomography (CT) and magnetic resonance (MR) images and 0.3 mm in physical space. The theoretical registration error using 4 such markers distributed around the head in a configuration that is clinically practical is approximately 0.5-0.6 mm. The mean CT-physical registration error for the: phantom experiments is 0.5 mm and for the clinical data obtained with rigid head fixation during scanning is 0.7 mm. The mean CT-MR registration error for the clinical data obtained without rigid head fixation during scanning is 1.4 mm, which is the highest mean error that the authors observed. These theoretical and experimental findings indicate that this system is an accurate navigational aid that can provide real-time feedback to the surgeon about anatomical structures encountered in the surgical field.

Journal ArticleDOI
TL;DR: A methodology for evaluating medical image segmentation algorithms wherein the only information available is boundaries outlined by multiple expert observers is proposed, and the results of the segmentation algorithm can be evaluated against the multiple observers' outlines.
Abstract: Image segmentation is the partition of an image into a set of nonoverlapping regions whose union is the entire image. The image is decomposed into meaningful parts which are uniform with respect to certain characteristics, such as gray level or texture. In this paper, we propose a methodology for evaluating medical image segmentation algorithms wherein the only information available is boundaries outlined by multiple expert observers. In this case, the results of the segmentation algorithm can be evaluated against the multiple observers' outlines. We have derived statistics to enable us to find whether the computer-generated boundaries agree with the observers' hand-outlined boundaries as much as the different observers agree with each other. We illustrate the use of this methodology by evaluating image segmentation algorithms on two different applications in ultrasound imaging. In the first application, we attempt to find the epicardial and endocardial boundaries from cardiac ultrasound images, and in the second application, our goal is to find the fetal skull and abdomen boundaries from prenatal ultrasound images.

Journal ArticleDOI
TL;DR: This paper considers models based on radiative transfer theory and its derivatives, which are either stochastic in nature (random walk, Monte Carlo, and Markov processes) or deterministic (partial differential equation models and their solutions).
Abstract: The desire for a diagnostic optical imaging modality has motivated the development of image reconstruction procedures involving solution of the inverse problem. This approach is based on the assumption that, given a set of measurements of transmitted light between pairs of points on the surface of an object, there exists a unique three-dimensional distribution of internal scatterers and absorbers which would yield that set. Thus imaging becomes a task of solving an inverse problem using an appropriate model of photon transport. In this paper we examine the models that have been developed for this task, and review current approaches to image reconstruction. Specifically, we consider models based on radiative transfer theory and its derivatives, which are either stochastic in nature (random walk, Monte Carlo, and Markov processes) or deterministic (partial differential equation models and their solutions). Image reconstruction algorithms are discussed which are based on either direct backprojection, perturbation methods, nonlinear optimization, or Jacobian calculation. Finally we discuss some of the fundamental problems that must be addressed before optical tomography can be considered to be an understood problem, and before its full potential can be realized.

Journal ArticleDOI
TL;DR: In this article, the location of the displacement-correlation peak in PIV images is optimized by the use of a window offset equal to the integer-pixel displacement, and the effect is predicted by an analytical model for the statistical properties of estimators for the displacement.
Abstract: This paper describes how the accuracy for estimating the location of the displacement-correlation peak in (digital) particle image velocimetry (PIV) can be optimized by the use of a window offset equal to the integer-pixel displacement. The method works for both cross-correlation analysis of single-exposure image pairs and multiple-exposure images. The effect is predicted by an analytical model for the statistical properties of estimators for the displacement, and it is observed in the analysis of synthetic PIV images of isotropic turbulence, and in actual measurements of grid-generated turbulence and of fully-developed turbulent pipe flow.

Journal ArticleDOI
TL;DR: In this paper, the possibility of using a television camera to shoot directly numerous real images produced by a lens array was studied, and the results showed that with this new direct pickup method, they can obtain an IP image like those obtained by using the conventional IP method.
Abstract: We studied integral photography (IP), which creates three-dimensional autostereoscopic images. In particular we studied the possibility of a new method that uses a television camera to shoot directly numerous real images produced by a lens array. Unlike the conventional IP method in which the film is placed immediately behind a lens array, this method employs a television camera, which enables us to shoot moving pictures. Of a number of factors affecting the process of image pickup, we examined some optical factors and compared them with those obtained by the conventional IP method. The results show that with this new direct pickup method that uses a television camera, we can obtain an IP image like those obtained by using the conventional IP method. Further, we conducted an experiment with an high-definition TV camera, confirming the production of an autostereoscopic image by using a display device that combines a liquid-crystal panel and pinholes.

Journal ArticleDOI
TL;DR: A new registration algorithm based on spline representations of the displacement field which can be specialized to solve all of the problems in multiframe image analysis, including the computation of optic flow, stereo correspondence, structure from motion, and feature tracking.
Abstract: The problem of image registration subsumes a number of problems and techniques in multiframe image analysis, including the computation of optic flow (general pixel-based motion), stereo correspondence, structure from motion, and feature tracking. We present a new registration algorithm based on spline representations of the displacement field which can be specialized to solve all of the above mentioned problems. In particular, we show how to compute local flow, global (parametric) flow, rigid flow resulting from camera egomotion, and multiframe versions of the above problems. Using a spline-based description of the flow removes the need for overlapping correlation windows, and produces an explicit measure of the correlation between adjacent flow estimates. We demonstrate our algorithm on multiframe image registration and the recovery of 3D projective scene geometry. We also provide results on a number of standard motion sequences.

Journal ArticleDOI
TL;DR: Fundamental concepts of color perception and measurement are first presented using vector-space notation and terminology in order to establish the background and lay down terminology.
Abstract: This paper surveys current technology and research in the area of digital color imaging. In order to establish the background and lay down terminology, fundamental concepts of color perception and measurement are first presented using vector-space notation and terminology. Present-day color recording and reproduction systems are reviewed along with the common mathematical models used for representing these devices. Algorithms for processing color images for display and communication are surveyed, and a forecast of research trends is attempted. An extensive bibliography is provided.

Patent
17 Jul 1997
TL;DR: In this paper, a method of constructing an image mosaic comprising the steps of selecting source images, aligning the source image, selecting source segments, enhancing the images, and merging the images to form the image mosaic is disclosed.
Abstract: A method of constructing an image mosaic comprising the steps of selecting source images, aligning the source images, selecting source segments, enhancing the images, and merging the images to form the image mosaic is disclosed. An apparatus for constructing an image mosaic comprising means for selecting source images (102), means for aligning the source images (103), means for selecting source image segments (104), means for enhancing the images (105), and means for merging (106) the images to form the image mosaic is also disclosed. The process may be performed automatically by the system or may be guided interactively by a human operator. Applications include the construction of photographic quality prints from video and digital camera images.

Journal ArticleDOI
TL;DR: Experimental results with real video demonstrate that a significant increase in the image resolution can be achieved by taking the motion blurring into account especially when there exists large interframe motion.
Abstract: Printing from an NTSC source and conversion of NTSC source material to high-definition television (HDTV) format are some of the applications that motivate superresolution (SR) image and video reconstruction from low-resolution (LR) and possibly blurred sources. Existing methods for SR image reconstruction are limited by the assumptions that the input LR images are sampled progressively, and that the aperture time of the camera is zero, thus ignoring the motion blur occurring during the aperture time. Because of the observed adverse effects of these assumptions for many common video sources, this paper proposes (i) a complete model of video acquisition with an arbitrary input sampling lattice and a nonzero aperture time, and (ii) an algorithm based on this model using the theory of projections onto convex sets to reconstruct SR still images or video from an LR time sequence of images. Experimental results with real video are provided, which clearly demonstrate that a significant increase in the image resolution can be achieved by taking the motion blurring into account especially when there exists large interframe motion.

Journal ArticleDOI
TL;DR: This work has developed a rapid and automatic method for performing affine registrations that uses a Bayesian scheme to incorporate prior knowledge of the variability in the shape and size of heads.