scispace - formally typeset
Search or ask a question

Showing papers on "Image resolution published in 2010"


Journal ArticleDOI
TL;DR: This paper presents a new approach to single-image superresolution, based upon sparse signal representation, which generates high-resolution images that are competitive or even superior in quality to images produced by other similar SR methods.
Abstract: This paper presents a new approach to single-image superresolution, based upon sparse signal representation. Research on image statistics suggests that image patches can be well-represented as a sparse linear combination of elements from an appropriately chosen over-complete dictionary. Inspired by this observation, we seek a sparse representation for each patch of the low-resolution input, and then use the coefficients of this representation to generate the high-resolution output. Theoretical results from compressed sensing suggest that under mild conditions, the sparse representation can be correctly recovered from the downsampled signals. By jointly training two dictionaries for the low- and high-resolution image patches, we can enforce the similarity of sparse representations between the low-resolution and high-resolution image patch pair with respect to their own dictionaries. Therefore, the sparse representation of a low-resolution image patch can be applied with the high-resolution image patch dictionary to generate a high-resolution image patch. The learned dictionary pair is a more compact representation of the patch pairs, compared to previous approaches, which simply sample a large amount of image patch pairs , reducing the computational cost substantially. The effectiveness of such a sparsity prior is demonstrated for both general image super-resolution (SR) and the special case of face hallucination. In both cases, our algorithm generates high-resolution images that are competitive or even superior in quality to images produced by other similar SR methods. In addition, the local sparse modeling of our approach is naturally robust to noise, and therefore the proposed algorithm can handle SR with noisy inputs in a more unified framework.

4,958 citations


Journal ArticleDOI
TL;DR: The MPRAGE sequence was modified to generate two different images at different inversion times, MP2RAGE, to create T(1)-weighted images where the result image was free of proton density contrast, T(2) contrast, reception bias field, and, to first order, transmit field inhomogeneity.

1,041 citations


Journal ArticleDOI
TL;DR: Compared with existing algorithms, KRR leads to a better generalization than simply storing the examples as has been done in existing example-based algorithms and results in much less noisy images.
Abstract: This paper proposes a framework for single-image super-resolution. The underlying idea is to learn a map from input low-resolution images to target high-resolution images based on example pairs of input and output images. Kernel ridge regression (KRR) is adopted for this purpose. To reduce the time complexity of training and testing for KRR, a sparse solution is found by combining the ideas of kernel matching pursuit and gradient descent. As a regularized solution, KRR leads to a better generalization than simply storing the examples as has been done in existing example-based algorithms and results in much less noisy images. However, this may introduce blurring and ringing artifacts around major edges as sharp changes are penalized severely. A prior model of a generic image class which takes into account the discontinuity property of images is adopted to resolve this problem. Comparison with existing algorithms shows the effectiveness of the proposed method.

938 citations


Journal ArticleDOI
TL;DR: A comprehensive optimization method to arrive at the spatial and spectral layout of the color filter array of a GAP camera is presented and a novel algorithm for reconstructing the under-sampled channels of the image while minimizing aliasing artifacts is developed.
Abstract: We propose the concept of a generalized assorted pixel (GAP) camera, which enables the user to capture a single image of a scene and, after the fact, control the tradeoff between spatial resolution, dynamic range and spectral detail. The GAP camera uses a complex array (or mosaic) of color filters. A major problem with using such an array is that the captured image is severely under-sampled for at least some of the filter types. This leads to reconstructed images with strong aliasing. We make four contributions in this paper: 1) we present a comprehensive optimization method to arrive at the spatial and spectral layout of the color filter array of a GAP camera. 2) We develop a novel algorithm for reconstructing the under-sampled channels of the image while minimizing aliasing artifacts. 3) We demonstrate how the user can capture a single image and then control the tradeoff of spatial resolution to generate a variety of images, including monochrome, high dynamic range (HDR) monochrome, RGB, HDR RGB, and multispectral images. 4) Finally, the performance of our GAP camera has been verified using extensive simulations that use multispectral images of real world scenes. A large database of these multispectral images has been made available at http://wwwl.cs.columbia.edu/ CAVE/projects/gap_camera/ for use by the research community.

833 citations


Journal ArticleDOI
TL;DR: Digital holography is an emerging field of new paradigm in general imaging applications as discussed by the authors, and a review of a subset of the research and development activities in digital holographic microscopy techniques and applications is presented.
Abstract: Digital holography is an emerging field of new paradigm in general imaging applications. We present a review of a subset of the research and development activities in digital holography, with emphasis on microscopy techniques and applications. First, the basic results from the general theory of holography, based on the scalar diffraction theory, are summarized, and a general description of the digital holographic microscopy process is given, including quantitative phase microscopy. Several numerical diffraction methods are described and compared, and a number of representative configurations used in digital holography are described, including off-axis Fresnel, Fourier, image plane, in-line, Gabor, and phase-shifting digital holographies. Then we survey numerical techniques that give rise to unique capabilities of digital holography, including suppression of dc and twin image terms, pixel resolution control, optical phase unwrapping, aberration compensation, and others. A survey is also given of representative application areas, including biomedical microscopy, particle field holography, micrometrology, and holographic tomography, as well as some of the special techniques, such as holography of total internal reflection, optical scanning holography, digital interference holography, and heterodyne holography. The review is intended for students and new researchers interested in developing new techniques and exploring new applications of digital holography.

672 citations


Journal ArticleDOI
Bin Yang1, Shutao Li1
TL;DR: A sparse representation-based multifocus image fusion method that can simultaneously resolve the image restoration and fusion problem by changing the approximate criterion in the sparse representation algorithm is proposed.
Abstract: To obtain an image with every object in focus, we always need to fuse images taken from the same view point with different focal settings. Multiresolution transforms, such as pyramid decomposition and wavelet, are usually used to solve this problem. In this paper, a sparse representation-based multifocus image fusion method is proposed. In the method, first, the source image is represented with sparse coefficients using an overcomplete dictionary. Second, the coefficients are combined with the choose-max fusion rule. Finally, the fused image is reconstructed from the combined sparse coefficients and the dictionary. Furthermore, the proposed fusion scheme can simultaneously resolve the image restoration and fusion problem by changing the approximate criterion in the sparse representation algorithm. The proposed method is compared with spatial gradient (SG)-, morphological wavelet transform (MWT)-, discrete wavelet transform (DWT)-, stationary wavelet transform (SWT)-, curvelet transform (CVT)-, and nonsubsampling contourlet transform (NSCT)-based methods on several pairs of multifocus images. The experimental results demonstrate that the proposed approach performs better in both subjective and objective qualities.

571 citations


Journal ArticleDOI
TL;DR: In this article, compressive sensing (CS) methods for tomographic reconstruction of a building complex from the TerraSAR-X spotlight data are presented, and the theory of 4-D (differential, i.e., space-time) CS TomoSAR and compares it with parametric (nonlinear least squares) and nonparametric (singular value decomposition) reconstruction methods.
Abstract: Synthetic aperture radar (SAR) tomography (TomoSAR) extends the synthetic aperture principle into the elevation direction for 3-D imaging. The resolution in the elevation direction depends on the size of the elevation aperture, i.e., on the spread of orbit tracks. Since the orbits of modern meter-resolution spaceborne SAR systems, like TerraSAR-X, are tightly controlled, the tomographic elevation resolution is at least an order of magnitude lower than in range and azimuth. Hence, super-resolution reconstruction algorithms are desired. The high anisotropy of the 3-D tomographic resolution element renders the signals sparse in the elevation direction; only a few pointlike reflections are expected per azimuth-range cell. This property suggests using compressive sensing (CS) methods for tomographic reconstruction. This paper presents the theory of 4-D (differential, i.e., space-time) CS TomoSAR and compares it with parametric (nonlinear least squares) and nonparametric (singular value decomposition) reconstruction methods. Super-resolution properties and point localization accuracies are demonstrated using simulations and real data. A CS reconstruction of a building complex from TerraSAR-X spotlight data is presented.

460 citations


Journal ArticleDOI
TL;DR: A sub-pixel shifting based super-resolution algorithm is implemented to effectively recover much higher resolution digital holograms of the objects, permitting sub-micron spatial resolution to be achieved across the entire sensor chip active area.
Abstract: We demonstrate lensfree holographic microscopy on a chip to achieve approximately 0.6 microm spatial resolution corresponding to a numerical aperture of approximately 0.5 over a large field-of-view of approximately 24 mm2. By using partially coherent illumination from a large aperture (approximately 50 microm), we acquire lower resolution lensfree in-line holograms of the objects with unit fringe magnification. For each lensfree hologram, the pixel size at the sensor chip limits the spatial resolution of the reconstructed image. To circumvent this limitation, we implement a sub-pixel shifting based super-resolution algorithm to effectively recover much higher resolution digital holograms of the objects, permitting sub-micron spatial resolution to be achieved across the entire sensor chip active area, which is also equivalent to the imaging field-of-view (24 mm2) due to unit magnification. We demonstrate the success of this pixel super-resolution approach by imaging patterned transparent substrates, blood smear samples, as well as Caenoharbditis Elegans.

454 citations


Journal ArticleDOI
TL;DR: First 3-D and 4-D reconstructions of an entire building complex with very high level of detail from spaceborne SAR data by pixelwise TomoSAR are presented.
Abstract: Synthetic aperture radar tomography (TomoSAR) extends the synthetic aperture principle into the elevation direction for 3-D imaging. It uses stacks of several acquisitions from slightly different viewing angles (the elevation aperture) to reconstruct the reflectivity function along the elevation direction by means of spectral analysis for every azimuth-range pixel. The new class of meter-resolution spaceborne SAR systems (TerraSAR-X and COSMO-Skymed) offers a tremendous improvement in tomographic reconstruction of urban areas and man-made infrastructure. The high resolution fits well to the inherent scale of buildings (floor height, distance of windows, etc.). This paper demonstrates the tomographic potential of these SARs and the achievable quality on the basis of TerraSAR-X spotlight data of urban environment. A new Wiener-type regularization to the singular-value decomposition method-equivalent to a maximum a posteriori estimator-for TomoSAR is introduced and is extended to the differential case (4-D, i.e., space-time). Different model selection schemes for the estimation of the number of scatterers in a resolution cell are compared and proven to be applicable in practice. Two parametric estimation algorithms of the scatterers' elevation and their velocities are evaluated. First 3-D and 4-D reconstructions of an entire building complex (including its radar reflectivity) with very high level of detail from spaceborne SAR data by pixelwise TomoSAR are presented.

411 citations


Journal ArticleDOI
TL;DR: Two new modifications to improve the spectral quality of the Intensity-Hue-Saturation method are introduced and an adaptive IHS is proposed that incorporates these two techniques.
Abstract: The goal of pan-sharpening is to fuse a low spatial resolution multispectral image with a higher resolution panchromatic image to obtain an image with high spectral and spatial resolution. The Intensity-Hue-Saturation (IHS) method is a popular pan-sharpening method used for its efficiency and high spatial resolution. However, the final image produced experiences spectral distortion. In this letter, we introduce two new modifications to improve the spectral quality of the image. First, we propose image-adaptive coefficients for IHS to obtain more accurate spectral resolution. Second, an edge-adaptive IHS method was proposed to enforce spectral fidelity away from the edges. Experimental results show that these two modifications improve spectral resolution compared to the original IHS and we propose an adaptive IHS that incorporates these two techniques. The adaptive IHS method produces images with higher spectral resolution while maintaining the high-quality spatial resolution of the original IHS.

390 citations


Journal ArticleDOI
TL;DR: A unique method for real‐time MRI that reduces image acquisition times to only 20 ms is described, approaching the ultimate limit of MRI technology, and yields high image quality in terms of spatial resolution, signal‐to‐noise ratio and the absence of artifacts.
Abstract: The desire to visualize noninvasively physiological processes at high temporal resolution has been a driving force for the development of MRI since its inception in 1973. In this article, we describe a unique method for real-time MRI that reduces image acquisition times to only 20 ms. Although approaching the ultimate limit of MRI technology, the method yields high image quality in terms of spatial resolution, signal-to-noise ratio and the absence of artifacts. As proposed previously, a fast low-angle shot (FLASH) gradient-echo MRI technique (which allows for rapid and continuous image acquisitions) is combined with a radial encoding scheme (which offers motion robustness and moderate tolerance to data undersampling) and, most importantly, an iterative image reconstruction by regularized nonlinear inversion (which exploits the advantages of parallel imaging with multiple receiver coils). In this article, the extension of regularization and filtering to the temporal domain exploits consistencies in successive data acquisitions and thereby enhances the degree of radial undersampling in a hitherto unexpected manner by one order of magnitude. The results obtained for turbulent flow, human speech production and human heart function demonstrate considerable potential for real-time MRI studies of dynamic processes in a wide range of scientific and clinical settings. Copyright © 2010 John Wiley & Sons, Ltd.

Journal ArticleDOI
TL;DR: Time series of images from FORMOSAT-2 and LANDSAT are used to develop and test a Multi-Temporal Cloud Detection (MTCD) method and results show that the MTCD method provides a better discrimination of clouded and unclouded pixels than the usual methods based on thresholds applied to reflectances or reflectance ratios.

Proceedings ArticleDOI
13 Jun 2010
TL;DR: This paper proposes an approach to extend edge-directed super-resolution to include detail from an image/texture example provided by the user (e.g., from the Internet), and can achieve quality results at very large magnification, which is often problematic for both edge- directed and learning-based approaches.
Abstract: Edge-directed image super resolution (SR) focuses on ways to remove edge artifacts in upsampled images. Under large magnification, however, textured regions become blurred and appear homogenous, resulting in a super-resolution image that looks unnatural. Alternatively, learning-based SR approaches use a large database of exemplar images for “hallucinating” detail. The quality of the upsampled image, especially about edges, is dependent on the suitability of the training images. This paper aims to combine the benefits of edge-directed SR with those of learning-based SR. In particular, we propose an approach to extend edge-directed super-resolution to include detail from an image/texture example provided by the user (e.g., from the Internet). A significant benefit of our approach is that only a single exemplar image is required to supply the missing detail – strong edges are obtained in the SR image even if they are not present in the example image due to the combination of the edge-directed approach. In addition, we can achieve quality results at very large magnification, which is often problematic for both edge-directed and learning-based approaches.

Journal ArticleDOI
TL;DR: Experiments show that the proposed method without residue compensation generates higher-quality images and costs less computational time than some recent face image super-resolution (hallucination) techniques.

Proceedings ArticleDOI
13 Jun 2010
TL;DR: It is shown the surprising result that 3D scans of reasonable quality can also be obtained with a sensor of such low data quality, and a new combination of a 3D superresolution method with a probabilistic scan alignment approach that explicitly takes into account the sensor's noise characteristics.
Abstract: We describe a method for 3D object scanning by aligning depth scans that were taken from around an object with a time-of-flight camera. These ToF cameras can measure depth scans at video rate. Due to comparably simple technology they bear potential for low cost production in big volumes. Our easy-to-use, cost-effective scanning solution based on such a sensor could make 3D scanning technology more accessible to everyday users. The algorithmic challenge we face is that the sensor's level of random noise is substantial and there is a non-trivial systematic bias. In this paper we show the surprising result that 3D scans of reasonable quality can also be obtained with a sensor of such low data quality. Established filtering and scan alignment techniques from the literature fail to achieve this goal. In contrast, our algorithm is based on a new combination of a 3D superresolution method with a probabilistic scan alignment approach that explicitly takes into account the sensor's noise characteristics.

Journal ArticleDOI
TL;DR: This work analyzes the focused plenoptic camera in optical phase space and presents basic, blended, and depth-based rendering algorithms for producing high-quality, high-resolution images in real time.
Abstract: Plenoptic cameras, constructed with internal microlens arrays, capture both spatial and angular information, i.e., the full 4-D radiance, of a scene. The design of traditional plenoptic cameras assumes that each microlens image is completely defocused with respect to the image created by the main camera lens. As a result, only a single pixel in the final image is rendered from each microlens image, resulting in disappointingly low resolution. A recently devel- oped alternative approach based on the focused plenoptic camera uses the microlens array as an imaging system focused on the im- age plane of the main camera lens. The flexible spatioangular trade- off that becomes available with this design enables rendering of final images with significantly higher resolution than those from traditional plenoptic cameras. We analyze the focused plenoptic camera in optical phase space and present basic, blended, and depth-based rendering algorithms for producing high-quality, high-resolution im- ages. We also present our graphics-processing-unit-based imple- mentations of these algorithms, which are able to render full screen refocused images in real time. © 2010 SPIE and IS&T.

Journal ArticleDOI
TL;DR: The proposed algorithm significantly improves the fusion quality in terms of: entropy, mutual information, discrepancy, and average gradient; compared to fusion methods including, IHS, Brovey, discrete wavelet transform (DWT), a-trous wavelet and RIM.

Journal ArticleDOI
TL;DR: An improved version of CS-based high-resolution imaging to overcome strong noise and clutter by combining coherent projectors and weighting with the CS optimization for ISAR image generation is presented.
Abstract: The theory of compressed sampling (CS) indicates that exact recovery of an unknown sparse signal can be achieved from very limited samples. For inversed synthetic aperture radar (ISAR), the image of a target is usually constructed by strong scattering centers whose number is much smaller than that of pixels of an image plane. This sparsity of the ISAR signal intrinsically paves a way to apply CS to the reconstruction of high-resolution ISAR imagery. CS-based high-resolution ISAR imaging with limited pulses is developed, and it performs well in the case of high signal-to-noise ratios. However, strong noise and clutter are usually inevitable in radar imaging, which challenges current high-resolution imaging approaches based on parametric modeling, including the CS-based approach. In this paper, we present an improved version of CS-based high-resolution imaging to overcome strong noise and clutter by combining coherent projectors and weighting with the CS optimization for ISAR image generation. Real data are used to test the robustness of the improved CS imaging compared with other current techniques. Experimental results show that the approach is capable of precise estimation of scattering centers and effective suppression of noise.

Journal ArticleDOI
TL;DR: Two new regularization items are proposed, termed as locally adaptive bilateral total variation and consistency of gradients, to keep edges and flat regions, which are implicitly described in LR images, sharp and smooth, respectively, respectively.

Journal ArticleDOI
TL;DR: A new upsampling method is proposed to recover some of this high frequency information by using a data-adaptive patch-based reconstruction in combination with a subsampling coherence constraint to outperform classical interpolation methods in terms of quantitative measures and visual observation.

Journal ArticleDOI
TL;DR: It can not only be proven that the Ehlers fusion is superior to all other tested algorithms, it is also the only method that guarantees excellent colour preservation for all dates and sensors used in this study.
Abstract: The main objective of this article is quality assessment of pansharpening fusion methods. Pansharpening is a fusion technique to combine a panchromatic image of high spatial resolution with multispectral image data of lower spatial resolution to obtain a high-resolution multispectral image. During this process, the significant spectral characteristics of the multispectral data should be preserved. For images acquired at the same time by the same sensor, most algorithms for pansharpening provide very good results, i.e. they retain the high spatial resolution of the panchromatic image and the spectral information from the multispectral image (single-sensor, single-date fusion). For multi-date, multi-sensor fusion, however, these techniques can still create spatially enhanced data sets, but usually at the expense of the spectral consistency. In this study, eight different methods are compared for image fusion to show their ability to fuse multitemporal and multi-sensor image data. A series of eight multitemp...

Journal ArticleDOI
TL;DR: The quantitative peak signal‐to‐noise ratio (PSNR) and visual results show the superiority of the proposed technique over the conventional and state‐of‐art image resolution enhancement techniques.
Abstract: In this paper, we propose a new super-resolution technique based on interpolation of the high-frequency subband images obtained by discrete wavelet transform (DWT) and the input image. The proposed technique uses DWT to decompose an image into different subband images. Then the high-frequency subband images and the input low-resolution image have been interpolated, followed by combining all these images to generate a new super-resolved image by using inverse DWT. The proposed technique has been tested on Lena, Elaine, Pepper, and Baboon. The quantitative peak signal-to-noise ratio (PSNR) and visual results show the superiority of the proposed technique over the conventional and state-of-art image resolution enhancement techniques. For Lena's image, the PSNR is 7.93 dB higher than the bicubic interpolation.

Journal ArticleDOI
TL;DR: A snapshot Image Mapping Spectrometer (IMS) with high sampling density is developed for hyperspectral microscopy, measuring a datacube of dimensions 285 × 285 × 60 (x, y, λ).
Abstract: A snapshot Image Mapping Spectrometer (IMS) with high sampling density is developed for hyperspectral microscopy, measuring a datacube of dimensions 285 × 285 × 60 (x, y, λ). The spatial resolution is ~0.45 µm with a FOV of 100 × 100 µm2. The measured spectrum is from 450 nm to 650 nm and is sampled by 60 spectral channels with average sampling interval ~3.3 nm. The channel’s spectral resolution is ~8nm. The spectral imaging results demonstrate the potential of the IMS for real-time cellular fluorescence imaging.

Proceedings ArticleDOI
13 Jun 2010
TL;DR: The proposed retrieval framework works well in image retrieval task owing to the encoding of geometric information of objects for capturing objects' spatial transformation, the supervised feature selection and combination strategy for enhancing the discriminative power, and the representation of bag-of-features for effective image matching and indexing for large scale image retrieval.
Abstract: In this paper, we study the problem of large scale image retrieval by developing a new class of bag-of-features to encode geometric information of objects within an image. Beyond existing orderless bag-of-features, local features of an image are first projected to different directions or points to generate a series of ordered bag-of-features, based on which different families of spatial bag-of-features are designed to capture the invariance of object translation, rotation, and scaling. Then the most representative features are selected based on a boosting-like method to generate a new bag-of-features-like vector representation of an image. The proposed retrieval framework works well in image retrieval task owing to the following three properties: 1) the encoding of geometric information of objects for capturing objects' spatial transformation, 2) the supervised feature selection and combination strategy for enhancing the discriminative power, and 3) the representation of bag-of-features for effective image matching and indexing for large scale image retrieval. Extensive experiments on 5000 Oxford building images and 1 million Panoramio images show the effectiveness and efficiency of the proposed features as well as the retrieval framework.

Journal ArticleDOI
TL;DR: A resolution progressive compression scheme which compresses an encrypted image progressively in resolution, such that the decoder can observe a low-resolution version of the image, study local statistics based on it, and use the statistics to decode the next resolution level.
Abstract: Lossless compression of encrypted sources can be achieved through Slepian-Wolf coding. For encrypted real-world sources, such as images, the key to improve the compression efficiency is how the source dependency is exploited. Approaches in the literature that make use of Markov properties in the Slepian-Wolf decoder do not work well for grayscale images. In this correspondence, we propose a resolution progressive compression scheme which compresses an encrypted image progressively in resolution, such that the decoder can observe a low-resolution version of the image, study local statistics based on it, and use the statistics to decode the next resolution level. Good performance is observed both theoretically and experimentally.

Journal ArticleDOI
TL;DR: Improved resolution and contrast versus noise properties can be achieved with the proposed method with similar computation time as the conventional approach, and comparison of the measured spatially variant and invariant reconstruction revealed similar performance with conventional image metrics.
Abstract: Accurate system modeling in tomographic image reconstruction has been shown to reduce the spatial variance of resolution and improve quantitative accuracy. System modeling can be improved through analytic calculations, Monte Carlo simulations, and physical measurements. The purpose of this work is to improve clinical fully-3-D reconstruction without substantially increasing computation time. We present a practical method for measuring the detector blurring component of a whole-body positron emission tomography (PET) system to form an approximate system model for use with fully-3-D reconstruction. We employ Monte Carlo simulations to show that a non-collimated point source is acceptable for modeling the radial blurring present in a PET tomograph and we justify the use of a Na22 point source for collecting these measurements. We measure the system response on a whole-body scanner, simplify it to a 2-D function, and incorporate a parameterized version of this response into a modified fully-3-D OSEM algorithm. Empirical testing of the signal versus noise benefits reveal roughly a 15% improvement in spatial resolution and 10% improvement in contrast at matched image noise levels. Convergence analysis demonstrates improved resolution and contrast versus noise properties can be achieved with the proposed method with similar computation time as the conventional approach. Comparison of the measured spatially variant and invariant reconstruction revealed similar performance with conventional image metrics. Edge artifacts, which are a common artifact of resolution-modeled reconstruction methods, were less apparent in the spatially variant method than in the invariant method. With the proposed and other resolution-modeled reconstruction methods, edge artifacts need to be studied in more detail to determine the optimal tradeoff of resolution/contrast enhancement and edge fidelity.

Journal ArticleDOI
TL;DR: The quantitative peak signal-to-noise ratio (PSNR) and visual results show the superiority of the proposed technique over the conventional bicubic interpolation, wavelet zero padding, and Irani and Peleg based image resolution enhancement techniques.
Abstract: In this letter, a satellite image resolution enhancement technique based on interpolation of the high-frequency subband images obtained by dual-tree complex wavelet transform (DT-CWT) is proposed. DT-CWT is used to decompose an input low-resolution satellite image into different subbands. Then, the high-frequency subband images and the input image are interpolated, followed by combining all these images to generate a new high-resolution image by using inverse DT-CWT. The resolution enhancement is achieved by using directional selectivity provided by the CWT, where the high-frequency subbands in six different directions contribute to the sharpness of the high-frequency details such as edges. The quantitative peak signal-to-noise ratio (PSNR) and visual results show the superiority of the proposed technique over the conventional bicubic interpolation, wavelet zero padding, and Irani and Peleg based image resolution enhancement techniques.

Proceedings ArticleDOI
13 Jun 2010
TL;DR: This work presents a flexible method for fusing information from optical and range sensors based on an accelerated high-dimensional filtering approach, and describes how to integrate priors on object motion and appearance and how to achieve an efficient implementation using parallel processing hardware such as GPUs.
Abstract: We present a flexible method for fusing information from optical and range sensors based on an accelerated high-dimensional filtering approach. Our system takes as input a sequence of monocular camera images as well as a stream of sparse range measurements as obtained from a laser or other sensor system. In contrast with existing approaches, we do not assume that the depth and color data streams have the same data rates or that the observed scene is fully static. Our method produces a dense, high-resolution depth map of the scene, automatically generating confidence values for every interpolated depth point. We describe how to integrate priors on object motion and appearance and how to achieve an efficient implementation using parallel processing hardware such as GPUs.

Journal ArticleDOI
01 Jul 2010
TL;DR: This paper presents a new method for extracting roads in Very High Resolution remotely sensed images based on advanced directional morphological operators that outperform standard approaches using rotating rectangular structuring elements.
Abstract: Very high spatial resolution (VHR) images allow to feature man-made structures such as roads and thus enable their accurate analysis. Geometrical characteristics can be extracted using mathematical morphology. However, the prior choice of a reference shape (structuring element) introduces a shape-bias. This paper presents a new method for extracting roads in Very High Resolution remotely sensed images based on advanced directional morphological operators. The proposed approach introduces the use of Path Openings and Path Closings in order to extract structural pixel information. These morphological operators remain flexible enough to fit rectilinear and slightly curved structures since they do not depend on the choice of a structural element shape. As a consequence, they outperform standard approaches using rotating rectangular structuring elements. The method consists in building a granulometry chain using Path Openings and Path Closing to construct Morphological Profiles. For each pixel, the Morphological Profile constitutes the feature vector on which our road extraction is based.

Journal ArticleDOI
TL;DR: An approach for automatic vehicle detection from optical satellite images using implicit modeling and the use of a priori knowledge of typical vehicle constellation leads to an enhanced overall completeness compared to approaches which are only based on statistical classification techniques.
Abstract: Current traffic research is mostly based on data from fixed-installed sensors like induction loops, bridge sensors, and cameras. Thereby, the traffic flow on main roads can partially be acquired, while data from the major part of the entire road network are not available. Today's optical sensor systems on satellites provide large-area images with 1-m resolution and better, which can deliver complement information to traditional acquired data. In this paper, we present an approach for automatic vehicle detection from optical satellite images. Therefore, hypotheses for single vehicles are generated using adaptive boosting in combination with Haar-like features. Additionally, vehicle queues are detected using a line extraction technique since grouped vehicles are merged to either dark or bright ribbons. Utilizing robust parameter estimation, single vehicles are determined within those vehicle queues. The combination of implicit modeling and the use of a priori knowledge of typical vehicle constellation leads to an enhanced overall completeness compared to approaches which are only based on statistical classification techniques. Thus, a detection rate of over 80% is possible with very high reliability. Furthermore, an approach for movement estimation of the detected vehicle is described, which allows the distinction of moving and stationary traffic. Thus, even an estimate for vehicles' speed is possible, which gives additional information about the traffic condition at image acquisition time.