scispace - formally typeset
Search or ask a question
Author

Alan C. Bovik

Bio: Alan C. Bovik is an academic researcher from University of Texas at Austin. The author has contributed to research in topics: Image quality & Video quality. The author has an hindex of 102, co-authored 837 publications receiving 96088 citations. Previous affiliations of Alan C. Bovik include University of Illinois at Urbana–Champaign & University of Sydney.


Papers
More filters
Proceedings ArticleDOI
27 Apr 1993
TL;DR: Useful approximations to the responses of discrete linear systems and certain discrete nonlinear systems are developed for input complex AM-FM signals of the form s(m)=a(m) exp (j phi (m)).
Abstract: Useful approximations to the responses of discrete linear systems and certain discrete nonlinear systems are developed for input complex AM-FM signals of the form s(m)=a(m) exp (j phi (m)). These are used to derive limits on simple AM-FM demodulation mechanisms related to the Teager-Kaiser operator. >

19 citations

Book ChapterDOI
01 Dec 2009

19 citations

Journal ArticleDOI
TL;DR: A new image quality assessment (IQA) measure is proposed that supports the visual qualitative analysis of pansharpened outcomes by using the statistics of natural images, commonly referred to as natural scene statistics (NSS), to extract statistical regularities from PS images.
Abstract: Pan-sharpening (PS) is a method of fusing the spatial details of a high-resolution panchromatic (PAN) image with the spectral information of a low-resolution multi-spectral (MS) image. Visual inspection is a crucial step in the evaluation of fused products whose subjectivity renders the assessment of pansharpened data a challenging problem. Most previous research on the development of PS algorithms has only superficially addressed the issue of qualitative evaluation, generally by depicting visual representations of the fused images. Hence, it is highly desirable to be able to predict pan-sharpened image quality automatically and accurately, as it would be perceived and reported by human viewers. Such a method is indispensable for the correct evaluation of PS techniques that produce images for visual applications such as Google Earth and Microsoft Bing. Here, we propose a new image quality assessment (IQA) measure that supports the visual qualitative analysis of pansharpened outcomes by using the statistics of natural images, commonly referred to as natural scene statistics (NSS), to extract statistical regularities from PS images. Importantly, NSS are measurably modified by the presence of distortions. We analyze six PS methods in the presence of two common distortions, blur and white noise, on PAN images. Furthermore, we conducted a human study on the subjective quality of pristine and degraded PS images and created a completely blind (opinion-unaware) fused image quality analyzer. In addition, we propose an opinion-aware fused image quality analyzer, whose predictions with respect to human perceptual evaluations of pansharpened images are highly correlated.

19 citations

Journal ArticleDOI
TL;DR: A vision system for measuring the area of an arbitrarily shaped object that consists of a gray-level thresholding technique combined with a region correction procedure based on mathematical morphology and has been successfully applied to a number of images of medical interest.

19 citations

Book ChapterDOI
29 Apr 2011
TL;DR: This chapter is concerned with a subset of image/video quality assessment algorithms (IQA/VQA) which are referred to as full reference (FR) algorithms; in these algorithms; the original, pristine stimulus is available along with the stimulus whose quality is to be assessed.
Abstract: ‘Quality’ according to the International Standards Organization (ISO) is the degree to which a set of inherent characteristics of a product fulfils customer requirements [1]. Even though this definition seems relatively straightforward at first, introspection leads one to the conclusion that the ambiguity inherent in the definition makes the quality assessment task highly subjective and hence difficult to model. Indeed, over the years researchers in the field of visual quality assessment have found that judging the quality of an image or a video is a challenging task. The highly subjective nature of the task, coupled with the human visual systems’ peculiarities make this an interesting problem to study and in this chapter we attempt to do just that. This chapter is concerned with the algorithmic evaluation of quality of an image or video, which is referred to as objective quality assessment. What makes this task difficult is that the measure of quality produced by the algorithm should match up to that produced by a human assessor. In order to obtain a statistically relevant measure of what a human thinks the quality of an image or video is; a set of images or videos are shown to a group of human observers who are asked to rate the quality on a particular scale. The mean rating for an image or video is referred to as the mean opinion score (MOS) and is representative of the perceptual quality of that visual stimulus. Such assessment of quality is referred to as subjective quality assessment. In order to gauge the performance of an objective algorithm, the scores produced by the algorithm are correlated with MOS; a higher correlation is indicative of better performance. In this chapter, we focus on a subset of image/video quality assessment algorithms (IQA/VQA) which are referred to as full reference (FR) algorithms. In these algorithms; the original, pristine stimulus is available along with the stimulus whose quality is to be assessed. The FR IQA/VQA algorithm accepts as input the pristine reference stimulus and its distorted version and produces a score that is representative of the visual quality of the distorted stimulus [2]. One of the primary questions that arise when we talk of visual quality assessment is : ‘why not use mean square error (MSE) for this purpose?’. MSE

19 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: In this article, a structural similarity index is proposed for image quality assessment based on the degradation of structural information, which can be applied to both subjective ratings and objective methods on a database of images compressed with JPEG and JPEG2000.
Abstract: Objective methods for assessing perceptual image quality traditionally attempted to quantify the visibility of errors (differences) between a distorted image and a reference image using a variety of known properties of the human visual system. Under the assumption that human visual perception is highly adapted for extracting structural information from a scene, we introduce an alternative complementary framework for quality assessment based on the degradation of structural information. As a specific example of this concept, we develop a structural similarity index and demonstrate its promise through a set of intuitive examples, as well as comparison to both subjective ratings and state-of-the-art objective methods on a database of images compressed with JPEG and JPEG2000. A MATLAB implementation of the proposed algorithm is available online at http://www.cns.nyu.edu//spl sim/lcv/ssim/.

40,609 citations

Book
01 Jan 1998
TL;DR: An introduction to a Transient World and an Approximation Tour of Wavelet Packet and Local Cosine Bases.
Abstract: Introduction to a Transient World. Fourier Kingdom. Discrete Revolution. Time Meets Frequency. Frames. Wavelet Zoom. Wavelet Bases. Wavelet Packet and Local Cosine Bases. An Approximation Tour. Estimations are Approximations. Transform Coding. Appendix A: Mathematical Complements. Appendix B: Software Toolboxes.

17,693 citations

Proceedings ArticleDOI
21 Jul 2017
TL;DR: Conditional adversarial networks are investigated as a general-purpose solution to image-to-image translation problems and it is demonstrated that this approach is effective at synthesizing photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks.
Abstract: We investigate conditional adversarial networks as a general-purpose solution to image-to-image translation problems. These networks not only learn the mapping from input image to output image, but also learn a loss function to train this mapping. This makes it possible to apply the same generic approach to problems that traditionally would require very different loss formulations. We demonstrate that this approach is effective at synthesizing photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks. Moreover, since the release of the pix2pix software associated with this paper, hundreds of twitter users have posted their own artistic experiments using our system. As a community, we no longer hand-engineer our mapping functions, and this work suggests we can achieve reasonable results without handengineering our loss functions either.

11,958 citations

Posted Content
TL;DR: Conditional Adversarial Network (CA) as discussed by the authors is a general-purpose solution to image-to-image translation problems, which can be used to synthesize photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks.
Abstract: We investigate conditional adversarial networks as a general-purpose solution to image-to-image translation problems. These networks not only learn the mapping from input image to output image, but also learn a loss function to train this mapping. This makes it possible to apply the same generic approach to problems that traditionally would require very different loss formulations. We demonstrate that this approach is effective at synthesizing photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks. Indeed, since the release of the pix2pix software associated with this paper, a large number of internet users (many of them artists) have posted their own experiments with our system, further demonstrating its wide applicability and ease of adoption without the need for parameter tweaking. As a community, we no longer hand-engineer our mapping functions, and this work suggests we can achieve reasonable results without hand-engineering our loss functions either.

11,127 citations

Journal ArticleDOI
01 Apr 1988-Nature
TL;DR: In this paper, a sedimentological core and petrographic characterisation of samples from eleven boreholes from the Lower Carboniferous of Bowland Basin (Northwest England) is presented.
Abstract: Deposits of clastic carbonate-dominated (calciclastic) sedimentary slope systems in the rock record have been identified mostly as linearly-consistent carbonate apron deposits, even though most ancient clastic carbonate slope deposits fit the submarine fan systems better. Calciclastic submarine fans are consequently rarely described and are poorly understood. Subsequently, very little is known especially in mud-dominated calciclastic submarine fan systems. Presented in this study are a sedimentological core and petrographic characterisation of samples from eleven boreholes from the Lower Carboniferous of Bowland Basin (Northwest England) that reveals a >250 m thick calciturbidite complex deposited in a calciclastic submarine fan setting. Seven facies are recognised from core and thin section characterisation and are grouped into three carbonate turbidite sequences. They include: 1) Calciturbidites, comprising mostly of highto low-density, wavy-laminated bioclast-rich facies; 2) low-density densite mudstones which are characterised by planar laminated and unlaminated muddominated facies; and 3) Calcidebrites which are muddy or hyper-concentrated debrisflow deposits occurring as poorly-sorted, chaotic, mud-supported floatstones. These

9,929 citations