A. N. Rajagopalan
Other affiliations: University of Maryland, College Park, Technische Universität München, Politehnica University of Timișoara ...read more
Bio: A. N. Rajagopalan is an academic researcher from Indian Institute of Technology Madras. The author has contributed to research in topics: Image restoration & Motion blur. The author has an hindex of 33, co-authored 199 publications receiving 4537 citations. Previous affiliations of A. N. Rajagopalan include University of Maryland, College Park & Technische Universität München.
Papers published on a yearly basis
TL;DR: A view-based approach to recognize humans from their gait by employing a hidden Markov model (HMM) and the statistical nature of the HMM lends overall robustness to representation and recognition.
Abstract: We propose a view-based approach to recognize humans from their gait. Two different image features have been considered: the width of the outer contour of the binarized silhouette of the walking person and the entire binary silhouette itself. To obtain the observation vector from the image features, we employ two different methods. In the first method, referred to as the indirect approach, the high-dimensional image feature is transformed to a lower dimensional space by generating what we call the frame to exemplar (FED) distance. The FED vector captures both structural and dynamic traits of each individual. For compact and effective gait representation and recognition, the gait information in the FED vector sequences is captured in a hidden Markov model (HMM). In the second method, referred to as the direct approach, we work with the feature vector directly (as opposed to computing the FED) and train an HMM. We estimate the HMM parameters (specifically the observation probability B) based on the distance between the exemplars and the image features. In this way, we avoid learning high-dimensional probability density functions. The statistical nature of the HMM lends overall robustness to representation and recognition. The performance of the methods is illustrated using several databases.
26 Mar 1999
TL;DR: A Partial Derivatives of Various Quantities in CRB, a MAP-MRF approach to Depth Recovery and Restoration using MRF Models.
Abstract: 1 Passive Methods for Depth Recovery.- 1.1 Introduction.- 1.2 Different Methods of Depth Recovery.- 1.2.1 Depth from Stereo.- 1.2.2 Structure from Motion.- 1.2.3 Shape from Shading.- 1.2.4 Range from Focus.- 1.2.5 Depth from Defocus.- 1.3 Difficulties in Passive Ranging.- 1.4 Organization of the Book.- 2 Depth Recovery from Defocused Images.- 2.1 Introduction.- 2.2 Theory of Depth from Defocus.- 2.2.1 Real Aperture Imaging.- 2.2.2 Modeling the Camera Defocus.- 2.2.3 Depth Recovery.- 2.2.4 Sources of Errors.- 2.3 Related Work.- 2.4 Summary of the Book.- 3 Mathematical Background.- 3.1 Introduction.- 3.2 Time-Frequency Representation.- 3.2.1 The Complex Spectrogram.- 3.2.2 The Wigner Distribution.- 3.3 Calculus of Variations.- 3.4 Markov Random Fields and Gibbs Distributions.- 3.4.1 Theory of MRF.- 3.4.2 Gibbs Distribution.- 3.4.3 Incorporating Discontinuities.- 4 Depth Recovery with a Block Shift-Variant Blur Model.- 4.1 Introduction.- 4.2 The Block Shift-Variant Blur Model.- 4.2.1 Estimation of Blur.- 4.2.2 Special Cases.- 4.3 Experimental Results.- 4.4 Discussion.- 5 Space-Variant Filtering Models for Recovering Depth.- 5.1 Introduction.- 5.2 Space-Variant Filtering.- 5.3 Depth Recovery Using the Complex Spectrogram.- 5.4 The Pseudo-Wigner Distribution for Recovery of Depth.- 5.5 Imposing Smoothness Constraint.- 5.5.1 Regularized Solution Using the Complex Spectrogram..- 5.5.2 The Pseudo-Wigner Distribution and Regularized Solution.- 5.6 Experimental Results.- 5.7 Discussion.- 6 ML Estimation of Depth and Optimal Camera Settings.- 6.1 Introduction.- 6.2 Image and Observation Models.- 6.3 ML-Based Recovery of Depth.- 6.4 Computation of the Likelihood Function.- 6.5 Optimality of Camera Settings.- 6.5.1 The Cramer-Rao Bound.- 6.5.2 Optimality Criterion.- 6.6 Experimental Results.- 6.7 Discussion.- 7 Recursive Computation of Depth from Multiple Images.- 7.1 Introduction.- 7.2 Blur Identification from Multiple Images.- 7.3 Minimization by Steepest Descent.- 7.4 Recursive Algorithm for Computing the Likelihood Function.- 7.4.1 Single Observation.- 7.4.2 Two Observations.- 7.4.3 General Case of M Observations.- 7.5 Experimental Results.- 7.6 Discussion.- 8 MRF Model-Based Identification of Shift-Variant PSF.- 8.1 Introduction.- 8.2 A MAP-MRF Approach.- 8.3 The Posterior Distribution and Its Neighborhood.- 8.4 MAP Estimation by Simulated Annealing.- 8.5 Experimental Results.- 8.6 Discussion.- 9 Simultaneous Depth Recovery and Image Restoration.- 9.1 Introduction.- 9.2 Depth Recovery and Restoration using MRF Models.- 9.3 Locality of the Posterior Distribution.- 9.4 Parameter Estimation.- 9.5 Experimental Results.- 9.6 Discussion.- 10 Conclusions.- A Partial Derivatives of Various Quantities in CRB.- References.
TL;DR: The dynamic time-warping (DTW) approach is used for matching so that non-linear time normalization may be used to deal with the naturally-occuring changes in walking speed.
Abstract: Human gait is an attractive modality for recognizing people at a distance. In this paper we adopt an appearance-based approach to the problem of gait recognition. The width of the outer contour of the binarized silhouette of a walking person is chosen as the basic image feature. Different gait features are extracted from the width vector such as the dowsampled, smoothed width vectors, the velocity profile etc. and sequences of such temporally ordered feature vectors are used for representing a person's gait. We use the dynamic time-warping (DTW) approach for matching so that non-linear time normalization may be used to deal with the naturally-occuring changes in walking speed. The performance of the proposed method is tested using different gait databases.
••20 May 2002
TL;DR: This paper proposes a view-based approach to recognize humans through gait using a continuous hidden Markov model that serves to compactly capture structural and transitional features that are unique to an individual.
Abstract: Gait is a spatio-temporal phenomenon that typifies the motion characteristics of an individual. In this paper, we propose a view-based approach to recognize humans through gait. The width of the outer contour of the binarized silhouette of a walking person is chosen as the image feature. A set of stances or key frames that occur during the walk cycle of an individual is chosen. Euclidean distances of a given image from this stance set are computed and a lower-dimensional observation vector is generated. A continuous hidden Markov model (HMM) is trained using several such lower-dimensional vector sequences extracted from the video. This methodology serves to compactly capture structural and transitional features that are unique to an individual. The statistical nature of the HMM renders overall robustness to gait representation and recognition. The human identification performance of the proposed scheme is found to be quite good when tested in natural walking conditions.
••14 Jun 2020
TL;DR: This work proposes an efficient pixel adaptive and feature attentive design for handling large blur variations across different spatial locations and proposes an effective content-aware global-local filtering module that significantly improves performance by considering not only global dependencies but also by dynamically exploiting neighboring pixel information.
Abstract: This paper tackles the problem of motion deblurring of dynamic scenes. Although end-to-end fully convolutional designs have recently advanced the state-of-the-art in non-uniform motion deblurring, their performance-complexity trade-off is still sub-optimal. Existing approaches achieve a large receptive field by increasing the number of generic convolution layers and kernel-size, but this comesat the expense of of the increase in model size and inference speed. In this work, we propose an efficient pixel adaptive and feature attentive design for handling large blur variations across different spatial locations and process each test image adaptively. We also propose an effective content-aware global-local filtering module that significantly improves performance by considering not only global dependencies but also by dynamically exploiting neighboring pixel information. We use a patch-hierarchical attentive architecture composed of the above module that implicitly discovers the spatial variations in the blur present in the input image and in turn, performs local and global modulation of intermediate features. Extensive qualitative and quantitative comparisons with prior art on deblurring benchmarks demonstrate that our design offers significant improvements over the state-of-the-art in accuracy as well as speed.
TL;DR: In this article, the authors categorize and evaluate face detection algorithms and discuss relevant issues such as data collection, evaluation metrics and benchmarking, and conclude with several promising directions for future research.
Abstract: Images containing faces are essential to intelligent vision-based human-computer interaction, and research efforts in face processing include face recognition, face tracking, pose estimation and expression recognition. However, many reported methods assume that the faces in an image or an image sequence have been identified and localized. To build fully automated systems that analyze the information contained in face images, robust and efficient face detection algorithms are required. Given a single image, the goal of face detection is to identify all image regions which contain a face, regardless of its 3D position, orientation and lighting conditions. Such a problem is challenging because faces are non-rigid and have a high degree of variability in size, shape, color and texture. Numerous techniques have been developed to detect faces in a single image, and the purpose of this paper is to categorize and evaluate these algorithms. We also discuss relevant issues such as data collection, evaluation metrics and benchmarking. After analyzing these algorithms and identifying their limitations, we conclude with several promising directions for future research.
01 Jan 2006
01 Apr 2003
TL;DR: The EnKF has a large user group, and numerous publications have discussed applications and theoretical aspects of it as mentioned in this paper, and also presents new ideas and alternative interpretations which further explain the success of the EnkF.
Abstract: The purpose of this paper is to provide a comprehensive presentation and interpretation of the Ensemble Kalman Filter (EnKF) and its numerical implementation. The EnKF has a large user group, and numerous publications have discussed applications and theoretical aspects of it. This paper reviews the important results from these studies and also presents new ideas and alternative interpretations which further explain the success of the EnKF. In addition to providing the theoretical framework needed for using the EnKF, there is also a focus on the algorithmic formulation and optimal numerical implementation. A program listing is given for some of the key subroutines. The paper also touches upon specific issues such as the use of nonlinear measurements, in situ profiles of temperature and salinity, and data which are available with high frequency in time. An ensemble based optimal interpolation (EnOI) scheme is presented as a cost-effective approach which may serve as an alternative to the EnKF in some applications. A fairly extensive discussion is devoted to the use of time correlated model errors and the estimation of model bias.
TL;DR: This survey reviews recent trends in video-based human capture and analysis, as well as discussing open problems for future research to achieve automatic visual analysis of human movement.