scispace - formally typeset
Search or ask a question
Author

A. G. Ramakrishnan

Bio: A. G. Ramakrishnan is an academic researcher from Indian Institute of Science. The author has contributed to research in topics: Handwriting recognition & Wavelet transform. The author has an hindex of 28, co-authored 220 publications receiving 3111 citations. Previous affiliations of A. G. Ramakrishnan include Birla Institute of Technology and Science.


Papers
More filters
Journal Article
TL;DR: The results suggest that kapalabhati modifies the autonomic status by increasing sympathetic activity with reduced vagal activity and suggests that HRV is a more useful psychophysiological measure than heart rate alone.
Abstract: The heart rate variability (HRV) is an indicator of the cardiac autonomic control. Two spectral components are usually recorded, viz. high frequency (0.15-0.50 Hz), which is due to vagal efferent activity and a low frequency component (0.05-0.15 Hz), due to sympathetic activity. The present study was conducted to study the HRV in two yoga practices which have been previously reported to have opposite effects, viz, sympathetic stimulation (kapalabhati, breathing at high frequency, i.e., 2.0 Hz) and reduced sympathetic activity (nadisuddhi, alternate nostril breathing). Twelve male volunteers (age range, 21 to 33 years) were assessed before and after each practice on separate days. The electrocardiogram (lead I) was digitized on-line and off-line analysis was done. The results showed a significant increase in low frequency (LF) power and LF/HF ratio while high frequency (HF) power was significantly lower following kapalabhati. There were no significant changes following nadisuddhi. The results suggest that kapalabhati modifies the autonomic status by increasing sympathetic activity with reduced vagal activity. The study also suggests that HRV is a more useful psychophysiological measure than heart rate alone.

149 citations

Journal ArticleDOI
TL;DR: The significant feature of the proposed technique is that, while the error is nearly uniform throughout the cycle, the diagnostically crucial QRS region is kept free of maximal reconstruction error.
Abstract: Presents a novel coding scheme for the electrocardiogram (ECG). Following beat delineation, the periods of the beats are normalized by multirate processing. After amplitude normalization, a discrete wavelet transform is applied to each beat. Due to the period and amplitude normalization, the wavelet transform coefficients bear a high correlation across beats at identical locations. To increase the compression ratio, the residual sequence obtained after linear prediction of the significant wavelet coefficients is transmitted to the decoder. The difference between the actual period and the mean beat period, and that between the actual scale factor and the average amplitude scale factor are also transmitted for each beat. At the decoder, the inverse wavelet transform is computed from the reconstructed wavelet transform coefficients. The original amplitude and period of each beat are then recovered. The approximation achieved, at an average rate of 180 b/s, is of high quality. The authors have evaluated the normalized maximum amplitude error and its position in each cycle, in addition to the normalized root mean square error. The significant feature of the proposed technique is that, while the error is nearly uniform throughout the cycle, the diagnostically crucial QRS region is kept free of maximal reconstruction error.

133 citations

Proceedings ArticleDOI
15 Oct 2003
TL;DR: A technique for automating the differential count of blood takes, as input, color images of stained peripheral blood smears and identifies the class of each of the white blood cells (WBC), in order to determine the count of cells in each class.
Abstract: A technique for automating the differential count of blood is presented. The proposed system takes, as input, color images of stained peripheral blood smears and identifies the class of each of the white blood cells (WBC), in order to determine the count of cells in each class. The process involves segmentation, feature extraction and classification. WBC segmentation is a two-step process carried out on the HSV-equivalent of the image, using k-means clustering followed by the EM-algorithm. Features extracted from the segmented cytoplasm and nucleus, are motivated by the visual cues of shape, color and texture. Various classifiers have been explored on different combinations of feature sets. The results presented are based on trials conducted with normal cells. For training the classifiers, a library set of 50 patterns, with about 10 samples from each class, is used. The test data, disjoint from the training set, consists of 34 patterns, fairly represented by every class. The best classification accuracy of 97% is obtained using neural networks, followed by 94% using SVM.

129 citations

Journal ArticleDOI
TL;DR: The combination of Gabor features with nearest neighbor or SVM classifier shows promising results; i.e., over 98% for bi-script and tri-script cases and above 89% for the eleven-script scenario.

123 citations

Journal ArticleDOI
TL;DR: The proposed dynamic plosion index (DPI) algorithm, based on integrated linear prediction residual (ILPR) which resembles the voice source signal, is tested for its robustness in the presence of additive white and babble noise and on simulated telephone quality speech.
Abstract: Epoch is defined as the instant of significant excitation within a pitch period of voiced speech. Epoch extraction continues to attract the interest of researchers because of its significance in speech analysis. Existing high performance epoch extraction algorithms require either dynamic programming techniques or a priori information of the average pitch period. An algorithm without such requirements is proposed based on integrated linear prediction residual (ILPR) which resembles the voice source signal. Half wave rectified and negated ILPR (or Hilbert transform of ILPR) is used as the pre-processed signal. A new non-linear temporal measure named the plosion index (PI) has been proposed for detecting ‘transients’ in speech signal. An extension of PI, called the dynamic plosion index (DPI) is applied on pre-processed signal to estimate the epochs. The proposed DPI algorithm is validated using six large databases which provide simultaneous EGG recordings. Creaky and singing voice samples are also analyzed. The algorithm has been tested for its robustness in the presence of additive white and babble noise and on simulated telephone quality speech. The performance of the DPI algorithm is found to be comparable or better than five state-of-the-art techniques for the experiments considered.

117 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis.
Abstract: Machine Learning is the study of methods for programming computers to learn. Computers are applied to a wide range of tasks, and for most of these it is relatively easy for programmers to design and implement the necessary software. However, there are many tasks for which this is difficult or impossible. These can be divided into four general categories. First, there are problems for which there exist no human experts. For example, in modern automated manufacturing facilities, there is a need to predict machine failures before they occur by analyzing sensor readings. Because the machines are new, there are no human experts who can be interviewed by a programmer to provide the knowledge necessary to build a computer system. A machine learning system can study recorded data and subsequent machine failures and learn prediction rules. Second, there are problems where human experts exist, but where they are unable to explain their expertise. This is the case in many perceptual tasks, such as speech recognition, hand-writing recognition, and natural language understanding. Virtually all humans exhibit expert-level abilities on these tasks, but none of them can describe the detailed steps that they follow as they perform them. Fortunately, humans can provide machines with examples of the inputs and correct outputs for these tasks, so machine learning algorithms can learn to map the inputs to the outputs. Third, there are problems where phenomena are changing rapidly. In finance, for example, people would like to predict the future behavior of the stock market, of consumer purchases, or of exchange rates. These behaviors change frequently, so that even if a programmer could construct a good predictive computer program, it would need to be rewritten frequently. A learning program can relieve the programmer of this burden by constantly modifying and tuning a set of learned prediction rules. Fourth, there are applications that need to be customized for each computer user separately. Consider, for example, a program to filter unwanted electronic mail messages. Different users will need different filters. It is unreasonable to expect each user to program his or her own rules, and it is infeasible to provide every user with a software engineer to keep the rules up-to-date. A machine learning system can learn which mail messages the user rejects and maintain the filtering rules automatically. Machine learning addresses many of the same research questions as the fields of statistics, data mining, and psychology, but with differences of emphasis. Statistics focuses on understanding the phenomena that have generated the data, often with the goal of testing different hypotheses about those phenomena. Data mining seeks to find patterns in the data that are understandable by people. Psychological studies of human learning aspire to understand the mechanisms underlying the various learning behaviors exhibited by people (concept learning, skill acquisition, strategy change, etc.).

13,246 citations

Christopher M. Bishop1
01 Jan 2006
TL;DR: Probability distributions of linear models for regression and classification are given in this article, along with a discussion of combining models and combining models in the context of machine learning and classification.
Abstract: Probability Distributions.- Linear Models for Regression.- Linear Models for Classification.- Neural Networks.- Kernel Methods.- Sparse Kernel Machines.- Graphical Models.- Mixture Models and EM.- Approximate Inference.- Sampling Methods.- Continuous Latent Variables.- Sequential Data.- Combining Models.

10,141 citations

Journal ArticleDOI
01 Oct 1980

1,565 citations

Proceedings ArticleDOI
23 Aug 2015
TL;DR: A new Challenge 4 on Incidental Scene Text has been added to the Challenges on Born-Digital Images, Focused Scene Images and Video Text and tasks assessing End-to-End system performance have been introduced to all Challenges.
Abstract: Results of the ICDAR 2015 Robust Reading Competition are presented. A new Challenge 4 on Incidental Scene Text has been added to the Challenges on Born-Digital Images, Focused Scene Images and Video Text. Challenge 4 is run on a newly acquired dataset of 1,670 images evaluating Text Localisation, Word Recognition and End-to-End pipelines. In addition, the dataset for Challenge 3 on Video Text has been substantially updated with more video sequences and more accurate ground truth data. Finally, tasks assessing End-to-End system performance have been introduced to all Challenges. The competition took place in the first quarter of 2015, and received a total of 44 submissions. Only the tasks newly introduced in 2015 are reported on. The datasets, the ground truth specification and the evaluation protocols are presented together with the results and a brief summary of the participating methods.

1,224 citations

Proceedings ArticleDOI
25 Aug 2013
TL;DR: The datasets and ground truth specification are described, the performance evaluation protocols used are details, and the final results are presented along with a brief summary of the participating methods.
Abstract: This report presents the final results of the ICDAR 2013 Robust Reading Competition. The competition is structured in three Challenges addressing text extraction in different application domains, namely born-digital images, real scene images and real-scene videos. The Challenges are organised around specific tasks covering text localisation, text segmentation and word recognition. The competition took place in the first quarter of 2013, and received a total of 42 submissions over the different tasks offered. This report describes the datasets and ground truth specification, details the performance evaluation protocols used and presents the final results along with a brief summary of the participating methods.

1,191 citations