scispace - formally typeset
Search or ask a question
Author

Vijayarajan Rajangam

Bio: Vijayarajan Rajangam is an academic researcher from VIT University. The author has contributed to research in topics: Image fusion & Feature vector. The author has an hindex of 1, co-authored 6 publications receiving 3 citations.

Papers
More filters
Journal ArticleDOI
TL;DR: An attempt to recognize seven emotional states from speech signals, known as sad, angry, disgust, happy, surprise, pleasant, and neutral sentiment, is investigated, which employs a non-linear signal quantifying method based on randomness measure,known as the entropy feature, for the detection of emotions.
Abstract: Emotion recognition system from speech signal is a widely researched topic in the design of the Human–Computer Interface (HCI) models, since it provides insights into the mental states of human beings. Often, it is required to identify the emotional condition of the humans as cognitive feedback in the HCI. In this paper, an attempt to recognize seven emotional states from speech signals, known as sad, angry, disgust, happy, surprise, pleasant, and neutral sentiment, is investigated. The proposed method employs a non-linear signal quantifying method based on randomness measure, known as the entropy feature, for the detection of emotions. Initially, the speech signals are decomposed into Intrinsic Mode Function (IMF), where the IMF signals are divided into dominant frequency bands such as the high frequency, mid-frequency , and base frequency. The entropy measures are computed directly from the high-frequency band in the IMF domain. However, for the mid- and base-band frequencies, the IMFs are averaged and their entropy measures are computed. A feature vector is formed from the computed entropy measures incorporating the randomness feature for all the emotional signals. Then, the feature vector is used to train a few state-of-the-art classifiers, such as Linear Discriminant Analysis (LDA), Naive Bayes, K-Nearest Neighbor, Support Vector Machine, Random Forest, and Gradient Boosting Machine. A tenfold cross-validation, performed on a publicly available Toronto Emotional Speech dataset, illustrates that the LDA classifier presents a peak balanced accuracy of 93.3%, F1 score of 87.9%, and an area under the curve value of 0.995 in the recognition of emotions from speech signals of native English speakers.

37 citations

Journal ArticleDOI
TL;DR: In this paper, a method of concatenating spatial pyramid Zernike moments based shape features and Law's texture features is proposed to uniquely capture the macro and micro details of each facial expression.
Abstract: Facial expression recognition (FER) requires better descriptors to represent the face patterns as the facial region changes due to the movement of the face muscles during an expression In this paper, a method of concatenating spatial pyramid Zernike moments based shape features and Law’s texture features is proposed to uniquely capture the macro and micro details of each facial expression The proposed method employs multilayer perceptron and radial basis function feed forward artificial neural networks for recognizing the facial expressions The suitability of the features in recognizing the expressions is explored across the datasets independent of the subjects or persons The experiments conducted on JAFFE and KDEF datasets demonstrate that the concatenated feature vectors are capable of representing the facial expressions with better accuracy and least errors The radial basis function based classifier delivers a performance with an average recognition accuracy of 9586% and 8887% on the JAFFE and KDEF datasets respectively for subject dependent FER

22 citations

Journal ArticleDOI
TL;DR: In this paper, a CNN was used to extract features from the skin lesions and concatenated with traditional features like texture and colour features extracted from the lesion region of the input images.
Abstract: Skin cancer is one of the most deadly diseases around the world, wherein one of the three cancers is skin cancer Early detection of skin cancer is paramount for better treatment planning This paper investigates a Convolutional Neural Network (CNN), specifically, You Only Look Once (YOLO), to extract features from the skin lesions The features, obtained from the CNN, are concatenated with traditional features like texture and colour features extracted from the lesion region of the input images Later, the concatenated features are fed to a Fully Connected Network, which is trained with the specific ground truths to achieve higher classification accuracy The proposed method improves the detection and classification of skin lesions when compared with other models and YOLO without traditional features The performance measures of the fusion network are able to achieve the accuracy of 94%, precision of 085, recall of 088, and area under the curve of 095

9 citations

Journal ArticleDOI
TL;DR: In this article, an improved ECG-derived respiration (EDR) based on empirical wavelet transform (EWT) and kernel principal component analysis (KPCA) is proposed.
Abstract: Many methods have been developed to derive respiration signals from electrocardiograms (ECGs). However, traditional methods have two main issues: (1) focusing on certain specific morphological characteristics and (2) not considering the nonlinear relationship between ECGs and respiration. In this paper, an improved ECG-derived respiration (EDR) based on empirical wavelet transform (EWT) and kernel principal component analysis (KPCA) is proposed. To tackle the first problem, EWT is introduced to decompose the ECG signal to extract the low-frequency part. To tackle the second issue, KPCA and preimaging are introduced to capture the nonlinear relationship between ECGs and respiration. The parameter selection of the radial basis function kernel in KPCA is also improved, ensuring accuracy and a reduction in computational cost. The correlation coefficient and amplitude square coherence coefficient are used as metrics to carry out quantitative and qualitative comparisons with three traditional EDR algorithms. The results show that the proposed method performs better than the traditional EDR algorithms in obtaining single-lead-EDR signals.

4 citations

Book ChapterDOI
01 Jan 2021
TL;DR: This chapter analyzes the performance of a pre-trained VGG19 deep learning network that extracts features from the base and detail layers of the source images for constructing a weight map to fuse the source image details.
Abstract: Multimodal imaging systems assist medical practitioners in cost-effective diagnostic methods in clinical pathologies. Multimodal imaging of the same organ or the region of interest reveals complementing anatomical and functional details. Multimodal image fusion algorithms integrate complementary image details into a composite image that reduces clinician's time for effective diagnosis. Deep learning networks have their role in feature extraction for the fusion of multimodal images. This chapter analyzes the performance of a pre-trained VGG19 deep learning network that extracts features from the base and detail layers of the source images for constructing a weight map to fuse the source image details. Maximum and averaging fusion rules are adopted for base layer fusion. The performance of the fusion algorithm for multimodal medical image fusion is analyzed by peak signal to noise ratio, structural similarity index, fusion factor, and figure of merit. Performance analysis of the fusion algorithms is also carried out for the source images with the presence of impulse and Gaussian noise.

2 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: In this paper , a review of state-of-the-art machine and deep learning-based methods for emotion recognition has been presented, based on EEG, speech, facial expression, and multimodal features.

38 citations

Proceedings ArticleDOI
TL;DR: This paper proposed Contrastive Learning of Multi-view facial Expressions (CL-MEx) to exploit facial images captured simultaneously from different angles towards FER, which achieved state-of-the-art performance on two multi-view FER datasets.
Abstract: Facial expression recognition (FER) has emerged as an important component of human-computer interaction systems. Despite recent advancements in FER, performance often drops significantly for non-frontal facial images. We propose Contrastive Learning of Multi-view facial Expressions (CL-MEx) to exploit facial images captured simultaneously from different angles towards FER. CL-MEx is a two-step training framework. In the first step, an encoder network is pre-trained with the proposed self-supervised contrastive loss, where it learns to generate view-invariant embeddings for different views of a subject. The model is then fine-tuned with labeled data in a supervised setting. We demonstrate the performance of the proposed method on two multi-view FER datasets, KDEF and DDCF, where state-of-the-art performances are achieved. Further experiments show the robustness of our method in dealing with challenging angles and reduced amounts of labeled data.

26 citations

Proceedings ArticleDOI
18 Oct 2021
TL;DR: This paper proposed Contrastive Learning of Multi-view facial Expressions (CL-MEx) to exploit facial images captured simultaneously from different angles towards FER, which achieved state-of-the-art performance on two multi-view FER datasets.
Abstract: Facial expression recognition (FER) has emerged as an important component of human-computer interaction systems. Despite recent advancements in FER, performance often drops significantly for non-frontal facial images. We propose Contrastive Learning of Multi-view facial Expressions (CL-MEx) to exploit facial images captured simultaneously from different angles towards FER. CL-MEx is a two-step training framework. In the first step, an encoder network is pre-trained with the proposed self-supervised contrastive loss, where it learns to generate view-invariant embeddings for different views of a subject. The model is then fine-tuned with labeled data in a supervised setting. We demonstrate the performance of the proposed method on two multi-view FER datasets, KDEF and DDCF, where state-of-the-art performances are achieved. Further experiments show the robustness of our method in dealing with challenging angles and reduced amounts of labeled data.

21 citations

Journal ArticleDOI
TL;DR: A comprehensive review of remote patient monitoring (RPM) systems including adopted advanced technologies, AI impact on RPM, challenges and trends in AI-enabled RPM is presented in this article .
Abstract: The adoption of artificial intelligence (AI) in healthcare is growing rapidly. Remote patient monitoring (RPM) is one of the common healthcare applications that assist doctors to monitor patients with chronic or acute illness at remote locations, elderly people in‐home care, and even hospitalized patients. The reliability of manual patient monitoring systems depends on staff time management which is dependent on their workload. Conventional patient monitoring involves invasive approaches which require skin contact to monitor health status. This study aims to do a comprehensive review of RPM systems including adopted advanced technologies, AI impact on RPM, challenges and trends in AI‐enabled RPM. This review explores the benefits and challenges of patient‐centric RPM architectures enabled with Internet of Things wearable devices and sensors using the cloud, fog, edge, and blockchain technologies. The role of AI in RPM ranges from physical activity classification to chronic disease monitoring and vital signs monitoring in emergency settings. This review results show that AI‐enabled RPM architectures have transformed healthcare monitoring applications because of their ability to detect early deterioration in patients' health, personalize individual patient health parameter monitoring using federated learning, and learn human behavior patterns using techniques such as reinforcement learning. This review discusses the challenges and trends to adopt AI to RPM systems and implementation issues. The future directions of AI in RPM applications are analyzed based on the challenges and trends.

9 citations