S
Shashidhar G. Koolagudi
Researcher at National Institute of Technology, Karnataka
Publications - 162
Citations - 2559
Shashidhar G. Koolagudi is an academic researcher from National Institute of Technology, Karnataka. The author has contributed to research in topics: Mel-frequency cepstrum & Speech corpus. The author has an hindex of 21, co-authored 145 publications receiving 1910 citations. Previous affiliations of Shashidhar G. Koolagudi include Indian Institute of Technology Kharagpur & Graphic Era University.
Papers
More filters
Journal ArticleDOI
Emotion recognition from speech: a review
TL;DR: The recent literature on speech emotion recognition has been presented considering the issues related to emotional speech corpora, different types of speech features and models used for recognition of emotions from speech.
Proceedings ArticleDOI
Tomato Leaf Disease Detection Using Convolutional Neural Networks
Prajwala Tm,Alla Pranathi,Kandiraju SaiAshritha,Nagaratna B. Chittaragi,Shashidhar G. Koolagudi +4 more
TL;DR: This paper adopts a slight variation of the convolutional neural network model called LeNet to detect and identify diseases in tomato leaves and has achieved an average accuracy of 94–95 % indicating the feasibility of the neural network approach even under unfavourable conditions.
Journal ArticleDOI
Emotion recognition from speech using global and local prosodic features
TL;DR: The results indicate that, the recognition performance using local Prosodic features is better compared to the performance of global prosodic features.
Book ChapterDOI
IITKGP-SESC: Speech Database for Emotion Analysis
Shashidhar G. Koolagudi,Sudhamay Maity,Vuppala Anil Kumar,Saswat Chakrabarti,K. Sreenivasa Rao +4 more
TL;DR: The design, acquisition, post processing and evaluation of the proposed speech database (IITKGP-SESC), recorded in Telugu language using the professional artists from All India Radio, Vijayawada, India is described.
Journal ArticleDOI
Emotion recognition from speech using source, system, and prosodic features
TL;DR: From the results, it is observed that, each of the proposed speech features has contributed toward emotion recognition and the combination of features improved the emotion recognition performance, indicating the complementary nature of the features.