scispace - formally typeset
Search or ask a question
Author

W. Fellenz

Bio: W. Fellenz is an academic researcher from King's College London. The author has contributed to research in topics: Time delay neural network & Content-addressable storage. The author has an hindex of 4, co-authored 7 publications receiving 2251 citations.

Papers
More filters
Journal ArticleDOI
TL;DR: Basic issues in signal processing and analysis techniques for consolidating psychological and linguistic analyses of emotion are examined, motivated by the PKYSTA project, which aims to develop a hybrid system capable of using information from faces and voices to recognize people's emotions.
Abstract: Two channels have been distinguished in human interaction: one transmits explicit messages, which may be about anything or nothing; the other transmits implicit messages about the speakers themselves. Both linguistics and technology have invested enormous efforts in understanding the first, explicit channel, but the second is not as well understood. Understanding the other party's emotions is one of the key tasks associated with the second, implicit channel. To tackle that task, signal processing and analysis techniques have to be developed, while, at the same time, consolidating psychological and linguistic analyses of emotion. This article examines basic issues in those areas. It is motivated by the PKYSTA project, in which we aim to develop a hybrid system capable of using information from faces and voices to recognize people's emotions.

2,255 citations

01 Jan 1999
TL;DR: An empirical approach to identifying the kind of task that an emotion recognition system could usefully address is described and the results confirm that an approach of this kind is feasible.
Abstract: We describe an empirical approach to identifying the kind of task that an emotion recognition system could usefully address. Three levels of information are elicited – a basic emotion vocabulary, a basic representation in ‘evaluation-activation space’ of the meaning of each word, and a richer ‘schema’ representation. The results confirm that an approach of this kind is feasible. Key-Words: emotion recognition, semantics, neural network CSCC'99 Proceedings, Pages:5311-5316

81 citations

Proceedings ArticleDOI
24 Jul 2000
TL;DR: A framework for the processing of face image sequences and speech, using different dynamic techniques to extract appropriate features for emotion recognition, using neural network techniques and fuzzy logic is proposed.
Abstract: We propose a framework for the processing of face image sequences and speech, using different dynamic techniques to extract appropriate features for emotion recognition. The features will be used by a hybrid classification procedure, employing neural network techniques and fuzzy logic, to accumulate the evidence for the presence of an emotional expression of the face and the speaker's voice.

38 citations

Proceedings ArticleDOI
24 Jul 2000
TL;DR: Three unsupervised learning algorithms are introduced and related numerical examples for a multilayer perceptron, recurrent neural networks, and a specially devised vector quantizer are shown.
Abstract: We split the rule extraction task into a subsymbolic and a symbolic phase and present a set of neural networks for filling the former. Under the two general commitments of: (i) having a learning algorithm that is sensitive to feedback signals coming from the latter phase, and (ii) extracting Boolean variables whose meaning is determined by the further symbolic processing, we introduce three unsupervised learning algorithms and show related numerical examples for a multilayer perceptron, recurrent neural networks, and a specially devised vector quantizer.

5 citations

Proceedings ArticleDOI
10 Jul 1999
TL;DR: It is shown that the hidden layer associative network is less dependent on the number of ones in the storage patterns and shows an improved retrieval capacity from incomplete retrieval patterns.
Abstract: We present a simple modification to binary associative memories employing a designated intermediate layer representation which reduces crosstalk between stored patterns and enhances performance and average storage capacity of the network. It is shown that the hidden layer associative network is less dependent on the number of ones in the storage patterns and shows an improved retrieval capacity from incomplete retrieval patterns.

4 citations


Cited by
More filters
Journal ArticleDOI

3,628 citations

Journal ArticleDOI
TL;DR: In this paper, the authors discuss human emotion perception from a psychological perspective, examine available approaches to solving the problem of machine understanding of human affective behavior, and discuss important issues like the collection and availability of training and test data.
Abstract: Automated analysis of human affective behavior has attracted increasing attention from researchers in psychology, computer science, linguistics, neuroscience, and related disciplines. However, the existing methods typically handle only deliberately displayed and exaggerated expressions of prototypical emotions despite the fact that deliberate behaviour differs in visual appearance, audio profile, and timing from spontaneously occurring behaviour. To address this problem, efforts to develop algorithms that can process naturally occurring human affective behaviour have recently emerged. Moreover, an increasing number of efforts are reported toward multimodal fusion for human affect analysis including audiovisual fusion, linguistic and paralinguistic fusion, and multi-cue visual fusion based on facial expressions, head movements, and body gestures. This paper introduces and surveys these recent advances. We first discuss human emotion perception from a psychological perspective. Next we examine available approaches to solving the problem of machine understanding of human affective behavior, and discuss important issues like the collection and availability of training and test data. We finally outline some of the scientific and engineering challenges to advancing human affect sensing technology.

2,503 citations

Journal ArticleDOI
05 Nov 2008
TL;DR: A new corpus named the “interactive emotional dyadic motion capture database” (IEMOCAP), collected by the Speech Analysis and Interpretation Laboratory at the University of Southern California (USC), which provides detailed information about their facial expressions and hand movements during scripted and spontaneous spoken communication scenarios.
Abstract: Since emotions are expressed through a combination of verbal and non-verbal channels, a joint analysis of speech and gestures is required to understand expressive human communication. To facilitate such investigations, this paper describes a new corpus named the “interactive emotional dyadic motion capture database” (IEMOCAP), collected by the Speech Analysis and Interpretation Laboratory (SAIL) at the University of Southern California (USC). This database was recorded from ten actors in dyadic sessions with markers on the face, head, and hands, which provide detailed information about their facial expressions and hand movements during scripted and spontaneous spoken communication scenarios. The actors performed selected emotional scripts and also improvised hypothetical scenarios designed to elicit specific types of emotions (happiness, anger, sadness, frustration and neutral state). The corpus contains approximately 12 h of data. The detailed motion capture information, the interactive setting to elicit authentic emotions, and the size of the database make this corpus a valuable addition to the existing databases in the community for the study and modeling of multimodal and expressive human communication.

2,359 citations

Journal ArticleDOI
TL;DR: Basic issues in signal processing and analysis techniques for consolidating psychological and linguistic analyses of emotion are examined, motivated by the PKYSTA project, which aims to develop a hybrid system capable of using information from faces and voices to recognize people's emotions.
Abstract: Two channels have been distinguished in human interaction: one transmits explicit messages, which may be about anything or nothing; the other transmits implicit messages about the speakers themselves. Both linguistics and technology have invested enormous efforts in understanding the first, explicit channel, but the second is not as well understood. Understanding the other party's emotions is one of the key tasks associated with the second, implicit channel. To tackle that task, signal processing and analysis techniques have to be developed, while, at the same time, consolidating psychological and linguistic analyses of emotion. This article examines basic issues in those areas. It is motivated by the PKYSTA project, in which we aim to develop a hybrid system capable of using information from faces and voices to recognize people's emotions.

2,255 citations

Journal ArticleDOI
TL;DR: A survey of speech emotion classification addressing three important aspects of the design of a speech emotion recognition system, the choice of suitable features for speech representation, and the proper preparation of an emotional speech database for evaluating system performance are addressed.

1,735 citations