scispace - formally typeset
T

Tobias Gehrig

Researcher at Karlsruhe Institute of Technology

Publications -  30
Citations -  753

Tobias Gehrig is an academic researcher from Karlsruhe Institute of Technology. The author has contributed to research in topics: Extended Kalman filter & Word error rate. The author has an hindex of 15, co-authored 30 publications receiving 712 citations.

Papers
More filters
Proceedings ArticleDOI

A joint particle filter for audio-visual speaker tracking

TL;DR: This paper uses features from multiple cameras and microphones, and process them in a joint particle filter framework, which performs sampled projections of 3D location hypotheses and scores them using features from both audio and video.
Proceedings ArticleDOI

Kalman filters for time delay of arrival-based source localization.

TL;DR: In this paper, a Kalman filter is employed to directly update the speaker's position estimate based on the observed time delay of arrival (TDOA) estimation, which consists of the observation associated with an extended Kalman Filter whose state corresponds to the speaker position.
Journal ArticleDOI

Kalman filters for time delay of arrival-based source localization

TL;DR: The proposed algorithm, although relying on an iterative optimization scheme, proved efficient enough for real-time operation and provides source localization accuracy superior to the standard spherical and linear intersection techniques.
Proceedings ArticleDOI

Kalman filters for audio-video source localization

TL;DR: This work proposes an algorithm to incorporate detected face positions in different camera views into the Kalman filter without doing any explicit triangulation, which yields a robust source localizer that functions reliably both for segments wherein the speaker is silent, which would be detrimental for an audio only tracker, and wherein many faces appear, which will confuse a video only tracker.
Proceedings Article

Multi-view facial expression recognition using local appearance features

TL;DR: This paper presents a multi-view facial expression classification system that utilizes local features extracted around automatically located facial landmarks using pose-dependent active appearance models, and evaluates the influence of AAM fitting errors, F-score feature selection, and expression intensity levels on classification accuracy.