Sciweavers

121 search results - page 6 / 25
» Multimodal Meeting Tracker
Sort
View
ICASSP
2009
IEEE
14 years 3 months ago
Multi-modal speaker diarization of real-world meetings using compressed-domain video features
Speaker diarization is originally defined as the task of determining “who spoke when” given an audio track and no other prior knowledge of any kind. The following article sho...
Gerald Friedland, Hayley Hung, Chuohao Yeo
ICPR
2008
IEEE
14 years 3 months ago
Clip retrieval using multi-modal biometrics in meeting archives
We present a system to retrieve all clips from a meeting archive that show a particular individual speaking, using a single face or voice sample as the query. The system incorpora...
Himanshu Vajaria, Sudeep Sarkar, Rangachar Kasturi
ICIP
2007
IEEE
13 years 8 months ago
Robust Multi-Modal Group Action Recognition in Meetings from Disturbed Videos with the Asynchronous Hidden Markov Model
The Asynchronous Hidden Markov Model (AHMM) models the joint likelihood of two observation sequences, even if the streams are not synchronised. We explain this concept and how the...
Marc Al-Hames, Claus Lenz, Stephan Reiter, Joachim...
TSD
2004
Springer
14 years 1 months ago
Multimodal Phoneme Recognition of Meeting Data
This paper describes experiments in automatic recognition of context-independent phoneme strings from meeting data using audiovisual features. Visual features are known to improve ...
Petr Motlícek, Jan Cernocký
ICMI
2005
Springer
215views Biometrics» more  ICMI 2005»
14 years 2 months ago
Multimodal multispeaker probabilistic tracking in meetings
Tracking speakers in multiparty conversations constitutes a fundamental task for automatic meeting analysis. In this paper, we present a probabilistic approach to jointly track th...
Daniel Gatica-Perez, Guillaume Lathoud, Jean-Marc ...