Abstract. In this paper, we present initial results towards boosting posterior based speech recognition systems by estimating more informative posteriors using multiple streams of ...
The automatic processing of speech collected in conference style meetings has attracted considerable interest with several large scale projects devoted to this area. This paper des...
Thomas Hain, Lukas Burget, John Dines, Iain McCowa...
Abstract. In this paper we describe the 2005 AMI system for the transcription of speech in meetings used in the 2005 NIST RT evaluations. The system was designed for participation ...
Thomas Hain, Lukas Burget, John Dines, Giulia Gara...
Abstract. We aim to create a model of emotional reactive virtual humans. This model will help to define realistic behavior for virtual characters based on emotions and events in t...
This paper presents the design and results of the Rich Transcription Spring 2005 (RT-05S) Meeting Recognition Evaluation. This evaluation is the third in a series of community-wide...
Jonathan G. Fiscus, Nicolas Radde, John S. Garofol...
Abstract. An important step to bring speech technologies into wide deployment as a functional component in man-machine interfaces is to free the users from close-talk or desktop mi...
Stephen M. Chu, Etienne Marcheret, Gerasimos Potam...
Abstract. In this paper, we report on the infrastructure we have developed to support our research on multimodal cues for understanding meetings. With our focus on multimodality, w...
Lei Chen 0004, R. Rose, Ying Qiao, Irene Kimbara, ...
Abstract. The AMI Meeting Corpus is a multi-modal data set consisting of 100 hours of meeting recordings. It is being created in the context of a project that is developing meeting...
Jean Carletta, Simone Ashby, Sebastien Bourban, Mi...
Acoustic events produced in meeting-room-like environments may carry information useful for perceptually aware interfaces. In this paper, we focus on the problem of combining diffe...