Corpus-based dialogue systems rely on statistical models, whose parameters are inferred from annotated dialogues. The dialogues are usually annotated using Dialogue Acts (DA), and the manual annotation is difficult and time-consuming. Therefore, several semiautomatic annotation processes have been proposed to speed-up the process. The standard annotation model is based on Hidden Markov Models (HMM). In this work, we explore the impact of different types of HMM on annotation accuracy using these models on two dialogue corpora of dissimilar features. The results show that some types of models improve standard HMM in a human-computer task-oriented dialogue corpus, but their impact is lower in a human-human non-task-oriented dialogue corpus.
Carlos D. Martínez-Hinarejos, Vicent Tamari