Computer Assisted Language Learning (CALL) applications for improving the oral skills of low-proficient learners have to cope with nonnative speech that is particularly challenging. Since unconstrained nonnative ASR is still problematic, a possible solution is to elicit constrained responses from the learners. In this paper we describe experiments aimed at selecting utterances from lists of responses. The first experiment on utterance selection indicates that the decoding process can be improved by optimizing the language model and the acoustic models, thus reducing the utterance error rate from 29-26% to 10-8%. Since giving feedback on incorrectly recognized utterances is confusing, we verify the correctness of the utterance before providing feedback. The results of the second experiment on utterance verification indicate that combining duration related features with a likelihood ratio (LR) yields an equal error rate (EER) of 10.3%, which is significantly better than the EER for the ...