About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Publication
ICASSP 1992
Conference paper
Adaptation of large vocabulary recognition system parameters
Abstract
This paper reports on a series of experiments in which the Hidden Markov Model baseforms and the language model probabilities were updated from spontaneously dictated speech captured during recognition sessions with the IBM Tangora system. The basic technique for baseform modification consisted of constructing new fenonic baseforms for all recognized words. To modify the language model probabilities, a simplified version of a cache language model was implemented. The word error rate across six talkers was 3.7%. Baseform adaptation reduced the average error rate to 3.5%, and employing the cache language model reduced the error rate to 3.2%. Combining both techiques further reduced the error rate to 3.1% - a respectable improvement over the original error rate, especially given that the system was speaker-trained prior to adaptation.