A Dynamic Language Model for Speech Recognition

A Dynamic Language Model for Speech Recognition

A Dynamic Language Model for Speech Recognition F. Jelinek, B. Merialdo, S. Roukos, and M. Strauss I IBM Research Division, Thomas J. Watson Research Center, Yorktown Heights, NY 10598 ABSTRACT eral words are bursty by nature, i.e., one expects the In the case of a trlgr~m language model, the proba- word "language" to occur in this paper at a signif- bility of the next word conditioned on the previous two icantly higher rate than the average frequency esti- words is estimated from a large corpus of text. The re- mated from a large collection of text. To capture the sulting static trigram language model (STLM) has fixed "dynamic" nature of the trigram probabilities in a probabilities that are independent of the document being particular document, we present a "cache;' trigram dictated. To improve the language mode] (LM), one can language model (CTLM) that uses a window of the n adapt the probabilities of the trigram language model to match the current document more closely. The partially most recent words to determine the probability dis- dictated document provides significant clues about what tribution of the next word. words ~re more likely to be used next. Of many meth- ods that can be used to adapt the LM, we describe in this The idea of using a window of the recent history to paper a simple model based on the trigram frequencies es- adjust the LM probabilities was proposed in [2, 3]. In timated from the partially dictated document. We call this [2] tile dynamic component adjusted the conditional model ~ cache trigram language model (CTLM) since we probability, p,(w,+l ] g,+l), of the next word, wn+l, are c~chlng the recent history of words. We have found given a predicted part-oPspeech (POS), g,~+l, in a tri- that the CTLM red,aces the perplexity of a dictated doc- ument by 23%. The error rate of a 20,000-word isolated part-of-speech language model. Each POS had a sep- word recognizer decreases by about 5% at the beginning of arate cache where the frequencies of all the words that a document and by about 24% after a few hundred words. occurred with a POS determine the dynamic compo- nent of the language model. As a word is observed it is tagged and the appropriate POS cache is updated. At least 5 occurrences per cache are required before INTRODUCTION activating it. P~eliminary results for a couple POS caches indicate that with appropriate smoothing the A language model is used in speech recognition perplexity, for example, on NN-words is decreased by systems and automatic translation systems to improve a factor of 2.5 with a cache-based conditional word the performance of such systems. A trigram language probability given the POS category instead of a static model [1], whose parameters are estimated from a probability model. large corpus of text (greater than a few million words), has been used successfully in both applications. The In [3], the dynamic model uses two bigram lan- trigram language model has a probability distribu- guage models, p,(w,.+l ] w,, D), where D=I for words tion for the next word conditioned on the previous w,+l that have occurred in the cache window and two words. This static distribution is obtained a.s an D=0 for words that have occurred in the cache win- average over many documents. Yet we know that sev- dow. For cache sizes from 128 to 4096, the reported I Now at Rutgers University, N J, work performed while vis- res,lts indicate an improvement in the average rank iting IBM of tile correct word predicted by the model by 7% to 293 Test static dynalnic ~c where p,(...) is the usual static trigram language model. Set Perplexity Perplexity We use the forward-backward algorithm to estimate A 9i 75 0.07 the interpolation parameter ),c [1]. This parameter B 53 49 0.12 varies between 0.07 and 0.28 depending on the par- C 262 202 0.07 ticular static trigram language model (we used tri- gram language models estimated from different size Table 1: Perplexity of static and dynamic language mod- corpora) and the cache size (varying from 200 to 1000 els. words.) We have evaluated this cache language model by Perplexity 262 217 computing the perplexity on three test sets: Table 2: Perplexity ~s ~ function of cache size on test set C. • Test sets A and B are each about 100k words of text that were excised from a corpus of docu- Static Unigram Trigram ments from an insurance company that was used Cache Cache for building the static trigram language model 262 230 202 for a 20,000-word vocabulary. • Test set C which consists of 7 documents (about Table 3: Perplexity of unigram and trigram caches. 4000 words) that were dictated in a field trial in the same insurance company on TANGORA 17% over the static model assuming one knows if the (the 20,000-word isolated word recognizer devel- next word is in the cache or not. oped at IBM.) In this paper, we will present a new cache lan- Table ] shows the perplexity of the static and dy- guage model and compare its performance to a tri- namic language models for the three test sets. The gram language model. In Section 2, we present our cache size was 1000 words and was updated word proposed dynamic component and some results com- synchronously. The static language model was esti- paring static and dynamic trigram language models mated from about 1.2 million words of insurance doc- using perplexity. In Section 3, we present our method uments. The dynamic language model yields from 8% for incorporating the dynamic language model in an to 23% reduction in perplexity, with the larger reduc- isolated 20,000 word speech recognizer and its effect tion occurring with the test sets with larger perplex- on recognition performance. ity. The interpolation weight Ac was estimated using set B when testing on sets A and C and set. A when CACHE LANGUAGE MODEL testing on set B. Table 2 shows the effect of cache size Using a window of the n most recent words, we can on perplexity where it appears that a larger cache is estimate a unigram frequency distribution f,(w,+l), more useful. These results were on test set C. On a bigram frequency distribution, fn(w,÷l I w,), and a test set C, the rate that the next word is in the cache trigram frequency distribution, f,(w,+l } w,,w,-1). ranges from 75% for a cache window of 200 words to The resulting 3 dynamic estimators are linearly smoothed 83% for a window of 1000. Table 3 compares a cache together to obtain a dynamic trigram model denoted with unigrams only with a full trigram cache (for the by pc,(w,+l I w,., w,-a). The dynamic trigram model trigram cache, the weights for the unigram, bigram, assigns a non-zero probability for the words that have and trigram frequencies were 0.25, 0.25, 0.5 respec- occurred in the window of the previous n words. Since tively and were selected by hand.) A second set of the next word may not be in the cache and since the weights (0.25,0.5,0.25) produced a perplexity of 190 cache contains very few trigrams, we interpolate lin- for the trigram cache. In all the above experiments, early the dynamic model with the the static trigram the cache was not flushed between documents. In the language model: next section, we compare the different models in an I = (1) isolated speech recognition experiment. [ w,,, w._l) + We have tried using a fancier interpolation scheme (1 - I w,,.,w,,.-1) where the reliance on the cache depends on the cur- 294 Text 0-100 100-200 200-300 300-400 400-500 500-800 Length % Reduction in Error 6.1% 5.3% 4.7% 10.5% 16.3% 23.8% Rate Table 4: Percentage reduction in error rate with trigr~m cache. rent word wn with the expectation that some words each document. will tend to be followed by bursty words whereas other words will tend to be followed by non-bursty words. In these experiments, the weights for interpolating We typically used about 50 buckets (or weighting pa- the dynamic unigram, bigram, and trigram hequen- rameters). However, we have found that the perplex- cies were 0.4, 0.5, and 0.1, respectively. The weight of ity on independent data to be no better than the single the cache probability, At, relative to the static trigram parameter interpolation. probability was 0.2. Small changes in this weight does not seem to affect recognition performance. The po- tential benefit of a cache depends on the amount of ISOLATED SPEECH RECOGNITION text that has been observed. Table 4 shows the per- We incorporated the cache language model into centage reduction in error rate as a function of the the TANGORA isolated speech recognition system. length of the observed text. We divided the docu- We evaluated two cache update strategies. In the first ments into 100-word bins and computed the error rate one, the cache is updated at the end of every utter- in each bin. For the static language model, the error ance, i.e., when the speaker turns off the microphone. rate should be constant except for statistical fluctua- An utterance may be a partial sentence or a complete tions, whereas one expects that the error rate of the sentence or several sentences depending on how the cache to decrease with longer documents.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    3 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us