Pronunciation Learning for Automatic Speech Recognition by Ibrahim Badr Submitted to the Department of Electrical Engineering and Computer Science in partial fulfillment of the requirements for the degree of Master of Science in Electrical Engineering and Computer Science ARCHNES MASE at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY June 2011 @ Massachusetts Institute of Technology 2011. All rights reserved. . ............ A uthor ............. Department of Electrical Engineering and Computer Science May 18, 2011 Certified by................. .. . .. . .. . . . James Glass Principal Research Scientist Thesis Supervisor Accepted by ............. / j) eslie A. Kolodziejski Chair of the Committee on Graduate Students 2 Pronunciation Learning for Automatic Speech Recognition by Ibrahim Badr Submitted to the Department of Electrical Engineering and Computer Science on May 18, 2011, in partial fulfillment of the requirements for the degree of Master of Science in Electrical Engineering and Computer Science Abstract In many ways, the lexicon remains the Achilles heel of modern automatic speech recognizers (ASRs). Unlike stochastic acoustic and language models that learn the values of their parameters from training data, the baseform pronunciations of words in an ASR vocabulary are typically specified manually, and do not change, unless they are edited by an expert. Our work presents a novel generative framework that uses speech data to learn stochastic lexicons, thereby taking a step towards alleviating the need for manual intervention and automnatically learning high-quality baseform pronunciations for words. We test our model on a variety of domains: an isolated-word telephone speech corpus, a weather query corpus and an academic lecture corpus. We show significant improvements of 25%, 15% and 2% over expert-pronunciation lexicons, respectively. We also show that further improvements can be made by combining our pronunciation learning framework with acoustic model training. Thesis Supervisor: James Glass Title: Principal Research Scientist Acknowledgments I am grateful to my advisor Jim Glass for his guidance and support. His patience and encouragement has helped me significantly in my past two years at MIT. While I was still an undergraduate at the American University of Beirut, Jim accepted I intern at SLS as a visiting student, despite the fact that I had no significant speech recognition, machine learning or even programming experience to speak of. What convinced Jim to do so based only on a half hour telephone conversation, a recommendation from my algorithms professor and some patched up chess algorithm code I may never know. I will always be thankful for the opportunity, studying and working at MIT has been a humbling and life changing experience. Much thanks to Ian McGraw for his patience in teaching me all the skills needed to work at SLS, from how to restart the browser on CSAIL Debian and proper coding style to belief propagation and discriminative acoustic modeling. I am grateful for all the help I have received from all the people at SLS, the staff, as well as, the students. I would like to thank Scott Cyphers, Stephanie Seneff, Lee Hetherington, Chao Wang, Hung-An Chang, Paul Hsu, Alex Gruenstein, as well as, everyone else in the group, too many to name here, for all their help and support. I would also like to thank Michael Collins for letting me unofficially sit in on his NLP course while I was still a visiting student. What I learned during that semester has heavily influenced my research. Last but not least, I would like to thank my friends and family for their continued encouragement and support - "nervat bi3tizir inch ibnik il sa2it" - This research was sponsored by Nokia as part of a joint MIT-Nokia research collaboration. fi Contents 1 Introduction 1.1 15 M otivation . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . 16 1.1.1 The Lexicon: A Hand-Written ASR Component . .. . . . 16 1.1.2 The OOV problem and Dynamic Vocabulary Learning . . . 17 . 1.2 Contributions of this Thesis . . . . . . . . . . . . . . . . . . . . . . 19 1.3 Summary of Each Chapter . . . . . . . . . . . . . . . . . . . . . . . 19 2 Background 21 2.1 Pronunciation Modeling. 2.2 Letter-to-Sound Models.... . . . . . . . .. . . . . . . ........ . ... . . . . . . . 21 . . . . . . . . . . . . 22 2.2.1 The Graphoneme model . . . . . . . . . . . . . . . . . . . . . . . 22 2.2.2 Other Work on Letter-to-Sound Prediction . . . . . . . . . . . . . . 23 .. . . . 24 2.3 Using Acoustics to Learn Pronunciations ...... 2.4 Global Linear Models . . . . . . . . . . . . . . . . . 2.5 Summary .. . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . .. . . . . . . . . 3 Learning Pronunciations from Isolated-Word Spoken Examples 3.1 3.2 Graphone-guided Phonetic Recognition... . . . .... .. . . . . .... ... 3.1.1 Cascading Recognizers....... 3.1.2 Pronunciation Mixture Model . . . . . . ........... . . . . . . . Experimental Setup and Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. 3.2.1 Experimental Setup... 3.2.2 Results using Un-Weighted Pronunciations..... . . . .. 25 26 3.2.3 Results using Stochastic Lexicon . . . . . . . . . . . . . . . . . . . . 34 3.3 Training Pronunciations using Noisy Acoustic Examples . . . . . . . . . . . 35 3.4 Analysis of Learned Baseforms . . . . . . . . . . . . . . . . . . . . . . . . . . 37 3.5 Phone Versus Phoneme Pronunciations . . . . . . . . . . . . . . . . . . . . . 37 3.6 Sum m ary 40 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Discriminative Pronunciation Training for Isolated-Word Speech Recognition 5 41 4.1 Global Linear Models for Lexicon Learning . . . . . . . . . . . . . . . . . . . 42 4.2 Implementing the Algorithm as an FST . . . . . . . . . . . . . . . . . . . . . 44 4.3 Experimental Setup and Results . . . . . . . . . . . . . . . . . . . . . . . . . 46 4.4 Summary......... . . . . . . . . . 47 . . . . . . . . . . . . . . . . . . . Pronunciation Mixture Model for Continuous Speech Recognition 49 5.1 EM for Estimating Phoneme Pronunciations . . . . . . . . . . . . . . . . . . 50 5.1.1 . . .. 52 . . . . . . . . . . . . . . . 52 EM for Estimating Phone Pronunciations . . . . . . . ... . 5.2 Implementation Details...... . . . . . . . . 5.3 Experimental Setup......... . . . . . . . . 5.4 5.5 5.6 . . . . . . . . . . . . . . . 5.3.1 Experimental Procedure....... . . 5.3.2 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 55 56 Analysis of Learned Baseforms......... . . . . . . . . . . . . . . . .. 58 5.4.1 Non-Native Speaker Pronunciation....... . . . . . . . . . . .. 60 5.4.2 Slow vs. Fast Speech . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 5.4.3 Word Confusion in Weather Domain and Lecture Domain Lexicon . . 63 Acoustic Model and Lexicon Co-Training . . . . . . . . . . . . . . . . . . . . 66 5.5.1 Real-W orld Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 5.5.2 Baseline Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 5.5.3 PM M Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 Sum m ary . . . . . . . . . . . . . . . . . . 6 Summary and Future Work . . . . . . . . . . . . . . . 68 71 A Weather Domain Lexicon 73 10 List of Figures 1-1 Major Components of a Speech Recognition System: The Acoustic Model, Lexicon and the Language Model [1] 1-2 . . . . . . . . . . . . . . . . . . . . . . Diagram of Speech Recognition System behavior when recognizing an utterance containing an OOV words [10] 3-1 18 . . . . . . . . . . . . . . . . . . ... Diagram of Cascading Recognizers Model: Each utterance uses recognition . . . . . . . . . results of preceding utterances to decode..... . . . . . 3-2 16 29 Diagram of Pronunciation Mixture Model: Pronunciation weights are initialized with graphoneme letter-to-sound model. All utterances vote in parallel to assign weights for pronunciations (E-step), Votes are then summed and normalized (M-step). New weights are used to initialize the next iteration. 3-3 . 30 Word Error Rate (WER) as a function of the number of example utterances used to adapt the underlying lexicon. We show results for using the Cascaded Recognizers Model, the PMM, the Graphone model and the Expert Dictionary 33 3-4 Word Error Rate (WER) as a function of the number of example utterances used to adapt the underlying lexicon. We show results for using the un- weighted learned pronunciations while varying the T threshold, as well as, using the PMM (stochastic lexicon) . . . . . . . . . . . . . . . . . . . . ... 5-1 Constrained Graphoneme Recognizer n-best hypothesis for "houston texas" along with total, AM and LM score . . . . . . . . . . . . . . . . . . . . . . . 5-2 35 54 Plot of H(w~b) and H(blw) as a function of EM iterations for both the Weather and Lecture dom ain. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65 12 List of Tables 3.1 By varying a threshold T over the weights learned in the PMM, we can incorporate multiple baseform pronunciations for individual words... 3.2 33 PMM results incorporating spoken examples collected via Amazon Mechanical Turk................. 3.3 . . ... . ..... .... . . . . . . . . . . . 37 Example baseform changes between expert dictionary and top PMM hypothesis. Phonemes involved in the difference have been capitalized... . ... 38 3.4 Top 10 phoneme confusion pairs for Phonebook lexicon . . . . . . . . . . . . 38 3.5 Top 5 Deleted Phonemes in Phonebook lexicon.... . . . . . . . . . . 38 3.6 Top 5 Inserted Phonemes in Phonebook lexicon . . . . . . . . . . . . . 39 3.7 PMM results for learning phone pronunciations from PhoneBook corpus. 4.1 Results for Discriminative training in WEB. on PhoneBook corpus . . . . . . 5.1 Results in WER on both the Weather and Lecture corpora using the PMM, . 39 46 the expert dictionary and the Graphoneme model (%) . . . . . . . . . . . . . 57 5.2 Lexicon statistics for the weather and lecture domains. 58 5.3 Top 10 confusion pairs for Weather Domain lexicon . . . . . . . . . . . ... 59 5.4 Top 5 Deleted Phonemes for Weather Corpus lexicon . . . . . . . . . . . . . 59 5.5 Top 5 Inserted Phonemes for Weather Corpus lexicon . . . . . . . . . . . . . 60 5.6 Example phoneme deletions between expert dictionary and top PMM hypoth- . . . . . . . . . . . . esis for W eather Corpus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7 Example of phoneme insertions between expert dictionary and top PMM hypothesis for Weather Corpus. 5.8 60 . . . . . . . . . . . . . . . . . . . . . . . . . . 61 . . . . 61 Top 5 phoneme confusion pairs for Indian accented non-native talker 5.9 Example of differences between native and Indian accented non-native speaker pronunciations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 5.10 Top 5 phoneme confusion pairs for Dutch accented non-native talker . . . . . 62 5.11 Example of differences between native and Dutch accented non-native speaker pronunciations............. . . . . . . . . . . . . . . . . . . . . . . 5.12 Top 10 phoneme confusion pairs between "fast" speech and "slow" speech 62 . . 64 5.13 Top 5 inserted phonemes in "fast" speech pronunciations . . . . . . . . . . . 64 5.14 Top 5 deleted phonemes in "fast" speech pronunciations 64 . . . . . . . . . . . 5.15 Examples of pronunciation differences between fast and slow speech pronunciations........ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 5.16 WER in percentage for E-book domain by iteratively training the acoustic models and the lexicon. A gradual reduction in WER can be observed. . . . 68 5.17 WER in percentage for E-book domain using pmmlem3. An increase in WER is observed possibly due to overfitting. . . . . . . . . . . . . . . . . . . . 68 Chapter 1 Introduction In many ways, the lexicon remains the Achilles heel of modern automatic speech recognizers (ASRs). Unlike stochastic acoustic and language models that learn the values of their parameters from training data, the baseform pronunciations of words in an ASR vocabulary are typically specified manually (usually along with the basic phoneme inventory itself), and do not change, unless they are edited by an expert. The lexicon is usually a long list of these hand-crafted pronunciations in the form of dictionary entries that map a word to one or more canonical pronunciations. A more desirable solution would be one whereby the basic linguistic units of a language, and the associated lexical pronunciations could be determined automatically from a large amount of speech data. In this thesis, we discuss methods for the latter, learning the lexical pronunciations of words given both their spelling and spoken utterances of isolated-word and/or continuous speech. This data-driven approach to lexicon generation might discard the notion of canonicalization altogether, and instead generate a stochastic lexicon with pronunciations weighted according to learned statistics. Like the acoustic and language models, ideally pronunciations would be learned from data closely matching the test domain. In our work, we use the same training data used for Acoustic Model training to learn our pronunciations. Since this data is readily available at no extra cost for all recognizers, we see our work as encouraging lexicon training to become a standard procedure when building an ASR system, while taking a step towards being able to train a speech recognizer entirely from an orthographically transcribed corpus. 11 .......... 111............ ............. 1111 .... ................ .......... ._ __I .. ... .... ..... I.... .... -----_ . ... ...... _ __ = _ 4. Figure 1-1; Major Components of a Speech Recognition System: The Acoustic Model, Lexicon and the Language Model [1] 1.1 1.1.1 Motivation The Lexicon: A Hand-Written ASR Component ASR is the process of decoding a spoken utterance into a string of words [10]. The spoken utterance is usually modeled as a weighted network of sub-lexical units, which are typically phones. A phone graph which models all speech corresponding to an input acoustic signal is constrained by the acoustic model, the lexicon and the language model. The acoustic model statistically models context-dependent or context-independent phones, typically Gaussian Mixture models are used. It is trained on acoustic-phonetic measurements of transcribed speech data. The lexicon maps words to their phonetic pronunciations, an example entry would be colonel : k er n ax 1, and is typically crafted manually by experts. The language model statistically models the probability of a word sequence and typically uses n-gram distributions to scores sentences. Figure 1-1 illustrates this process. The fundamental equation of ASR seeks the most likely word sequence W* = w*, ... , w* given an utterance u: W =argmaxP(Wu) = argmaxP(ulW)P(W) W W (1.1) Where P(ulW) is the acoustic model and P(W) is the language model. This equation implicitly implies that to each word corresponds one phonetic representation. A more accurate model would be: P(Wlu) = ( PW,B = Zu) P(u) B,u) - ZBCBP(W)P(BIW)P(uIB, P(u) (1.2) Here B is a phonetic pronunciation and B is the set of all phonetic pronunciations. P(Wlu) - ZBc P(W)P(B|W)P(u|B) P(u) (1.3) where we assume conditional independence P(ulB, W) = P(ulB). Hence: W arg max P(Wlu) = arg max w w P(u|B)P(B|W)P(W) (1.4) BEB P(B IW) is the lexicon. It assigns a probability to each phonetic pronunciation. Typically all baseforms are weighed equally by setting P(BlW) = 1 for all B corresponding to W in the lexicon, even if the lexicon admits multiple pronunciations per word [34, 25]. In this work we propose to learn the lexicon, thereby avoiding the use of experts, as well as having to weight all baseforms equally. Learning the lexicon from data opens up the possibility of using discriminative training in order to directly optimize on error rate. Discriminative training has been extensively studied for acoustic and language models and has been shown to perform significantly better than maximum likelihood methods. In this thesis, we explore incorporating Discriminative training to enhance our stochastic lexicons. 1.1.2 The OOV problem and Dynamic Vocabulary Learning Almost all speech recognizers have a finite vocabulary pronunciation dictionary. If the recognizer encounters an out of vocabulary (OOV) term, a word that is not in the lexicon, it cannot produce a transcription for it. This limitation is detrimental in many ASR. applications. Figure 1-2 demonstrates a potential outcome to an OOV word euthanasia. A Acoustic P Feature Extraction Speech Example Input: euthanasia euthanasia f Language Kecognizer Search Space Word Hypotheses Example Outputs: youth in Asia using Asia use in Asia Lexicon Figure 1-2: Diagram of Speech Recognition System behavior when recognizing an utterance containing an OOV words [10] misrecognized OOV word can easily cause the surrounding words or even the whole sentence to be misrecognized. Therefore, in order for a ASR systems to be used in the real world, techniques need to be developed to handle the OOV problem. A typical approach is to increase the vocabulary size whether manually or automatically. This cannot completely eliminate OOVs in an open domain recognition task. Other techniques include using confidence scoring to detect OOVs [14], as well as, filler models to hypothesize the pronunciation of an OOV word [6]. In our work, we focus on automatically learning quality baseforms for OOV words in order to expand our lexicon. The rate of OOV occurrence is tied to the design of the ASR vocabulary. Since constantly increasing the vocabulary size is bound to introduce acoustic ambiguity, ASRs should undergo a paradigm shift from vocabulary design to that of an adaptive system that can dynamically detect and learn new words. In a typical scenario, a speech recognizer remains static while it is deployed. Its acoustic models, language models and lexicon are trained or specified beforehand and remain unchanged while the recognizer is used. In our work, we explore techniques to dynamically expand a recognizer's lexicon, as well as, improve on the pronunciations already present in the lexicon. This will be done by either automatically collecting new data to be used to retrain the lexicon or using the speech data collected from our speech recognizer's tasks. 1.2 Contributions of this Thesis This thesis offers the following specific contributions: " A method for learning better-than-expert word pronunciations given their spellings, as well as, spoken utterances. - This method can learn from both isolated-word and continuous speech. - This method is robust and hence can learn from cheaply collect noisy speech. - This method can supplement the Acoustic and Language models without the use of any additional training data. * A method for discriminatively training lexicon pronunciation weights for isolated word recognition. " A method for iteratively training the Acoustic Models, as well as, the Lexicon. 1.3 Summary of Each Chapter The rest of this thesis is organized into 4 chapters. The content of each of the remaining chapters is summarized next: " Chapter 2: Background Chapter 2 starts with a description of how pronunciation variability is handled in current ASR systems. It then presents a review of the major advances in pronunciation learning for ASR. It also explains some of the mathematical frameworks applied in this thesis. " Chapter 3: Learning Pronunciations from Isolated-Word Spoken Examples This chapter presents two Bayesian models that learn pronunciations from isolated word examples. It reports on experiments that produce better-than-expert pronunciations for isolated word recognition, as well as, an inherent robustness to real world noisy data. " Chapter 4: Discriminative Pronunciation Training for Isolated-Word Speech Recognition In this chapter, we improve on our results for isolated-word recognition by discriminatevly training our learned pronunciations to minimize WER. " Chapter 5: Learning Pronunciations from Continuous Speech Here, we explore the use of continuous speech data to learn stochastic lexicons. Building on the work of the previous chapter, we extend our framework to two domains containing spontaneous speech: a weather information retrieval spoken dialogue system and an academic lectures domain. We find that our learned lexicons out-perform expert, hand-crafted lexicons in each domain. We also explore the relationship between the Lexicon and the Acoustic Models by iteratively re-training them. We show improvemerits on a Book Titles corpus and discuss limitations that should be addressed for these scenarios. " Chapter 6: Summary and Future Direction This chapter concludes the thesis, and suggests some ideas for future research. Chapter 2 Background 2.1 Pronunciation Modeling Pronunciation variation has been identified as a major cause of errors for a variety of ASR tasks [26]. Pronunciations are typically modeled in a speech recognizer by a phonemic dictionary which may be accompanied by a set of rewrite rules to account for phonological variation. The ASR system used in this paper incorporates manually crafted phonological rules that account for segmental mismatches between the underlying phonemic baseforms and surfacelevel phonetic units. These rules have been shown to outperform relying on context-dependent acoustic models to implicitly model phonetic variation [18]. In this work, we model the ASR's search space using a weighted finite-state transducer (FST) [20]. The FST search space has four primary hierarchical components: the language model (G), the phoneme lexicon (L), the phonological rules (P) that expand the phoneme pronunciations to their phone variations, and the mapping from phone sequences to contextdependent model labels (C). The full network can be represented as a composition of these components: N =CoP oLoG (2.1) In this work, we experiment with learning context-independent phoneme pronunciations along with their weights. That is, we try to replace the manually crafted FST L while keeping the pronunciations rules FST P unchanged. We also explore learning phone pronunciations directly, thus avoiding the use of the phonological rewrite rules altogether. 2.2 Letter-to-Sound Models Generating pronunciations for new words has been the subject of much work [9, 7, 22, 33, 28, 5, 23, 21] . The common approach is to use a form of letter-to-sound model to predict the pronunciations of new words using only their spelling. Despite the fact that for some languages, mapping a spelling to a pronunciation is relatively straightforward, English has shown itself to be rather challenging. Many models have been used for this task: in our work, we compare our results to the graphone model which has been shown to produce state of the art scores for letter-to-sound tasks [7, 36]. We address the mathematical formulation of the graphone model in this next section. 2.2.1 The Graphoneme model We utilize the joint-multigram approach employed in [9, 7] to model the relationship between graphemes and phonetic units. In this work, we use the term graphone to denote a model that maps graphemes to phones, and graphoneme to refer to a model that maps graphemes to phonemes. We begin by constructing an n-gram model over graphoneme sequences. We let w denote a grapheme sequence drawn from the set of all possible grapheme sequences W and b denote a phoneme sequence drawn from the set of all possible phoneme sequences, B. A joint model of the letter-to-sound task can be formalized as: b* arg max P(w, b) bEB Generally speaking, a graphoneme, g = (w, b) (2.2) gC C W x B, is a sub-word unit that maps a grapheme subsequence, w, to a phoneme subsequence, b. By analogy, a graphone is an alternative sub-word unit that maps a grapheme subsequence to a phone subsequence. In this work, we restrict our attention to singular graphones or graphonemes, in which a mapping is made between at most one grapheme and at most one phonetic unit. The empty subsequence c is allowed, however a mapping from c to e is omitted. Taken together, a sequence of graphonemes, g, inherently specifies a unique sequence of graphemes w and phonemes b; however, there may be multiple ways to align the pair (w, b) into various graphoneme sequences g E S(w, b). The following table shows two possible graphoneme segmentations of the word "couple". w = c o b = k ah - k u1 ah k1 U/ a h1U/eC 1 p p ax p ax P/ p C e 1 11/ 11 gi C/ 92 c/k o/ c u/ahp/p c/ax 1/1 / Given this ambiguity, employing graphonemes in our joint model requires us to marginalize over all possible segmentations. Fortunately, the standard Viterbi approximation has been shown to incur only minor degradation in performance [71. P(w,b) = P(g) ~ max P(g) gGS(wvb) (2.3) gES(wb) In our work, we use the open source implementation provided by [7], which runs the Expectation-Maximization (EM) algorithm on a training corpus of word-phoneme pironunciation pairs to automatically infer graphoneme alignments. We then train a standard 5-gram language model over the automatically segmented corpus of graphonemes. This configuration has been shown to produce good results for singular graphonemes [36]. 2.2.2 Other Work on Letter-to-Sound Prediction While original work in Grapheme-to-Phoneme conversion consisted of rule-based methods [22], these quickly became overly complicated and tedious for practical use and were soon replaced by data-driven methods. An example of work that utilizes knowledge-based formal linguistic methods in a statistical framework is [33]. As a first step, a hand-written grammar parses words into a set of hand-written linguistically motivated sub-word "spellneme" units. Then, after parsing a large lexicon into these segmented words, a statistical n-gram model is trained and used later for decoding. Most other work on letter-to-sound generation can be classified as using a pronunciation by analogy technique or local classification technique. Local classification processes a word spelling sequentially from left-to-right. For each input character a decision is made by looking at the letter's context using decision trees [30] or neural networks [17]. As for pronunciation by analogy, the main theme is to scan the training lexicon for words or part-of-words that are in some sense similar to the word to be transcribed [28, 5]. The output pronunciation is then chosen to be analogous to the retrieved examples. While the most recent state-of-the-art work has been on probabilistic approaches such as the joint-sequence model discussed in the previous section. Some have also explored discriminative training in a joint-sequence setting [21], as well as, grapheme-to-phoneme conversion using a Statistical Machine Translation system (SMT) [23]. 2.3 Using Acoustics to Learn Pronunciations Other relevant work extends on the idea of letter-to-sound generation by incorporating spoken examples to refine the generated pronunciations [27, 11]. The work of [4] deduces the baseform b* given a, word or grapheme sequence w and an utterance u of the spoken word w. In this work, a decision tree is used to model P(w, b) which was later shown to produce poorer results when compared to graphone models. The work of [24] adapts the graphone model parameters using acoustic data and is applied to a name recognition task. Given a training set (Ui, wi) of pairs of words w and spoken examples w, they maximize the likelihood of their data: Al M log po(ui, w; 0) = i=1 log i=1 po(ui, w, b) bc3 by using the standard EM algorithm to adjust 0, the graphone n-grain parameters. They initially set the parameters of 0 to those of the graphone model and iterate until convergence. They also experiment with discriminative training and show that it produces better results than MLE. In our work, we use a similar framework to MLE geared towards predicting a pronunciation for a single word given multiple example utterances. The work of [35] uses the forced alignment of a phonetic decoder to generate a list of possible pronunciations for words, and then assigns weights using a Minimum-Classification-Error criterion. They then test on a business name query corpus. Curiously, work on pronunciation learning using acoustic examples is rarely applied across the entire lexicon to regularize the pronunciations with respect to the underlying acoustic models. By contrast, in our work we learn pronunciations across all lexical entries rather than the few out-of-vocabulary words for which we do not have an expert opinion. Thus, our test experiments directly compare a learned stochastic lexicon with manually-crafted pronunciations. 2.4 Global Linear Models Here we describe a general framework of linear models that can be applied to a diverse range of tasks, e.g. parsing or ASR hypothesis reranking. We will be modifying this framework to train our pronunciation lexicon. The framework is outlined in [12]. We learn a mapping from inputs x E X to outputs y c Y by assuming: " A set of training examples (xi, Ii) for i = 1 . N " A function GEN which enumerates at set of candidates GEN(x) for input x " A representation <b mapping (x, y) to a feature vector <D(x, y) E Rd " A parameter vector a E Rd We define a mapping from an input x to an output F(x) F(x) = arg max yEGEN(x) <D(x, y) -a (2.4) To illustrate how Global Linear Models can be applied to discriminative language model training for ASR, we reference the work of [32]. " Each (Xi, yi) is an utterance and its reference transcription " GEN(x) is the outputed word-lattice from a baseline recognizer e <(x, y) tracks unigram, bigram and trigram word counts in y, as well as, the score for (x, y) under the baseline recognizer To optimize the parameters a, we consider this variant of the perceptron algorithm: INPUT: Training examples (xi, yi) INITIALIZATION: Set a = 0 ALGORITHM: for t = 1. T,i = 1 --. N do zi =arg maxzEGEN(xj) (xi,Z) ' a if zi 4 y, then a =a + ID(Xi, yi) - T(Xi, Zi) end if end for OUTPUT: a For a complete discussion regarding the convergence of this algorithm, as well as, generalization to unseen data, we refer the reader to [12]. 2.5 Summary In this chapter, we presented a survey of approaches that handle pronunciation prediction from word spellings, as well as, word spelling supplemented with acoustic examples. We also explained how pronunciation variability is handled in the ASR system used in this thesis. We concluded with a description of Global Linear Models, a mathematical framework used for Discriminative training. Chapter 3 Learning Pronunciations from Isolated-Word Spoken Examples We first tackle the simplified task of learning word pronunciations from isolated-word speech and testing on isolated-word speech. Our research differs from previous work in the stochastic learning framework. In our case, we use an n-gram graphoneme-based model as the basis for our initial estimate of a pronunciation. The graphoneme model is used as a form of prior to condition our expectation of possible pronunciations of a new word, given its spelling. We then use spoken examples to further refine the pronunciations, and explore two different stochastic pronunciation models: the first cascades all the examples to find a single best pronunciation, while the second creates a pronunciation mixture model (PMM) to consider multiple pronunciations. We compare both approaches on the telephone-based PhoneBook corpus of isolated words [2] and find that they are able to recover expert-level pronunciation baseforms with relatively few example utterances. 3.1 Graphone-guided Phonetic Recognition We begin by exploring a model which incorporates one example utterance with the graphoneme model to find a single high probability baseform. Given a word with grapheme sequence w and an example utterance of the word u, we deduce the baseform for b* using a similar framework to that described in [4]. b* =argmaxP(blw,u) =argmaxP(b,w)p(ulb,w) beB3 bcB (3.1) We replace the decision tree originally described in [4] with a graphoneme N-gram model using independence assumptions and the Viterbi approximation. b arg max beIB7eS(w b) p(ub)P(g) (3.2) For each word, w, a recognizer, R., can be constructed using weighted finite state transducers (FSTs) to model the mapping of acoustic model labels to phoneme sequences, weighted by graphoneme n-gram parameters. The procedure described above only incorporates a single example utterance into the pronunciation generation framework. The following sections introduce two methods of utilizing a set of Al example utterances, u, , of a given word, w. 3.1.1 Cascading Recognizers As in Equation 3.1, we apply Bayes rule with the additional assumption of independence between example utterances given their pronunciations to model the probability of a baseform given the data: M argmaxP(blw,u') =argmaxP(b,w)fjp(ujlb,w) bE3 bE3 i=1 This multiple utterance recognizer can be implemented as a cascade of single utterance recognizers. Figure 3-1 illustrates this approach. While this framework ties up all utterances to predict a single baseform, in this next section, we explore a more parallel approach that can model multiple pronunciations simultaneously. Grahone ASA U2 00 + -E Un+4- : b Figure 3-1: Diagram of Cascading Recognizers Model: Each utterance uses recognition results of preceding utterances to decode. 3.1.2 Pronunciation Mixture Model A second formulation of pronunciation generation informed by multiple example utterances is that of a pronunciationmixture model (PMM). We parametrize our model with Ob,= P(b, w) under the assumption that a particular word w and baseform b have some joint probability, however small, of mapping to one another. In a setup similar to the work described in [24], expectation maximization is used to update these parameters based on the data (ul , w). Whereas Li et al. optimize graphoneme language model parameters, our goal here is to directly learn weights for word pronunciations, hence the PMM characterization. We begin by characterizing the log-likelihood of the data. M L(O) (log i=1 Al p(ui,w; 0) =(log i=1 Owb .p(u 1 w, b) bE B The parameters, 0, are initialized to our graphoneme n-gram model scores and run multiple iterations of the EM algorithm. The following equations specify the expectation and maximization steps respectively: -- -- -- - - - -- - ----- - Initial Parameters nrapho N-gramn I..------ ---.- - ---..-- - Figure 3-2: Diagram of Pronunciation Mixture Model: Pronunciation weights are initialized with graphoneme letter-to-sound model. All utterances vote in parallel to assign weights for pronunciations (E-step), Votes are then summed and normalized (M-step). New weights are used to initialize the next iteration. E-step: P(bluiw;0) w,b - P(uiIb, w) E ,,,p -p(uib,w) 1A M-step: 6 *b = P(blui, w; 0) i=1 We illustrate this parallel approach in Figure 3-2. As an initial experiment we pick the baseform b with the highest weight as the pronunciation of w and use it unweighted i.e. with a probability of 1. b arg max P(w, b; 0*) = arg max 0* bE13 bE13 (3.3) In later experiments, we use multiple pronunciations weighted by their mixture probabilities in a stochastic lexicon. 3.2 3.2.1 Experimental Setup and Results Experimental Setup To experiment with the two pronunciation models, we use a landmark-based speech recognizer [15]. MFCC averages are computed over varying durations around hypothesized acoustic-phonetic landmarks to generate 112-dimensional feature vectors, which are then whitened via a PCA rotation. The first 50 principal components are kept as the feature space over which diphone acoustic models are built. Each model is a diagonal Gaussian mixture with tip to 75 mixture components trained on a separate corpus of telephone speech. The search space in the recognizer is modeled using a flexible weighted FST toolkit [20]. We consider the task of isolated word recognition using the NYNEX PhoneBook corpus. PhoneBook is a phonetically-rich, isolated-word, telephone-speech database of American English spanning a variety of talkers and telephone transmission characteristics. The core section of PhoneBook consists of a total of 93,667 isolated-word utterances, totalling 23 hours of speech. This breaks down to 7,979 distinct words, each said by an average of 11.7 talkers, with 1,358 talkers each saying up to 75 words. All data were collected in 8-bit in-law digital form directly from a Ti telephone line. Talkers were adult native speakers of American English chosen to be demographically representative of the U.S [2]. To insure adequate data for our baseline experiments, we chose 2,000 words at random from the subset of the corpus that had example utterances from at least 13 distinct speakers. We held out two of the 13 utterances, one from a male speaker the other from a female speaker, to generate a 4,000 utterance test set. While the individual recognition experiments described in the next section are limited to the 2,000 selected words, a far larger lexicon was used to train the initial graphone language model parameters. For this work we used an internal dictionary that contains over 150,000 manually generated entries. To simulate the out-of-vocabulary scenario for which graphones are typically employed, we removed the 2,000 trial words from our lexicon, and further pruned similarly spelled words using a simple edit distance criterion. We then trained a 5-gram graphone language model according to the procedures described in [36]. We conduct two baseline experiments to frame our remaining results. The first is a graphoneme-only baseline in which we performed isolated word recognition over the 4,000 test utterances using a 2,000 word lexicon generated from the graphoneme model alone according to Equation 2.3. Since no acoustic information was used, this provides us with an upper-bound on word error rate (WER) of 16.7%. The second baseline was again the 2,000 word-recognition task; however, this time we explicitly used the manually generated pronunciations originally found in our lexicon [3], giving us a target WER of 12.4%, achievable directly by experts. It should be noted that about 160 words in the expert-lexicon had multiple baseforms associated with them. For example, the word "youths" was represented as both y uw dh z and y uw th s. Initial experiments indicated that allowing multiple baseforms could give an advantage to the expert-lexicon that could be leveraged in the other frameworks. We begin however by choosing only a single pronunciation for inclusion in an automatically generated lexicon. Even so, were able to show the feasibility of recovering and even surpassing the performance of manually generated baseforms. 3.2.2 Results using Un-Weighted Pronunciations Having established our baseline experiments, we evaluated both the cascading recognizer approach and the PMM by varying the number of training utterances for each, and evaluating the WER of the test set against the lexicons produced under each condition. The resulting plot is shown in Figure 3-3. It is encouraging to note that both models perform admirably, achieving expert-level pronunciations with just three example utterances per word. The cascading recognizer approach of Section 3.1.1 improves slightly faster than the PMM technique. With seven utterances, this model surpasses the expert baseform WER by nearly 1%. An inherent assumption of this model is that there is a single, correct underlying pronunciation. This fact may explain the slight advantage that this approach has, since our experimental design only allows a single baseform for each word in our automatically generated lexicon. A model which directly computes the single most likely baseform given the data is thus particularly well-suited to the task. Ideally, a pronunciation generation model would be able to cope with words that have Lexicon Adaptation using Acoustic Examples I7 '4* Graphones Only 016 C) 15 -X- Cascaded Recognizers 0, Pronunciation Mixture Model (D14 0 - 13 Expert Baseforms Cc W12 11 0 2 4 6 8 10 Number of utterances used to learn baseforms (M) Figure 3-3: Word Error Rate (WER) as a function of the number of example utterances used to adapt the underlying lexicon. We show results for using the Cascaded Recognizers Model, the PMM, the Graphone model and the Expert Dictionary r Avg. # b per w WER (%) 0.2 1.25 11.2 0.1 2.10 11.0 0.05 2.21 11.5 Table 3.1: By varying a threshold r over the weights learned in the PMM, we can incorporate multiple baseform pronunciations for individual words. multiple pronunciations, such as "either". It probably does not make sense, for example, to be multiplying the acoustic scores of one utterance pronounced iy dh er with a second pronounced ay dh er. Lastly, another potential pitfall of the cascaded recognizers approach is that unless special care is taken to customize the pruning procedure, acoustic variation will inherently cause the pronunciation search space to become successively smaller as the compositions prune lowprobability paths. This is especially problematic when considering noisy utterances. Indeed, even with the clean speech comprising the PhoneBook corpus, by the 11th utterance, N-best lists produced by the cascaded recognizers contained an average of just 10.7 entries. To illustrate the performance of the PMM, we show in Figure 3-3 the WER obtained by generating a lexicon according to Equation 3.3 (i.e. taking the highest scoring pronunciation and using it unweighted in the lexicon) after two iterations of EM. This stopping criterion was determined by constructing a development set of 1,500 previously discarded PhoneBook utterances and running recognition using lexicons generated after each EM iteration. Alternatively, EM could have been run to convergence and then smoothed, again with the aid of a development set. While the PMM requires slightly more data to achieve the lowest reported WER of the cascade approach (11.5%), it is eventually able to do so once all 11 training utterance are incorporated into the mix. It is clear from the figure that with only a single training example EM begins to over-fit the acoustic idiosyncrasies of that particular example. Though not shown in the figure, this effect is magnified for small amounts of training data when EM is run for a third and fourth iteration. One big advantage of the PMM approach is that it directly models multiple pronunciations for a single word, an avenue we begin to explore with a second set of experiments. We use a simple weight threshold 0 ,,b > r, to choose baseforms for inclusion. As in the single baseform case, we initially discard the weights once the baseforms have been chosen, but we ultimately use them during decoding later on in experiments with stochastic lexicons. Table 3.1 shows WER obtained by recognizers with lexicons trained with all 11 utterances generated under varying values of r. Choosing T = 0.1 yields the best reported WER of 11.0%, a 1.4% absolute improvement over the expert-baseline. It's interesting to note that this threshold implies an average of 2.1 pronunciations per word, almost double that of the expert lexicon which has 1.08. We also plot WER using varying thresholds as a function of the number of training utterances used in Figure 3-4. As can be seen, WER improves when using r = 0.1 even with less than 11 utterances. 3.2.3 Results using Stochastic Lexicon So far we have been using a threshold to figure out how many baseforms to include per word. During decoding we throw out each baseform's associated score and weight all of them equally. In this part we experiment with including as many baseforms as possible while keeping their associated weight. We run our experiments on the same test set as before by including the top 100 baseforms. The intuition is that since most of the weight will be concentrated on the top 5 baseforms Lexicon Adaptation using Acoustic Examples --- top baseform . -- tau-0.2 - - tau=0.1 -o-- tau-0.05 14- PMM -- 11 0 1 2 3 4 S 8 7 8 Number of uttrances used to learn baseforms (M) 9 10 11 Figure 3-4: Word Error Rate (WER) as a function of the number of example utterances used to adapt the underlying lexicon. We show results for using the un-weighted learned pronunciations while varying the T threshold, as well as, using the PMM (stochastic lexicon) including an arbitrary large amount should not hurt our scores. Again we notice improvements: the WER for the phoneme PMM drops to 10.6 while the WER for the phone PMM drops to 9.9 which results in a 25% relative improvement over the expert-crafted lexicon. Figure 3-4 plots WER using the stochastic lexicon while training on less than 11 utterances. As can be seen, the stochastic lexicon out-performs using an unweighted lexicon and requires only 6 training examples to achieve its best score. 3.3 Training Pronunciations using Noisy Acoustic Examples While the PhoneBook data used for training simulated ideal training conditions by matching the test data and the AM used, we now show the utility of un-matched speech collected cheaply from a cloud-service in generating high-quality pronunciations in a fully automatic fashion. Models that incorporate acoustic information into lexicon adaptation become particularly useful in domains where acoustic data is cheaper to obtain than expert input. In [24], example utterances of spoken names are obtained in an unsupervised fashion for a voicedialing application by filtering for interactions where the user confirmed that a call should be placed. Unfortunately, not all domains are amenable to such a convenient context-filter to find self-labeled utterances. To collect arbitrary acoustic data, we turned to the Amazon Mechanical Turk (AMT) cloud-service. AMT has been described as a work-force in the cloud since it enables requesters to post web-based tasks to any workers willing to accept micro-payments of as little as $0.005 upon completion. The service has become popular in the natural language processing community for collecting and annotating corpora, and has recently been gaining use in the speech community. In [29], we were able to collect over 100 hours of read speech, in under four days. In this work, we used a similar procedure to augment our PhoneBook corpus with another 10 example utterances for each of its 2,000 words at a cost of $0.01 per utterance. Whereas in [29] we took care to filter the collected speech to obtain high-quality sub-corpora, we took no such precautions when collecting these example utterances. Thus, in addition to other sources of mismatch between the data and our acoustic model, this noisy data poses a challenge to even a recognizer built on expert pronunciations. Due to acoustic mismatch between the utterances and our acoutic models, running the expert baseline recognizer over these 20,000 utterances yields a very high WER of 50.1%. Of course, since we could make no guarantees that the worker even read the word, the true error rate is unknown. It might seem, then, that using this data to generate valid pronunciations is a dubious exercise. Indeed, this data set confounds the cascading recognizer configuration since a single noisy utterance can throw off the entire cascade. Fortunately, the PMM approach has the nice property that a few noisy scores do not significantly affect the totals. Repeating a subset of the experiments of the previous section, we again show four iterations of the PMM approach, using the PhoneBook utterances alone, AMT-PhoneBook combined utterances, and the AMT-collected corpus alone. Despite the noisy nature of the cloud-collected corpus, Table 3.2 shows that there is little degradation in WER when using # Utts. Iter.1 Phonebook AMT Phonebook+AMT 11 10 21 12.3 12.3 12.3 2 11.5 12.0 11.6 3 11.7 13.0 11.6 4 12.0 15.3 12.0 Table 3.2: PMM results incorporating spoken examples collected via Amazon Mechanical Turk. all 21 utterances for every word. Perhaps more pleasing is the fact that generating pronunciations based on just the AMT-data still manages to out-perform even the expert generated pronunciations, achieving a WER of 12.0% when using the top un-weighted learned baseform, compared with 12.4% for the experts. 3.4 Analysis of Learned Baseforms In order to quantify some of the differences between the expert and learned baseforms, we ran NIST align software to tabulate differences between the reference expert baseform, and the top choice hypothesis of the PMM model. Of the 2000 baseform pairs, 83% were identical, while the remainder mostly contained a single substitution. Most of the substitutions involved vowels, typically a schwa. Only 2% of the data contained an additional insertion or deletion. Most of these involved retroflexed vowel sequences. Table 3.3 shows examples of common confusions including vowel and consonant substitutions, vowel/semi-vowel sequence perturbations, syllable deletions, and outright pronunciation corrections. Although the latter were few, it was encouraging to see that they did occur. Table 3.4 shows the top 10 most confused phoneme pairs, while Table 3.5 and Table 3.6 show the most deleted and inserted phonemes. 3.5 Phone Versus Phoneme Pronunciations As described in Section 2.1, our recognizer uses a lexicon of phoneme transcriptions to decode. This is made possible by the use of a manually crafted pronunciation rules FST [19] that expands sequences of phonemes to all their possible phone realizations. Given the promising Word parishoners [sic] traumatic winnifred crosby melrose arenas billowy whitener airsickness Isabel Dictionary Baseform p AE r ih sh ax n er z tr r AO rn ae tf &xkd w ih n ax f r AX dd k r ao Z b iy m eh I r ow Z ER iy n ax z b ih l OW iy w ay TF AX n er eh r SH ih kd n EH s AX S AA b eh l Top PMM Baseform p AX r ih sh ax n er z tr r AX m ae tf ax kd w ih n ax f r EH dd k r aa S b iy m eh I r ow S AX R iy n ax z b ih I AX W iy w ay TD n er eh r S ih kd n AX s IH Z AX b eh I Table 3.3: Example baseform changes between expert dictionary and top PMM hypothesis. Phonemes involved in the difference have been capitalized. Confusion Pairs AX # of Occurences =>IH 37 AO==>AA ER >B R =EB AX--> EH 31 26 23 10 Z = S 10 AX >AA IH==>AX AX==>AH AH=>AX 10 9 9 6 Table 3.4: Top 10 phoneme confusion pairs for Phonebook lexicon Deleted Phonemes AX # of Occurences 12 TD 5 Y 4 S 3 AO 2 Table 3.5: Top 5 Deleted Phonemes in Phonebook lexicon # Inserted Phonemes AO AX N TD L of Occurences 18 11 4 2 2 Table 3.6: Top 5 Inserted Phonemes in Phonebook lexicon I 0.2 0.15 0.1 0.05 0.025 0.01 Iter1 13.2 12.8 2 3 4 11.9 11.5 11.5 11.7 11.1 10.7 10.3 10.8 11.1 11.0 11.0 11.0 Table 3.7: PMM results for learning phone pronunciations from PhoneBook corpus. results shown in previous sections, we set out to learn phone transcriptions of words which would allow us to bypass the pronunciations rules file and move closer to a fully automatically learned speech recognizer. The FST implementation of the recognizer would then be simplified to R = C o L o G where L is now a lexicon FST that maps phone sequences to words. Note that since we are learning pronunciation in an isolated word setting we avoid having to worry about cross-word pronunciation variation. To automatically learn phone transcriptions of words we first have to train a graphone model instead of the graphoneme model we were using previously. This is done by using the pronunciation rules files to expand all the 200k phoneme transcriptions in our expert crafted lexicon to all their phone realizations and train our graphone model on this phone lexicon. We proceed by using our trained graphone model with a modified PMM implementation to learn phone transcriptions. The results are displayed in Table 3.7. The table displays WER scores by varying thresholds for baseform inclusion and number of EM iterations. We notice that by using a threshold of r > 0.025 we achieve significant gains over the phoneme PMM. It is interesting to note that our best phoneme PMM lexicon contained on average 2.1 baseforms per word. When we used the pronunciations rules file to expand these baseforms we ended up with an average of 5.3 baseforms per word. Our new phone PMM lexicon contains an average of 5.1 baseforms per word which could indicate that our pronuciations rules file tends to overgenerate phone realizations which in turn expands the search space and leads to lower scores. 3.6 Summary In this chapter, we have presented two Bayesian models that learn pronunciations from isolated word examples: The Cascaded Recognizer and the Pronunciation Mixture Model (PMM). We reported on experiments that produce better-than-expert pronunciations for isolated word recognition. We also showed an inherent robustness to real world noisy data when using the PMM. We experimented with using our learned pronunciations in both an unweighted dictionary, as well as, a stochastic lexicon. To understand why our learned pronuciations outperform expert baseforms, we also described some of the differences between our learned pronunciations and the expert pronunciations. Chapter 4 Discriminative Pronunciation Training for Isolated-Word Speech Recognition As a refinement step, we attempted to adjust the pronunciation weights proposed by the phoneme PMM using discriminative training to directly minimize errors. These new weights are then to be used in a stochastic lexicon during decoding. The hope is that discriminative training can address some of the concerns that maximum likelihood training cannot: e Word Confusability: On average the PMM introduces more pronunciations then the expert lexicon without taking into account that these extra pronunciations might be close to or the same as pronunciations for other words. Hence, the PMM could potentially be increasing word-confusability and therefore increasing WER during testing. e Viterbi Decoding: Our ASR system uses the Viterbi approximation while decoding i.e. it does not marginalize or sum over the probabilities of all pronunciations mapping to a word hypotheses. The Viterbi approximation forces the decoder to select the highest scoring pronunciation for any given word hypotheses. Since the PMM marginalizes over all pronunciations to maximize data likelihood, there is mismatch between our decoding and our training method. 4.1 Global Linear Models for Lexicon Learning In this section, we adapt the approach presented in section 2.4 to discriminatively train the weights of our pronunciation lexicon. Without loss of generality, we can modify Equation 3.1 as follows. w*, b* argmaxP(ub)P(b,w) w.b (4.1) We now recognize word-pronunciation pairs instead of plain words. For discriminative training, we consider the model below: W b arg max < a, CD(u, b, w) > w~b (4.2) While <D(u, b, w) can represent some complex features such as word, letter, phoneme or letterphoneme pair ngrams, we only use the ones listed below: * <bo(u, b, w) = log P(uIw) +log P(b, w) the acoustic model score and pronunciations score under our our PMM-lexicon recognizer. We define the rest of the features as: <b(u,b,w)= 1 ifb=pand w=1( 0 otherwise This reduces equation 4.2 to the following: w* b* arg max ao(log P(vb) + log P(b, w)) + ab,w w,b (4.4) The rest of our model is described as follows: " Our training examples are (ui, wi). We use the same training set used in our MLE approach, except that now we consider an utterance and its transcription individually, i.e we do riot group all utterances of the same word together. " GEN(u) is an n-best list or lattice of recognized word-pronunciation pairs (b, w), we use our MLE PMM recognizer to generate GEN(u). We optimize the parameters a as follows: INPUT: Training examples (ui, wi) INITIALIZATION: Set a = 0 ALGORITHM: for t = 1 -.T,i =1. ... N do (b*, w*) = arg nax(b,w)EGEN(ui) F(ui, b, w) ' a (boraie, wi) = arg mnax(b,w)EGEN(un) 4(ui, b, w) a if w* / wi then a = a + I(uj, borawe, wi) - ID(uj, b*, w*) end if end for OUTPUT: a Proof of convergence: Let GEN(uj) = GEN(uj) - {(b, w) Iw = wi}. In other words GEN(uj) is the set of incorrect candidates for an utterance ui. We will say that a training sequence (aj, wi) for i = 1... M is separable with margin a > 0 if there exists some vector U with |Ull = 1 such that Vi, V(p,l) c GEN(uj), Vb s.t. (b, wi) E GEN(uj) U - <b(ui, b, wi) - U . <b(ui, p, l) > a (4.5) The proof is adapted from proofs for the tagging case in [12]. Let ak be the weights before the k'th mistake is made. It follows that a' = 0. Suppose the k'th mistake is made at Take (b*, w*) to be the output proposed at this example, (b*, w*) = the i'th example. arg max(b,w)EGEN(uj) (ui, b, w) ' ak and borace to be the highest scoring correct pronunciation under this model, (baracie, wi) = arg max(b,W)EGEN(u-) 1(ui, b, w) ak. It follows from the w=wuiwi) - <b(ui, b*, w*). We take inner products algorithm updates that ak+1 - ak + <b(ui, borace, of both sides with the vector U: U -ak+1 = u - a + U .,<(ui, borae, Wi) - U . <b(ui, b*, w*) > U - ak + o (4.6) where the inequality follows because of the property of U. Because a' = 0, and therefore U . a, = 0, it follows by induction on k that for all k, U - ak+1 > ku. Because U- ak+1 < IU|||Iak+1||, it follows that Iak+1I > ka. We also derive an upper bound for |lak+ 12: Sak+1 12 _ k 2 + Ib((ui, borade, wi) - <b (ui, b*, +2ak - (<b(ui, borade, wi) Ilak2+R <b(ui, b*, w*)I|2 w*)) 2 Where R is a constant such that Vi, V(p, l) e GEN(uj), V(b, w) s.t. w = wi and (b,w) e GEN(uj) , ||<b(ui, b, w)-< (uip,l)|| (4.7) Note that 2ak . (<b(ui, borade, wi) - <b(ui, b*, w*)) 0 because (b*, w*) is the highest scoring candidate for ui under the parameters ak. It follows by induction that Iak+ 12 kR 2. Combining the bounds Ilak+1II > ku and IIak+1 II kR 2 give the results for all k that k2U 2 < ||ak+11 2 < kR 4.2 2 --> k < - (4.8) Implementing the Algorithm as an FST Our current implementation of the perceptron relies heavily on the FST toolkit [20]. We first build an FST T that expands a word to all possible word-pronunciations present in the PMM lexicon. Here a word-pronunciation is basically one entry that is a word concatenated with its pronunciation, i.e. apple -- apple-ae-p-ax-l. By inverting the FST, inv(T), we now have an FST that can map word-pronunciations back to words. GEN(u) is represented as a lattice of word pronunciation pairs D" generated by a recognizer that uses the PMM lexicon but is modified so as to hypothesize both a word and its pronunciation. We model the search space of this recognizer in a similar manner to that described in section 2.1: N = C o PoL' o GT (4.9) R Where C is the context-dependent labels and P are the pronunciation rules. We modify the baseforms FST L' so that it can map a string of phones to a word-pronunciation where the pronunciation side matches the input string of phones. Each mapping is weighted by its PMM weight P(blw). We now have to modify the LM FST so that the LM score is still used while maintaining the hypothesized word-pronunciation. This is done by composing G with inv(T) and then projecting its input to its output. We call this FST GT. We can also represent our feature vector weights as an FST. We build an FST L that maps every word-pronunciation w-p to itself weighted by a(L). We can now represent Equation 4.2 as: w*,b* = bestpath(aoLo D) (4.10) Where bestpath(.) finds the highest scoring path in a lattice. Our perceptron algorithm is now implemented as follows: INPUT: Training examples (ui, wi) INITIALIZATION: Set all arcs weights in L to zero and ao 1 ALGORITHM: for t= 1--T,i= 1-..Ndo (b*,w*) = bestpath(aoL o Du,) if w* f wi then (boracle, wi) = bestpath(aoL o Du, o F,,) Update ao and L appropriately. end if end for Where F., is an FST that only accepts word-pronunciations such that the word side matches wi. iter. 0 1 2 3 4 5 6 7 8 9 10 Train 4.7 4.0 3.9 3.0 2.7 2.7 2.6 2.7 2.5 2.6 2.5 Test 10.6 10.2 10.1 10.7 10.4 10.5 10.4 10.9 10.7 10.3 10.4 Test-Avg n/a 10.4 10.4 10.3 10.3 10.2 10.2 10.1 10.1 10.2 10.1 Table 4.1: Results for Discriminative training in WER on PhoneBook corpus. 4.3 Experimental Setup and Results We ran the perceptron for multiple iterations on the training set. We report word error rates on the training set and test set. We also test on a refinement of the perceptron algorithm called the "average parameters" method. We define a'!" to be the value for the s'th parameter after the i'th training example has been processed in pass t over the training data. Using the "averaged parameters" a, = 1/nT E_=1.T ...n a has been shown to perform significantly better than the final parameters a.". The results are shown in Table 4.1. Note here that iteration zero corresponds to using the same lexicon weights learned using the phoneme PMM discussed in section 3.2.3. As expected, WER on the training set goes down with every iteration since we are directly optimizing the feature weights on the training set. WER on the test set goes down until iteration 3 where we begin overfitting. Since the averaged parameters method corresponds to a smoothed out version of the learned parameters, we see a steady decline in WER. The best achievable score is 10.1%. Using the above described discriminative model results in an improvement in WER. While these results are very promising, we leave further investigations into the approach for future work. 4.4 Summary In this chapter, we introduced a novel mathematical framework for discriminatively training lexicon weights, along with convergence guarantees. We showed that more WER reductions could be acheived using this method. We leave further experiments using more complex features for future work. 48 Chapter 5 Pronunciation Mixture Model for Continuous Speech Recognition In this chapter, we extend our approach to the more complex task of learning pronunciations from continuous speech and testing on continuous speech. This is a more attractive application than isolated-word speech since continuous data is more readily available. To extend our isolated-word PMM requires additional considerations. We once again model the sequence of phonetic units as a hidden variable, however, the general ASR problem is now a search for the most likely string of words W* =w, W =argmaxP(Wlu)=argiax w w ... , w* given an utterance u: P(WBlu) (5.1) BGB where now, B e B is a sequence of phonemes or phones that might span multiple words and include silence. Thus, for the continuous case, we consider silence to be a phonetic unit and denote it with a '-'. For example, a possible phoneme sequence B for an utterance with transcription "the boy ate the apple" could be "dli ax b oy - - ey td - dh ah ae p ax 1". Equation 5.1 can be decomposed as follows: W =argmax( P(W)P(B|W)P(u|W,B) (5.2) BGB Where P(W) is the language model, P(B|W) can be computed using a stochastic lexicon and P(ulW, B) can be approximated with the acoustic model P(ulB). Note that the speech recognizer used in this thesis uses standard Viterbi approximations during decoding. This reduces Equation 5.2 to the following: arg max P(W)P(B|W)P(u|B) W (5.3) Wn) EM for Estimating Phoneme Pronunciations 5.1 We now extend the Pronunciation Mixture Model (PMM) framework developed for isolated word recognition in Chapter 3 to learn the appropriate weights that can model P(B|W) in continuous speech. Our training data is comprised of A utterances and their transcriptions {Ui, Wi} where W = wi, ,w. We parameterize the log-likelihood as follows: log P(ui, Wi6) log = i=1 i=1 P (ui, W, B, V|)10 FEB OeGI(Wi.B) where V, is an additional hidden variable defined to segment the phonetic sequence B into k baseforms while deleting the silences. Thus, @ is drawn from possible segmentations J(W, B) and can be indexed to retrieve a particular word-baseform pair. For example: " # 1(dh ax b oy - - ey td - dh ah ae p ax 1) = dh ax e i/ 2 (dh ax b oy - - ey td - dh ah ae p ax 1) =b oy e V)3 (dh ax b oy - - ey td - dh ah ae p ax 1) = ey td " V)4 (dh ax b oy - - ey td - dh ah ae p ax 1) = dh ah o V 5 (dh ax b oy - - ey td - dh ah ae p ax 1) = ae p ax 1 We can now further decompose the term as follows: P(ui, Wi,, B, ) = P(ujlB)P(w,. w bi, ,b) where bi = @tj(B). Our acoustic models are trained such that when B is the phoneme alphabet, a pronunciation bi is context independent and the equation above can be rewritten as: ki P(ui, l Wi, iB) kelhood, B, V@) = PO(uu (5.4) (B) j=1 where 6w,bjo = P(w) , bj). Our log-likelihood then becomes: >3 Ylog P(uiWji6) = >logE i=1 i=1 BEB ki P(uilB) # ETW(W ,B3) JOw,,,(B) j=1 The parameters, 0, are initialized to our graphoneme n-gram model scores and multiple iterations of the EM algorithm are run. E-step: Mo[w,p3= >3 P (B, @| Iui, W, 0) M[p, w, Wi,, B, V)] M-step: Mo[w,p] O7M _ .P Et',P'EV XB MOIL, p] where M[p, w, W , B, V)] is the number of times word w appears in W aligned with the pronunciation p. M[p, w, Wi, B, @b]= |{j :,@j(B) = p and W w|11 (5.5) The weights learned are directly used in a stochastic lexicon for decoding continuous speech. The term P(BIW) in Equation 5.2 can be computed as: k P(B|W) = (5.6) ,EP(Wi,B) j=1 'PEB1 uEjP 5.1.1 EM for Estimating Phone Pronunciations The phonetic rules our recognizer applies between the phoneme-based lexicon and the contextdependent acoustic models create context dependencies across words at the phone level. A misguided attempt to apply the PMM directly to learning phone pronunciations would ignore the independence assumption made in Equation 5.4, which is no longer valid. In this section, we explore a model that assigns a probability to a word-pronunciation pair given the last phoneme of the previous word's pronunciation: P(wi, bi,. ... , wn, bn) = P(wi, bilw1, bi, ... ,wi_ 1, bi-_1) n j=1 ~j7JP(wi,b~Ibs_1) i=1 ~~H P(wi, biILAST(bi_1)) i=1 Where LAST(b) is the last phone in the phone sequence b. Here we make several independence assumptions, the first is similar to the Markov assumption made in n-gram language modeling, the second references the fact that only the ending of the previous word's pronunciation can affect the current word's pronunciation. Our new features 6wb|p = P(w, b~p), where b E t3 and p is a single phone, can now be used in equation 5.4 as follows: |wi P(ui, W, B, )) = P(u I|B) 17 0 ik(B),wm L AST(pg - (b)) (5.7) j=1 5.2 Implementation Details Here we discuss some of the issues faced when implementing the continuous speech PPM for phoneme pronunciations. The framework is split into two steps: we first build a constrainedgraphoneme recognizer and run it on each training utterance to generate an n-best list of graphonemne hypotheses for the entire utterance. After creating an n-best list of graphoneme hypothesis for each utterance, we iterate through the entire list computing expected counts for every word and its pronunciation (E-step) and then normalizing the expected counts into a probability distributions (M-step). We first build a graphoneme recognizer for continuous speech. The search space is modelled by the following FST: GR=CoPoGP* (5.8) Where C maps phones to context-dependent model labels and P are the pronunciation rules. Here GP* is our graphoneme n-gram that has been expanded to allow the recognition of multiple words. This is done by taking the graphoneme n-gram used for the isolated-word PMM, concatenating it with an FST that emits a # symbol and then performing closure on this new FST. The # symbol is used as a word boundary. One problem with this approach was that the GP* FST was too big to build. To overcome this problem, we pruned out all the paths in the graphoneme n-gram that do not appear in the top 700 pronunciations for any word in our training set. This significantly reduced the size of GP* so that the recognizer search space could be efficiently built. We then distributively process each utterance and its word-transcription in the training set. We first expand the word transcription into all graphoneme sequences whose letter side match the letters of the transcription and encode those paths as an FST L. This FST is then composed with the GR FST in Equation 5.8 to constrain the search space of the graphoneme recognizer. We call this recognizer the constrained-graphoneme recognizer. Recognition is performed on the utterance in question and an n-best list of 5000 hypotheses is stored on disk. Figure 5-1 shows an example n-best list for the utterance: "houston texas". The three scores at the beginning of each line represent the total score, the acoustic model score and the graphoneme n-gram score respectively. The total score is the log-domain joint probability P(ui, Wi, B, V) used in Equation 5.4, the AM score is P(ui IB) and the graphoneme n-gram score is fl, P(wj, bj). After generating an n-best list for each utterance in our training data, we can proceed with the E-step of the first EM iteration. We keep a double hash table for every word and pronunciation that maps to the expected counts of seeing that word aligned with that specific pronunciation. For every n-best list, we first normalize the total scores to sum to one by 59.0655 83.9573 -24.8918 h/hh o/ /y u/uw s/s t/t o/ax u/n # t/t e/eh x/kd /s a/ax s/s 55.4080 83.1458 -27.7378 h/hh o/ /y u/uw s/s t/t o/ax n/n # t/t e/ey x/kd /s a/ax s/s 54.4675 83.9573 -29.4899 h/hh o/ /y u/uw s/s t/t- o/ax n/n # t/t e/eh x/kd /s a/ax s/s 50.8100 83.1458 -32.3359 h/hh o/ /y u/uw s/s t/t- o/ax n/n # t/t e/ey x/kd /s a/ax s/s 49.5057 78.9219 -29.4162 h/hh o/ /y u/uw s/s t/t o/ax n/n # t/t e/eh x/kd /s a/ax s/z 47.5087 83.4593 -35.9506 h/hh o/ /y u/uw s/s t/t o/ax n/n # t/t e/eh x/kd /s a/ax s/s 47.2153 83.9573 -36.7420 h/hh o/ /y u/uw s/s t/t o/ax n/n # t/tf e/eh x/kd /s a/ax s/s 47.0608 78.0224 -30.9616 h/hh o/ u/uw s/s t/t o/ax n/n # t/t e/eh x/kd /s a/ax s/s 46.9454 83.9573 -37.0119 h/hh o/ /y u/uw s/s t/tq o/ n/en # t/t e/eh x/kd /s a/ax s/s 46.8134 77.7686 -3019552 h/ /y o/ u/uw s/s t/t- o/ax n/n # t/t e/eh x/kd /s a/ax s/s Figure 5-1; Constrained Graphoneme Recognizer n-best hypothesis for "houston texas" along with total, AM and LM score. computing P(B, b|ui, W) = P(uiW.BV') . We then iterate through the n-best list and increment the double hashtable for every aligned word-pronunciation with its normalized total score. After computing all the expected counts across all the generated n-best lists, we normalize the counts according to the M-step equation. We now have a joint-probability distribution for word-pronunciations i.e. P(w, b) = 0 ,.b* To compute the remaining EM iterations, we again iterate through all the n-best lists to compute: P(B,I), u, Wi) = P(Blui)P(B,@, W) = P(Blui) JO 3 ,b (5.9) Where P(Blui) is the AM score extracted from the n-best list and Owjbj are the parameters computed during the previous EM iterations. After computing this joint probability for every hypothesis, we again normalize these probabilities to sum to one and compute expected counts for all the aligned word-pronunciations in the n-best lists in a similar manner to what was described before. Finally, we once again compute the updated features 0 using the M-step equation. 5.3 Experimental Setup Experiments using the SUMMIT landmark-based speech recognizer [15] were conducted in two domains: a weather query corpus [13] and an academic lecture corpus [31]. 5.3.1 Experimental Procedure To evaluate the performance of our PMM model we used the following procedure for both the weather and lecture domains. We begin by cleaning the acoustic model training set by removing utterances with non-speech artifacts to generate a training set for PMM pronunciations. We then prepare two recognizers, the first based on manually created pronunciations of all the words in the training set and the second a learned PMM recognizer that contains all the pronunciations generated by our PMM model. We then compare the Word Error Rate (WER) of both recognizers on a common test set. Thus, both recognizers use precisely the same vocabulary, but the pronunciations are chosen or weighted either by human or machine. Although the expert-lexicon leaves pronunciations unweighted, it does include a number of entries with multiple pronunciations. To keep the number of PMM pronunciations included in the search space to a reasonable size we use a 0.005 threshold to prune out low probability pronunciations. The weather query corpus is comprised of relatively short utterances, with an average of 6 words per utterance. After pruning the original training and test sets of all utterances containing non-speech artifacts, we ended up with a 87,600 utterance training set with an 1,805 word vocabulary and a 3,497 utterance test set. The acoustic models used with this corpus were trained on a large data set of telephone speech of which this corpus is a subset. The lecture corpus contains audio recordings and manual transcriptions for approximately 300 hours of MIT lectures from eight different courses and nearly 100 MITWorld seminars given on a variety of topics [31]. The lecture corpus is a difficult data set for ASR systems because it contains many disfluencies, poorly organized or ungrammatical sentences, and lecture specific keywords [8]. Compared to the weather corpus the sentences are much longer, with about 20 words per utterance on average. As in the weather domain, we discard utterances that contain non-speech artifacts from the training set used to train the acoustic model and end up with a 50K utterance training set. We then cleaned the remaining utterances to create a 6,822 utterance test set. We report results on training pronunciations and decoding with back-off maximum likelihood trained Acoustic Models [8]. We leave the use of discriminatively trained models for future work. The back-off maximum likelihood model uses a set of broad phonetic classes to divide the classification problem originating from context-dependent modeling into a set of subproblems. The reported results differ from those in [8] because we use a 25K word vocabulary of all the words in our training set. The original paper uses a 35K word vocabulary with some words absent from the training set. Graphone/Graphoneme Training A 150,000 word dictionary of manually generated phoneme pronunciations was used to train the graphoneme n-gram parameters according to the procedures described in [36]. To train our graphone model, one might be tempted to simply expand the phoneme-based lexicon according to the pronunciation rules learned in Section 2.1. Unfortunately, given the manner in which our acoustic models were trained, the beginnings and endings of pronunciations are context-dependent at the phone level. Thus, we must expand all the sentences in our weather and lecture training corpora first to their phoneme pronunciations using the manually crafted dictionary and then to their phone variations using the pronunciation rules. These phone pronunciations were properly generated since the pronunciation rules had access to the context of a word in a sentence. 5.3.2 Experimental Results The results for learning pronunciations can be seen in Figure 5.1. A baseline using a lexicon generated directly from graphonemes is shown to be significantly worse than both experts and the PMM. More interestingly, phoneme PMM pronunciations achieve more than a 1.2% WER reduction on the weather test set and a 0.4% WER reductions on the lectures test set over the hand-crafted lexicon. Both results were deemed statistically significant (p<0.001) using the Matched Pair Sentence Segment Word Error test. These results were achieved by running the EM algorithm until convergence which took around 8 iterations. It is important to note that Graphoneme L2S Expert Phoneme PMM Phone PMM Context PMM Weather 11.2 9.5 8.3 8.3 8.3 Lectures 47.6 36.7 36.3 36.1 37.7 Table 5.1: Results in WER on both the Weather and Lecture corpora using the PMM, the expert dictionary and the Graphoneme model (%) we are learning these pronunciations without human supervision and hence this technique can be reliably used to predict pronunciations for out-of-vocabulary words from continuous speech. We also show improvement over a baseline of expert crafted pronunciations by training on the same data used for training our acoustic models. This suggests that not only are we learning better-than-expert pronunciations, we are also allowing more information to be extracted from the training set that can complement the acoustic models. This smaller size vocabulary of the weather domain could explain the higher gains achieved: since the PMM includes multiple pronunciations per word, this might make them more confusable a fact that is more apparent in the 25k vocabulary Lecture domain. Another disadvantage of the PMM approach when applied to the Lecture domain is due to the longer sentences present in that corpus. The sheer size of some of the sentences caused the constrained graphoneme recognition step to fail. This happened to about 2k utterances of our 50k utterance training set because no complete path could be found through the FST lattice. We also test on learning phone pronunciations in a similar context-independent setup as that of phonemes (Section 5.1). The results are referenced as "Phone PMM" in Table 5.1. One advantage of learning phone pronunciations in a context-independent setup is that we are no longer using the pronunciation rules that might be over-expanding the search space. This fact is made apparent in Table 5.2 where we see an increase in the number of pronunciations per word but a smaller search space. A second reason to favor removal of phonological rules from the recognizer is simply that, when the lexicon is trained appropriately, they appear to be an unnecessary complication. It is also interesting to note that Phone PMM training is faster and requires only 4 EM iterations to achieve convergence. The disadvantage of the direct expansion approach we Weather Lexicon Expert Phoneme PMM Phone PMM Context PMM Lecture Lexicon Expert Phoneme PMM Phone PMM Context PMM Avg # Prons 1.2 3.15 4.0 63 Avg. # Prons 1.2 1.8 2.3 8.9 # # States 32K 51K 47K 253K States 226K 501K 243K 565K # Arcs 152K 350K 231K 936K # Arcs 1.6M 7.6M 1.2M 2.7M Size 3.5 MB 7.5 MB 5.2 MB 22 MB Size 36 MB 154 MB 28 MB 61MB Table 5.2: Lexicon statistics for the weather and lecture domains. have described so far is that phone pronunciations are context-dependent with respect to the underlying acoustic models, a fact that is not represented in the learned lexicon. We tried to model these cross-word dependencies by using the context-dependent model described in section 5.1.1. Whereas the move from graphonemes to graphones reduces the complexity of our recognizer, incorporating context dependent models grows the search space, as seen in Table 5.2. From the results in Table 5.1, however, it is clear that they are an unnecessary complication, and even hurt performance in the lectures domain. The degradation is likely due to a problem of context sparsity, which is caused by a greater mismatch between training and test in the lecture corpus. 5.4 Analysis of Learned Baseforms We first attempt to measure how much of the expert lexicon the PMM was able to recover. For the 1.8k words in the Weather Corpus vocabulary, the expert lexicon contained 2252 pronunciations (since the expert lexicon allowed multiple pronunciations for some words) of these, 2002 or 88.9% were present in the PMM lexicon with a threshold of 0.005 (all pronunciations with weights less than 0.005 were were not included in the PMM lexicon). The average rank of the pronunciations that were found in the PMM lexicon was 1.2 and the average weight was 0.73 . Given that most of the expert pronunciations were included in the PMM with a relatively high weight, we believe that little gain can be achieved by combining the expert lexicon and the PMM lexicon. Confusion Pairs AX=>IY AX=>IH AA ->AE R >ER ER >R AA =>AX IY=>Y AX=>EH UW >AX AX==>AA # of Occurences 22 19 16 15 14 10 9 9 9 8 Table 5.3: Top 10 confusion pairs for Weather Domain lexicon Deleted Phonemes AX AO TD R AA # of Occurences 15 13 4 3 3 Table 5.4: Top 5 Deleted Phonemes for Weather Corpus lexicon Similarly to Chapter 3, we again try to quantify some of the differences between the expert and learned baseforms. We run NIST align software to tabulate differences between the reference expert baseform, and the top choice hypothesis of the PMM model for the Weather Corpus lexicon (A list of learned pronunciations is listed in Appendix A). Of the 1805 baseform pairs, 81.5% were identical, while the remainder mostly contained a single substitution. We found the alignment results to be similar to the ones conducted on the isolated-word PMM lexicon. Again, most of the substitutions involved vowels, typically a schwa. Table 5.3 shows the top 10 most confused phoneme pairs, while Table 5.4 and Table 5.5 show the most deleted and inserted phonemes. About 3.4% and 3.1% of the data contained an additional insertion or deletion, respectively, more than three times the amount observed for the isolated-word case. We found that the increase in deletions could be explained by the fact that during continuous speech speakers are more likely to skip phonemes. Table 5.6 shows some examples where the PMM learned pronunciations with plausible deletions. Inserted Phonemes AX Y HH IY AO # of Occurences 13 7 7 6 5 Table 5.5: Top 5 Inserted Phonemes for Weather Corpus lexicon Word already antarctica asked barbara bratislava clothes Dictionary Baseform Top PMM Baseform ao L r eh df iy aa r eh df iy ae nt aa r KD t ax k ax ae nt aa r tf ax k ax ae s KD t ae s td b aa r b AX r ah b aa r b r ax b r AA tf ax s I aa v ax b r tf ax z 1 aa v ax k I ow DH z k l ow z Table 5.6: Example phoneme deletions between expert dictionary and top PMM hypothesis for Weather Corpus. When taking a closer look at the data to explain the increase in insertions, we found that out of the 62 insertions in our PMM lexicon about 16 were either at the beginning or at the end of the pronunciation. Upon closer examination, we found that most of these insertions were happening to words with a low number of occurrence in the training data and were erroneous. We believe that these mistakes are due to words absorbing the pronunciations of other surrounding words. For example, the word "missed" only appeared 3 times in the training set. Of the 3 occurrences, 2 were preceded by the word "I". In this case, the PMM gave a high probability to the pronunciation "ax m ih s td" i.e. the pronunciation of "I" concatenated with the pronunciation of "missed". Table 5.7 contains examples of some reasonable insertions, as well as, many erroneous insertions. We also compare the number of times these examples appear in the training set. 5.4.1 Non-Native Speaker Pronunciation In this section, we use the PMM to try to analyze pronunciation differences between native and non-native speakers. We train two PMM lexicons, one on 3k utterances of lectures given by an Indian accented English speaker and the other on 30k utterances of lectures given by Word iran wow deep guadaloupe towards missed Dictionary Baseforin ax r aa n w aw d iy pd g w aa df ax l uw pd t ao r dd z m ih s td Top PMM Baseform AY ax r ae n w AY aw d iy p AX L g w aa df ax I uw P EY t ao r dd S IY AX m ih s td rotterdam r aa tf er d ae m AX r aa tf er d ae m snowbase sea s n ow b ey s s iy EH s n ow b ey s S LY s iy # of Occurrences 16 19 2 2 1 3 1 1 2 Table 5.7: Example of phoneme insertions between expert dictionary and top PMM hypothesis for Weather Corpus. Confusion Pairs III => # of Occurences AX 8 AE=>AA S=>Z AE==>EH P==>PD 5 4 3 3 Table 5.8: Top 5 phoneme confusion pairs for Indian accented non-native talker a Dutch accented English speaker. Note that this data was not part of our original Lecture corpus training set. We again run NIST alignment on the top pronunciations learned from both speakers with the top pronunciations learned from the entire Lecture corpus. To make sure that the pronunciations learned from the lectures were correctly estimated, we restricted our analysis to words that had at least 5 examples in the training set. For Indian accented speech, we found that out of the 400 words we compared, about 87% of the pronunciations were the same while 12.3% contained substitutions. Only a small fraction of the differences were due to insertions or deletions. Table 5.8 lists the top 5 confusion pairs, mainly vowel substitutions, while Table 5.9 lists 4 word pronunciations that illustrates differences between the native and non-native speaker baseforms. As for Dutch accented speech, we found that out of the 2.5k words we compared, 88.8% of the pronunciations were similar and that most of the changes were again due to substitutions. Table 5.10 lists the top 5 confusion pairs, while Table 5.11 lists some examples. Word dependent therefore Native Speaker Pronunciation d AX p EH N D ax n td dh eh r f ao r Non-Native Speaker Pronunciations d IY p AX DF ax n td dhehrAX faor vary v AE r iy v EH r iy various only simple within v AE r iy ax s OW n I iy s IH m p ax l w IH dh IH n vEH AO n sAX w AX riyaxs 1 iy mpaxl dh AXn Table 5.9: Example of differences between native and Indian accented non-native speaker pronunciations. Confusion Pairs AX=>IY AX=>IH DD==>TD Z=>S AA >OW # of Occurences 8 5 4 3 3 Table 5.10: Top 5 phoneme confusion pairs for Dutch accented non-native talker Word fluid estimate reverse Native Speaker Pronunciation f I uw AX DD eh s t ax m AX td r AX v er s Non-Native Speaker Pronunciations f I uw IH TD eh st ax m EY td r IY v er s pops p AA pd s p OW pd s Table 5.11: Example of differences between native and Dutch accented non-native speaker pronunciations. 5.4.2 Slow vs. Fast Speech We now try to quantify the differences of word pronunciations during fast and slow speech. We conducted the following experiment: We ranked every training utterance in our Lecture corpus training set by its average word duration i.e we divided the total duration of the utterance by the number of words in its reference transcription to compute the average word duration. We then split the training corpus into two parts: " The "fast speech" training set contained the 25% of the data with the shortest word duration. " The "slow speech" training set contained the 25% of the data with the longest word duration. We discarded the remainder of the training data. After running the PMM on both training sets, we then used the NIST alignment software on the learned baseforms provided they had appeared at least 4 times in both training set. We ended up with a lexicon of 1783 words. Only about 32% of the baseforms were the same in both the "fast speech" and "slow speech" lexicon. 46% had substitutions, 21% of the baseforms in the "fast speech" lexicon had deletions and 14% had insertions. On average, fast pronunciations contained 5.2 phonemes while slow pronunciations contained 5.4 phonemes. Table 5.12, Table 5.13 and Table 5.14 list some of the top confusion pairs, insertions and deletions. We also tabulate some the example differences in Table 5.15. 5.4.3 Word Confusion in Weather Domain and Lecture Domain Lexicon Since our PMM defines a conditional distribution over words and pronunciations P(blw), by assuming that all words are equally likely i.e. P(w) = 1/IVI where V is our vocabulary, we can compute the conditional entropy H(blw) and H(wb). * H(blw) is a measure of disorder when trying to predict a pronunciation given a word. Hence, it is related to how many pronunciations are allowed per word and how uniform Confusion Pairs S=>Z IY=>AX IH =>AX R =>ER AX> EH AX=>IH Z=>S EH >AX ER > R IY> IH # of Occurences 46 38 35 35 35 33 26 26 25 20 Table 5.12: Top 10 phonenie confusion pairs between "fast" speech and "slow" speech Inserted Phonemes AX TD DD EH N # of Occurences 51 40 20 15 14 Table 5.13: Top 5 inserted phonemes in "fast" speech pronunciations Deleted Phonemes AX IY L TD S # of Occurences 130 33 33 27 27 Table 5.14: Top 5 deleted phonemes in "fast" speech pronunciations Word leadership obviously learned thousand "Slow" Pronunciation 1 iy df er sh IH PD aa bd v iy ax S I iy 1 EH R n dd th OW Y AX z ax n dd "Fast" Pronunciation 1 iy df er sh AX aa bd v iy S ax 1 iy 1 ER n dd th AW z ax n dd Table 5.15: Examples of pronunciation differences between fast and slow speech pronunciations. -- . ......................... ... ....... .................. ..... Eile Edit - -- ------__ _ ___ __ = Yiew insert Desktop 1oos Window ..- - -- Help H(blw) an(wt)for Weather and Lecture Domain 076 .5 - -. Weather H(bw) - Weather H(wb) Lecture H(blw) - - - Lecture H(wlb) -D- 0. - - 0-3 'U 0 1 1.5 2 2.5 3 8.5 4 Ntumberof EM itrations 4,5 5 5S. 6 Figure 5-2: Plot of H(wb) and H(bjw) as a function of EM iterations for both the Weather and Lecture domain. their distribution is. * H(wjb) is a measure of disorder when trying to predict a word given a pronunciation. This is a measure of how confusable words are when using a stochastic lexicon. In our lexicons we allow different words to map to the same pronunciation, hence H(w|b) allows us to measure the uncertainty of being able to map back from a pronunciation to the word that generated it. Figure 5-2 plots H(wlb) and H(blw) as a function of EM iterations for both the Weather lexicon and the Lecture lexicon. As can be seen, while the Weather domain lexicon has a higher H(blw) that seems to be increasing with EM iterations, it has a much lower H(wlb) than the Lecture domain. This confirms our initial concerns that the Lecture domain lexicon allowed for more confusion between words due to its large vocabulary size. 5.5 Acoustic Model and Lexicon Co-Training One of the research activities currently undertaken by the Spoken Language Systems (SLS) research group is utilizing crowd-sourcing for data, collection. In this section, we describe applying our PMM model to the E-book (electronic book) corpus. This corpus was collected using Amazon Mechanical Turk (AMT) and contains a large variety of words, including many that have never been used in SLS's prior speech applications. We first describe the data collection process and then discuss some of the experiments performed on this corpus. 5.5.1 Real-World Data The E-book corpus contains a list of 10K book titles, containing 30K individual words and about 9.5K unique words. The speech data was collected using the WAMI toolkit infrastructure [16] on Amazon Mechanical Turk (AMT). 10 utterances from 10 different speakers were collected for each book title resulting in a total of 100K utterances. As a final step, the data was then split up into a training set, development set and test set. For each of the 10K titles, 7 utterances from different speakers were kept for training, 1 utterance was used in the development set and the last 2 were held out for testing. 5.5.2 Baseline Approach In this part, we describe the usual procedure to build a speech recognizer for the E-book corpus. As a first attempt, we used our generic Acoustic Models (genericAM) trained on telephone speech, the expert dictionary (expertLexicon) to provide pronunciations for the 9.5K vocabulary, as well as, a Language Model n-gram trained on the training data. This approach has two limitations: the generic Acoustic Models used do riot match the E-book data because they have been trained on a different domain. The other limitation is due to the pronunciation dictionary: about 16% of the 9.5K vocabulary is out-of-vocabulary (OOV), hence the expert dictionary does not contain a pronunciation for those words. To better evaluate the performance of this recognizer, we split our test set into a 17K-utterance invocabulary test set (IV-test) where none of the words are OOV and a 3K-utterance OOV test set (OOV-test) where each utterance has at least one OOV word in its reference. Running the now 8K-vocabulary generic-AM recognizer on the test set we end up with a fairly poor WER of 24.7% on the IV-test and 98.1% on the OOV-test. In order to improve our baseline performance, we attempt to retrain our AMs using the Ebook training set. Retraining AMs requires a training set of utterances and phone-transcribed references. To generate the phone-transcriptions from the original word-transcription, we use a forced-alignment recognizer using our expert dictionary and our generic models. Again since about 16% of the words are OOV, we fail to generate forced-alignments for about 15K of our training utterances. Using these new phone alignments, we can now retrain our Acoustic Models (referred to as expertAM). A significant WER reduction can now be achieved: 11.1% on the IV-test and 90.0% on the OOV-test. The WEB. on the OOV test set remains high since each utterance has at least one word that is not in the ASR system's vocabulary. 5.5.3 PMM Approach In this section, we combine our PMM pronunciation acquisition framework with AM training and show improvements over the baseline approach. Using our generic AM, we first run the PMM for two EM iterations on our training set to generate pronunications for the 9.5K words in our vocabulary. We refer to this lexicon as pmmlemn2. We then use this new lexicon and the generic AM to run phone forced-alignment on the training set and retrain new acoustic models (referred to as pinmlem2AM). As a final step we again run the PMM for several iterations using the new retrained AM. We report our results in Table 5.16. Each row reports on WER for both the in-vocabulary and OOV test set using different AM models and different lexicons. Here pmmlem2AM are the acoustic models trained by running phone forced-alignment using the genericAM and the learned PMM lexicon from EM iteration 2. pmm2em3 is the lexicon learned by running the PMM until EM iteration 3 using the pmmlem2AM Acoustic Models. As can be seen, using this PMM and Acoustic Model Retraining framework can achieve a significant reduction in WER on both the OOV and the IV test sets. genericAM + expertLexicon expertAM + expertLexicon genericAM + pmmleml genericAM + pmmlem2 genericAM + pmmlem3 genericAM + pmmlem4 genericAM + pmmlem5 genericAM + pmmlem6 genericAM + pmmlem7 pmmlem2AM + pimnlem2 pmmlem2AM + pmm2eml pmmlem2AM + pmm2em2 pmmlem2AM + pmm2em3 pmmlem2AM + pmin2em4 IV-test 24.7 11.1 25.6 24.1 23.5 23.5 23.3 23.1 23.0 10.6 10.6 10.4 10.3 10.4 OOV-test 98.1 90.0 26.6 24.6 24.6 24.6 24.9 24.9 25.0 14.6 13.6 13.8 14.0 14.2 Table 5.16: WER in percentage for E-book domain by iteratively training the acoustic models and the lexicon. A gradual reduction in WER can be observed. pmmlem2AM pmmlem4AM + pmmlem3AM + pmmlem3AM + pmmlem3AM + + pmm2em3 pmmlem4 pmm2eml pmn2em2 pmm2em3 IV-test 10.3 10.8 11.0 10.7 10.7 OOV-test 14.0 14.6 13.6 13.9 14.0 Table 5.17: WER in percentage for E-book domain using pmmlem3. An increase in WER is observed possibly due to overfitting Overfitting Concerns One important issue to address when considering PMM-AM co-training is overfitting. To illustrate this point, we also tried using pmm1lem4 to retrain the AMs and observed degradation in performance. The results can be seen in Table 5.17. These initial results indicate that further research into this approach is needed. 5.6 Summary In this Chapter, we generalized our PMM approach to training and testing on continuous speech. We tested and showed WER improvements on both a weather query corpus and an academic Lecture corpus. After quantifying some of the differences between baseforms learned from continuous speech and expert pronunciations, we used our PMM approach to analyze baseforms learned from fast versus slow speech, as well as, non-native speakers. As a final experiment, we introduced the notion of iteratively co-training the acoutic model and the lexicon using the PMM. After testing on a corpus of book titles collected from AMT, we showed that further improvements can be achieved through this method. 70 Chapter 6 Summary and Future Work This work has introduced and compared several approaches to generating pronunciations by combining graphone techniques with acoustic examples. Furthermore, we have shown that even in the presence of significant noise, a pronunciation mixture model can reliably generate improved baseform pronunciations over those generated by experts given both isolated-word and continuous speech. Regardless, we believe that wrapping the lexicon into a statistical framework is a constructive step, which presents exciting new avenues of exploration. We have also shown that the PMM can be trained on the same data as the AM and LM, and hence requires no additional training data to build. These properties make pronunciation training a cheap and effective additional step to building an ASR system. We hope that this thesis, encourages the training of a matched lexicon when training AMs. Since our findings seem to extend to multiple corpora, the possibility of learning betterthan-expert baseforms in arbitrary domains opens up many possibilities for future work. For example, when faced with an out-of-vocabulary word with a known spelling, any system could programmatically post a task to AMT and collect example utterances to generate a high quality entry in the lexicon. There are two clear directions in which we hope to further this work. The first is to explore acoustic model and lexicon co-training in an iterative fashion, effectively taking a maximum-likelihood step along a set of coordinates in the probability space represented by the recognizer. The second is to move beyond maximnum-likelihood, and explore discriminative approaches to pronunciation learning on continuous speech. While we have only talked about discriminatively training lexicon weights for isolated-word, we hope to generalize this work for continuous speech by developing a frarnework for simultaneously discriminatively training LM weights, lexicon weights, as well as, word boundary and phone n-gram weights. Long term, the ultimate goal of this research might be to infer a phoneme set and learn pronunciations from orthographically transcribed data only. If it were feasible to simultaneously train the lexicon and the acoustic model, large vocabulary speech recognizers could be built for many different languages with little to no expert input, if given enough orthographically transcribed data. Appendix A Weather Domain Lexicon The table below lists the 1.8K words present in the Weather Query corpus vocabulary, along with their top PMM pronunciations, associated mixture weights and expert baseforms. We underline the words whose top PMM pronunciation does not match its expert counterpart. Word PMM Baseforms Weight Expert Baseform Word PMM Baseforms Weight Rxpert <oh> <um> aalten ow ahip m aa 1t iy n ao 1 tq en 1 1 0.58 0.27 ow ahip m aa 1 tq en <uh> a ahip ax ey ax b ae b aa 1 0.64 0.34 0.20 ah.fp (ax ey) aa b ae b ae b ax b ae b ey b y ax 0.19 0.18 0.16 0.15 0.90 0.86 0.12 aberdeen abilene about abu accent accumulated ao ae ae ax aa ae ax tf Itf ax n b er d iy n b ax I iy n b aw td b uw kd s ax n td k y uw m y ax I ey ax dd 0.10 0.99 0.97 0.99 0.98 0.99 0.98 ababa ae ae ax aa ae ax tf b er d iy n b ax I iy n b aw td b uw kd s eh n td k y uw m y ax ax dd abidjan able ey I ax b aa ae b ax ax b ax aa b ax iy jh aa n ax 1 b ax 1 Baseform (aa ae) b ( aa b ( aa I ax) aa b iy (jh ey b axl ax) I zh ) aa n Word accumulations activity addis address advisories afghanistan after again ahead air airline airplane akron alamos albania alberta alert alexandria algiers all along already also altoona altus [ PMM Baseforms ax k y uw rs y ax 1ey sh ax n s ax k y uw A y ax 1 ey sh ax n z ae kd t ih v ax tf iy aa df iy s aa df ax s ae df ax s Weight 0.89 ae d iy z 0.13 ae df iy z ae dr r eh s ae dd v ay z ax r iy z 0.09 0.99 0.93 ae f g ae n ax s t ae n ae f t er ax g eh n ax hh eh dd eh dd eh r eh r 1ay a eh r p 1 ey n ae kd r ax p ae 1ax m ow s aa 1ax m ow s ae 1 b ey n iy y ax ae 1b ey n iy ax aa 1b ey n y ax ae 1 b er tf 4x ax I er td ae 1ax gd a ae n dr r iy ax ae 1 ax kd s ae n dr r iy ax ae I jh y ib r z ae 1jh y er z aa 1jh y er z ao I ax 1ao ng aa r eh df iy ao 1s ow ao 1s ax ae 1 t uw n 4x ae 1 tf ax s 0.85 0.97 0.95 0.74 0.23 0.99 0.99 0.92 0.98 0.65 0.17 0.47 0.39 0.09 0.99 0.99 0.81 Expert Baseform ax k y uw m y ax sh ax n z 1ey 0.10 0.99 0.37 0.20 0.16 Weight I Expert Baseform ax b ah v 0.99 acapulco aa k ax p ax I k ow 0.31 accumulate aa k ax p ow k ow ae k ax p ax 1k ow ow k ax p ow k ow ax k y uw m y ax 1 ey td ax k y uw m y ax I ey sh ax n 0.29 0.11 0.11 0.99 active ae kd t iy v ae kd t ax v actually ae kd ch uw I iy 0.52 0.39 0.47 accumulation (ae ax) dr r eh s ae dd v ay z ax r iy z ae f g ae n ax s t ae n eh r eh r I ay n eh r p I ey n ae kd r ax n ae 1ax m ow ( z airlines airport g ow n z w er th r iy z r iy z r iy z r I ay n z r p ao r td alabama ae I ax b ae m ax 0.89 ae 1 ax b ae m aa ax 1 ae s k ax ao l b ax n iy ae I b ax k er k iy ax 1 er td s r ax 1er td s ae 1jh ih r iy ax ae 1z jh ih r iy ax ao I jh ax r iy ax ae g aa ng k w ax n 0.09 0.92 0.94 0.93 0.73 0.17 0.78 0.10 0.10 0.90 advisory africa afternoon ago ainsworth aires ae 1b ey n iy ax ae 1b er tf ax ax 1er td ae 1ax ( gd z kd s ) ae n dr r iy ax ae 1jh ih r z alaska albany albuquerque alerts ao I ax 1 ao ng ao l r eh df iy ao 1s ow algeria ae 1 t uw n ax ao 1tf ax s kd ch ax I iy kd ch uw ax 1 iy d ih sh ax n ax I df ax I ey dd df ax l ay dd dd I ey dd dd I ay dd dd v ay z ax r iy f r ax k ax ax ey eh ae ay eh eh additional adelaide s) ae ae ax ae ae ae ae ae ae algonquin ae f t er n uw n ( aa I ae ) k ax p ( ow uh 1) k ow ax k y uw m y ax 1 ey td 0.94 0.36 0.13 0.99 0.35 0.15 0.15 0.15 0.87 0.98 0.97 0.97 0.99 0.25 0.14 0.11 0.99 0.99 ae f t er ax g eh n ax hh eh dd J PMM Baseforms ax b ah v ae kd t ih v ax tf iy (aa I ae) df ax s 0.10 0.37 0.30 0.19 0.92 0.98 0.98 0.73 0.15 0.91 0.35 Word above ax k y uw n y ax 1 ey sh ax n ae kd t ax v ae kd ch ( uw ax 1 uw I I ax 1 ) iy ax d ih sh ax n ax 1 ae df ax 1 ey dd ae dd v ay z ax r iy ae f r ax k ax ae f t er n uw n ax g ow ey n z w er th ( eh I ae ) r ( iy Ieh ) z eh r 1 ay n z eh r p ao r td ae 1 ax b ae m ax ax 1 ae s k ax ao 1 b ax n iy ae 1 b ax k er k iy ax 1 er td s ae 1jh ih r iy ax ae I ax ) I g ( aa ) ng k w ax n amarillo american amman amsterdam anaheim anchorage anderson angola ann anniston another antarctica aa ae ae ax 1 td 1 t uh z m ax r ih I ow m eh r ax k ax n 0.33 0.18 0.82 0.99 ax m aa n ae m aa n ae m s t- er d ae m ae n ax hR ay m ae ng k r ax jh ae n d er s ax n ih n d er s ax n ae ng g ow ax 1 aa 0.80 0.11 0.83 0.99 0.88 0.51 0.48 0.52 ae ng g ow I ao 0.46 0.99 0.90 0.99 0.60 ae n ae n ax s t ax n ax n ah dh er ae nt aa r tf ax k ax ae nt aa r kd t ax k ax 0.38 ae ax ax ax m ax r ih 1 ow m eh r ( ax I ih) k n m aa n alright alto altos ae m s t- er d ae mj ae n ax hh ay m ae ng k ( er I r ) ax ae n d er s ax n allentown alot jh ae ng g ow 1 ax ao ax aa aa ae 0.74 0.11 0.87 0.49 0.39 0.82 0.13 0.99 1 r ay td I r ay td 1 tf ow I tf ow z 1tf ow z s ae m america m ax m eh r ax k ax amherst n ax s t ax n n ah dh er nt aa r ( tf I kd t k ax 0.09 0.99 0.85 0.11 am ae n ae ax ae ax ax 1 g aa ng k w ax n ae I ax n t aw n ax 1 aa td I aa td amount an analog ae ae ax ax m hh er s td n er s td m aw n td n ae n 0.65 0.29 0.92 0.85 0.12 ae n ax 1 aa gd 0.84 ah ae 1 ax n t aw n ax 1 aa td ao Ir ay td (ae I aa) 1 t ow ( ae aa) 1 t ow z ae m ax m eh r ( ax I ih ) k ax ae m ( hh er I er ) s td ax m aw n td ( ax | ae ) n ae n ax 1aa gd Word antonio anything anywhere appreciate april arabia arctic area argentina arkansas around arthur as asia asked aspen astoria at atlanta au august austin austria average aware back baghdad bahrain bakersfield baltimore, bangkok bangor barbados barcelona baseball PMM Baseforms ae n t ow n iy ow ax n t ow n iy ow eh n iy th ih ng eh n iy w eh r ax p r iy sh iy ey td p r iy sh iy ey td ax p r ax sh iy ey td ey pd r ax I ax r ey b iy ax aa r kd t ax kd. aa r tf ax kd aa r tf ih kd Weight eh r iy ax aa r jh ax n t iy n ax aa r k ax n s ao aa r k ax n s ax aa r k ax n s aa ax r aw n dd aa r th er ax z ae z ey zh ax ae s td 0.99 ae s kd td ae s p ax n ax s t ao r iy ax ax s t- ao r iy ax ae td ae td 1ae nt ax ax td 1 ae nt ax ow ao g ax s td ao g ah s td aa g ax s td 0.19 0.97 ao s t ax n aa s t ax n 0.83 ao s tr r iy ax aa s tr r iy ax ae v r ax jh w eh r ax w eh r 0.79 b ae kd b ae gd d ae dd b aa r ey n b aa hh ax r ey n b ey k er z f iy 1dd b ao I tf ax m ao r b ae ng k aa kd b ae ng g ao r b ae ng g er b aa r b ey d ow s b aa r b iy df ow z b aa r s ax 1ow n ax b ey s b ao 1 b ey z ax kd b ao 1 m b ey s b ao 1 0.99 0.92 0.78 0.19 0.92 0.99 0.45 Expert Baseform ae n t ow n iy ow and eh n iy th ih ng eh n iy w eh r ax p r iy sh iy ey td 0.34 0.20 0.98 0.99 0.76 0.09 0.96 0.99 0.75 0.22 0.92 0.73 0.55 annapolis annual eh r iy ax aa r jh ax n t iy n ax aa r k ax n s ao answer antigua ax r aw n dd aa r th er ( ax I ae ) z ey zh ax ae s kd td ae s p ax n ae s t ao r iy ax 0.44 0.95 0.79 ae td ( ax I ae ) td I ae nt ax 0.20 0.79 0.62 ow I ao ao g ( ah I ax ) s td 0.17 0.15 ao s t ax n 0.10 0.20 0.90 0.68 0.94 0.75 0.12 0.83 0.10 0.93 0.62 0.21 0.12 any anytime appleton approximate arab arbor are areas arizona arlington arrowhead aruba asheville ask asking assistance asuncion ao s tr r iy ax ae v r ax ( jh | zh ax w eh r 0.31 0.74 0.16 0.88 0.98 angeles ankara ey pd r ax 1 ax r ey b iy ax aa r ( tf I kd t ) ax kd 0.10 0.10 0.97 0.74 0.13 Word b ae kd b ae gd d ae dd b aa r ey n athens atlantic auckland augusta b b b b ey ao ae ae k er z f iy 1dd I tf ax m ao r ng k aa kd ng g ( ao r I er) b aa r b ey d ow s b aa r s ax 1ow n ax b ey s b ao 1 australia available aviv b bad bahamas baja bali PMM Baseforms ae n ax 1 ao gd ae n dd ax n dd ae n ae n jh ax 1 ax s ae ng k ax r ax aa ng k ax r ow aa n k aa r ax ae n k ae r ax ae ng kd y ax r ax ax n ae p ax I ax s ae n y uw ax 1 Weight 0.15 0.61 ae n y uh 1 ae n y uw w ax 1 ae n y ax 1 ae n s er ae n t iy gd w ax ae nt iy g ax eh n iy eh n iy t ay m ae p ax 1tf ax n ae pd 1tf ax n ax p r aa kd s ax m ax td ae r ax bd aa r b er aa r eh r iy ax z ae r ax z ow n ax aa r 1 ih ng t ax n ae r ow hh eh dd ax r uw b ax ae sh v ih 1 ae s kd ae s k ih ng ax s ih s t ax n s ax s ah n s iy ow n ax s ah n ch ow n ae s ax n zh ax n ax s ah n sh ax n ae th ax n z ae td I ae nt ax kd 0.13 ax td 1 ae nt ax kd ao kd I ax n dd aa kd I ax n dd ax g ah s t ax aa g ah s t ax ao s tr r ey 1y ax ax v ey 1 ax b ax 1 ax v ey 1b ax 1 ax v iy v b iy b an dd b ax hh aa m ax s b ax hh aa m ax z b aa hh aa, b aa 1 iy 0.47 0.44 0.41 0.73 0.14 0.78 0.65 0.34 0.95 0.99 0.99 0.72 0.26 0.99 0.98 Expert Baseform ( ax | ae ) n dd 0.18 0.10 0.97 0.45 0.18 0.12 0.11 0.09 0.97 0.60 ae n jh ax 1 ax,, ae ng k ax r ax ax n ae p ax 1ax s ae n ( y uh 1 y uw ax 1) 0.09 0.09 0.71 0.77 0.15 0.97 0.86 0.75 0.13 0.99 0.99 0.99 0.97 0.86 0.98 0.99 0.90 0.93 0.92 0.96 0.97 0.96 0.27 0.23 0.23 0.22 0.93 0.49 ae n s er ae n t iy gd w ax eh n iy eh n iy t ay m ae p ax 1tf ax n ax p r aa kd s ax m ey I ax ) td ae r ax bd aa r b er aa r eh r iy ax z ae r ax z ow n ax aa r 1 ih ng t ax n ae r ow hh eh dd ax r uw b ax ae sh v ih 1 ae s kd ae s k ih ng ax s ih s t ax n s ax s ah n s iy ow n ae th ax n z ( ax | ae ) td 1 ae nt ax kd ao kd I ax n dd ( ao I ax ) g ah s t ax ao s tr r ey 1y ax ax v ey 1ax b ax 1 ax v iy v b iy b ae dd b ax hh aa m ax z b aa hh aa b aa 1 iy Word baton be beaumont bedford ' PMM Baseforms b ae tq en Weight 0.92 b b b b 0.99 0.81 0.11 0.99 iy ow m aa n td aw m ah td eh dd f er dd ' Expert Baseform ( b ae tq en | b ax t aa n) b iy b ow m aa n td b eh dd f er dd '" Word bangalore bangladesh bar barbara ' PMM Baseforms b ae ng g ax 1 ao r Weight I Expert Baseform 0.75 aa ng g ax I ao r b b b b ay ih ng g ax 1 ao r ae ng g I ax d eh sh aa r aa r b r ax 0.23 b aa r b ax r ax b ae r ax m eh tr r ax kd b aa z ax 1 b ey z ax 1 b ey b iy ch b ax k ax m b ih n b ax g ih n b ey zh ih ng b ey jh ih ng b eh I ax r uw s 0.38 0.97 0.23 0.93 0.79 0.99 0.56 b iy f ao r b ax f ao r 0.79 0.20 b ( ax I iy ) f ao r b b b b b b b b b b ax ey ey eh eh eh eh eh eh eh g ih n ih ng r uw td uw td r uw td I f ae a td 1 f ae a I f r ax s td 1 gd r ey dd I g r ey dd Iv y uw 0.99 0.44 0.27 0.17 0.70 0.14 0.13 0.49 0.29 0.97 b ax g ih n ih ng b eh r uw td b b b b b b eh I m aa eh n dd er 1ih n er I ae n er I ax n er n n td 0.98 0.93 0.63 0.19 0.09 0.96 b eh 1 m aa n td b eh n dd b er 1ih n belgium belize b er n bellingham b iy eh I ax r uw s b eh 1jh ax m b ax 1 iy z b eh 1iy z bd 1 iy z b eh I ih ng hh ae m b b b b b ax s ay dd z iy s ay dd z ax th eh a d ax ax t ax z d ax ax s th eh a d ax 0.67 0.30 0.49 0.26 0.22 b ( ax | iy ) s ay dd z below berkeley bermuda bernardino b eh 1ih ng hh ax m b ax 1 ow b er kd 1 iy b er m y uw df ax b er n ax d iy n ow 0.18 0.87 betwe en billing birmir ngham b ax t w iy n b ih I ih ng z b er m ih ng 1h ae m 0.95 0.99 0.92 best better big b eh s td b eh tf er b ih gd 0.99 0.98 0.44 blah bloomington boca b 1 aa b 1uw m ih b ow k ax 0.98 0.95 0.99 b ax t w iy n b ih 1ih ng z b er m ih ng ( hh ae hh ax I ax) m b 1aa b 1uw m ih ng t ax n b ow k ax 0.17 binghamton b ey gd b ax gd b ih ng hh ax m t ax n boise b oy s iy b oy z iy b aa m b ey b ao r d ow b aa z n iy ax b aa z n iy y ax b aa z n y ax b ow th ow th b ow 1 ih ng 0.69 0.24 0.94 0.88 0.40 0.36 0.13 0.82 0.16 0.97 b oy ( z | s ) iy 0.44 b ow I ih ng bolivia b ih ng ax m t ax n b ih z m aa r kd b I ih z er dd b 1ah f b l ax f b l ao f b ow g ax t aa b ow gd ax t aa b ow g ax tf aa b ax I ih v iy ax b ow z m ax n 0.99 0.33 b ow z m ax n b r ax z ih I ( y ax | iy ax ) bonn b ow 1 ih v iy ax b aa n 0.24 0.99 b ao r n iy ow b ao s t ax n b aa s t ax n b ow 1df er b oy b oy s ax td s b oy sh ax n s b r ae dd f er dd b r tf ax z 1 aa v ax 0.99 b r aa tf ax z 1 aa v 0.29 before beginning beirut belfast belgrade bellevue barometric basel b eh 1g r ( aa Iey) dd bay beach become been begin beijing b eh 1 v y uw belarus b eh 1 f ae s td 0.58 0.11 0.96 0.97 0.97 0.96 0.99 0.57 b ae ng g 1ax d eh sh b aa r b aa r b (er Iax r r) ( ah I ax) b ae r ax m eh tr r ax kd b ( aa [ ae) z ax 1 b ey b iy ch b ( ax I iy) k ah m b ih n b ax g ih n b ey (zh I jh) ih ng 0.39 0.75 b ( eh 1 | ey 1 ) ax r uw S belmont bend berlin bern beside S' bethessda bombay bordeaux bosnia both bowling bozeman brasilia brazil breezy bring bristol british b r ax z ih b b b b b b b b b r r r r r r r r r ng t Iy ax n ax ax z ax Iiy ax ax z ih 1iy ax ax s iy 1iy ax ax z ih 1 iy z iy ih ng ax s t- ax 1 ih s t ax 1 ih tf ax sh b r ax tf ax sh 0.25 0.20 0.19 0.98 0.98 0.99 0.50 0.32 0.82 0.13 b ax th eh z d ax b aa m b ey b ao r d ow b aa z n iy ax I bismarck blizzard bluff bogota b ow th borneo boston b b b b r ax z ih 1 r iy z iy r ih ng r ih s t ax 1 b r ih tf ax sh boulder boy bradford bratislava 0.79 0.10 b eh Ijh ax m b eh 1 iy z 0.09 0.77 0.95 0.87 0.84 0.10 0.50 0.99 0.99 0.43 0.31 0.11 0.50 0.21 0.12 0.62 0.71 0.27 0.98 0.74 0.11 0.09 0.96 0.68 b eh 1ih ng ( hh ae I hh ax ax) m b ax 1ow b er kd 1 iy b er m y uw df ax b er n ( aa r I er I ax d iy n ow b eh s td b eh tf er b ih gd b ih ng (hh ax I ax) m t ax n b ih z m aa r kd b 1 ih z er dd b 1 ah f b ow g ax t aa b ( ow ax 1 ax 1 ) ih v iy b aa n b ao r n iy ow b ( aa I ao ) s t ax n b ow 1 df er b oy b r ae dd f er dd b r ( aa I ae ) tf ax ( s I z) 1 aa v ax Word brooklyn brunswick bucharest buenos bulgaria burlington but buzzards bye cairo caledonia cali call calling cambridge can canberra cannes canyon capital caracas caribbean carlo carolina casablanca causes cayman celsius centigrade central chance change charleston ' PMM Baseforms b r uh kd I ax n b r ah n z w ih kd Weight b r ah n z w ax kd b uw k ax r eh s td b w eh n ax s b w ey n ow s b w ey n ax s b ax 1 g eh r iy ax b er 1 ih ng t ax n b ah td b ax td b ah z er dd z b ao z er dd z b ay k ay r ow k ae 1 ax d ow n iy ax k ae 1ax d ow n y ax k aa 1 iy k ae 1iy k ao k ao 1 ih ng k ey m bd r ih jh k ey m b r ih jh k ae n k ax n k ae n b eh r ax k aa n 0.28 0.92 0.59 0.19 0.11 0.99 0.96 0.81 0.10 0.51 0.48 0.98 0.90 0.71 0.28 0.78 0.21 0.99 0.97 0.65 0.27 0.64 0.34 0.89 0.59 k ae k aa k ae k ey k ae k ax k ae k ax nz nz n y ax n n y ax n p ax tf ax 1 r aa k ax s r ae k ax s r ih b iy ax n 0.19 0.13 0.85 0.10 0.99 0.75 0.11 0.55 k ae r ax b iy ax n k ae r ax b iy ae n k ax r ih b iy ae n k aa r 1ow k ae r ax 1 ay n ax k ae s ax b 1 ae ng k ax k ae s ax b I aa n k ax k ae s ax b 1aa n k ae k aa s ax bd 1 aa ng k ax k ae s ax bd I ae ng k ax k aa z ax z k ay z ax z k ey m ax n k ey m eh n s eh 1s iy ax s s eh nt ax g r ey dd s eh n tr r ax 1 ch ae n s ch ey n jh ch aa r I s t- ax n ch aa r 1s t ax n 0.19 0.13 0.10 0.99 0.99 0.27 0.19 0.19 0.12 0.99 0.71 [ ' Expert Baseform b r uh kd I ax n b r ah n z w ( ax I ih) kd ' Word breckenridge bridgeport PMM Baseforms b r eh k ax n r ib jh b r ih jh p ao r td Weight b r ax z b ax n b r ih z b ey n b r ih z b ax n b r ih tq en b r uh kd 1 ay n b uh kd 1 ay n bd r uh kd 1 ay n b r ax k 1 ay n b r aw n z v ih 1 b r ah s ax 1z b uw df ax p eh s td b ah f ax 1ow b er b ae ng kd b er m aa b er m ah b ao r m aa b y uw td b ay s iy k ae 1 k ah tf ax k ae 1 k aa tf ax k ae 1 g ax r iy k ae 1 gd r iy k as 1 ax f ao r n y ax k ao I dd 0.44 0.37 0.09 0.99 0.52 0.16 0.11 0.09 0.87 0.90 0.93 0.96 0.99 0.53 0.33 0.13 0.99 0.86 0.98 0.53 0.38 0.73 0.16 0.99 0.90 k k k k k k k k aa ae ae ae ae ae ae ey 0.09 0.99 0.99 0.99 0.66 0.30 0.92 0.99 carson casper cayenne k k k k k k k k k aa r eh r y eh r ae r ax b uw aa r m eh 1 aa r m ax 1 aa r s ax n ae s p er ay ae n tf ax dd cedar brisbane, b uw k ax r eh s td b w ey n ( ax I ow ) s britain brookline, b ax 1g eh r iy ax b er 1 ih ng t ax n b ah td b ah z er dd z b ay k ay r ow k ae 1 ax d ow n iy ax brownsville brussels budapest buffalo burbank burma k aa 1 iy k ao I k ao 1 ih ng k ey m b r ih jh butte by c calcutta k ( ae I ax ) n calgary k as n b (eh r Ier ) ax k ( ae n ae n z | aa n california called k ae n y ax n k ae p ax tf ax 1 k ax r aa k ax s k ( as r ax | ax r ih ) b iy ax n cambodia camden canada cancun canton cape car care k aa r 1 ow k ae r ax 1ay n ax k ae s ax b 1 ae ng k ax 0.11 0.90 0.09 0.81 0.11 0.98 0.99 0.86 0.98 0.99 0.65 0.32 " caribou carmel k ao z ax z k ey m ax n s eh 1s iy ax s s eh nt ax g r ey dd s eh n tr r ax 1 ch ae n s ch ey n jh ch aa r 1 ( s z ) t ax n center centigrades champaign chances chapel charlotte charlottetown ' 1 dd m b ow df iy ax m d ax n n ax df ax n k uw n n k ow n n tq en pd 0.97 0.97 ' Expert Baseform h r eh k ax n r ih jh b r ih jh p ao r td b r ih z b ( ax I ey ) n b r ih tq en b r uh kd I ay n b b b b b b r aw n z v ih 1 r ah s ax 1 z uw df ax p eh s td ah f ax l ow er b ae ng kd er m ax b y uw td b ay s iy k ae 1k ah tf ax k ae 1 g ax r iy k ae 1 ax f ao r n y ax k ao 1 dd k k k k ae ae ae ae m b ow df iy ax m d ax n n ax df ax n k uw n k ae n tq en k ey pd 0.99 0.83 0.16 0.96 0.53 0.46 0.99 0.99 0.99 k aa r k eh r s iy df er 0.86 s iy df er s iy df er uw s eh nt er s eh n t iy gd r ey dd z s eh nt ax g r ey dd z s eh n t iy g r ey dd z s eh nt iy g r ey dd z sh ae m p ey n ch ae n s ax z ch ae p ax 1 sh aa r 1ax td sh er 1 ax t aw n 0.10 0.99 0.39 0.34 0.11 0.09 0.96 0.99 0.98 0.99 0.54 k ae r ax b uw k aa r m ( ax 1| eh 1) k aa r s ax n k ae s p er k ay ( ae I eh ) n s eh nt er s eh nt ax g r ey dd z sh ae m p ey n ch ae n s ax z ch ae p ax 1 sh aa r 1ax td sh aa r 1 ax t aw n Word charlottesville chatham PMM Baseforms sh aa r 1ax td s v ih 1 ch ae tf ax m k ae tf ax m ch eh kd sh ay ae n s ay eh n Weight 0.93 0.86 0.09 0.97 0.42 0.14 Expert Baseform sh aa r I ax td s v ih 1 ch ae tf ax m 0.99 0.97 0.72 0.17 sh ax k aa g ow ch ih I ch ay n iy z christi church cities claire clear cleveland sh ax k aa g ow ch ih 1 ch ay n iy z ch ay n iy z y ax tf ax v k r ih s t iy ch er ch s ih tf iy z k 1 eh r k 1ih r k 1 iy v I ax n dd 0.99 0.99 0.99 0.99 0.94 0.99 k r ih s t iy ch er ch s ih tf iy z k 1 eh r k 1 ih r k 1 iy v 1ax n dd closest cloud cloudy coastal cocoa cold coldest college cologne colombo colorado k 1 ow s ax s td k l aw dd k l aw df iy k ow s t ax 1 k ow k ow k ow I dd k ow I df ax s td k aa 1ax jh k ax 1ow n k ax 1 ah m b ow k aa 1ax r aa df ow 0.99 0.84 0.89 0.99 0.99 0.98 k 1 ow s ax s td k l aw dd k I aw df iy k ow s t ax 1 k ow k ow k ow I dd k ow 1 df ax s td k aa 1 ax jh k ax 1ow n k ax I ah m b ow k aa 1 ax r ( ae I aa) df ow 0.32 0.92 0.98 0.95 0.86 0.83 0.16 0.98 0.92 cost could countries county k aa 1ax r ae df ow k ax 1 ah m b ax s k ah m ih ng k ax m p y uw tf er k ax n d ih sh ax n k aa ng g ow k ao ng g ow k ax n eh tf ax k ax td k ow p ax n hh ey g ax n k ao s td k uh dd k ah n tr r iy z k aw nt iy 0.99 0.98 0.99 0.94 k ax n eh tf ax k ax td k ow p ax n hh ey g ax n k ao s td k ( uh I ax ) dd k ah n tr r iy z k aw nt iy course cozumel k ao r s k aa z ax m eb 1 0.94 0.43 k ao r s k ( aa I ow) z ax m eh k ow z ax n eh 1 k aa z ax n ax 1 k r iy kd k r ow ey sh ax k r ax w ey sh ax k r aa s k r ao s k ao s, k r uw z k uw p er t iy n ow k er ax n td 1 iy s ay pd r ax s d iy d ax k aa r 0.42 0.13 0.99 0.73 0.21 0.56 0.27 0.16 0.98 0.95 check cheyenne chicago chill chinese columbus coming computer condition congo connecticut copenhagen creek croatia crosse cruz cupertino currently cyprus d dakar 0.97 0.99 0.90 0.99 0.66 0.97 0.99 0.99 0.73 ch eh kd sh ay eh n Word chile sh ch ch ch ch ch china christchurch ch ih 1 ey ch iy Iey ch ay n ax k r ay s td ch er ch chattanooga checking chi I ah m b ax s m ih ng m p y uw tf er n d ih sh ax n ng g ow k k k k k k k k k k k 1 ih r w ao tf er I ay m ax td 1ow z 1 aw dd z ow s td ow td aa dd ow I df er ax 1 eh kd sh ax n aa 1ax n z ao I ax n z 0.16 0.99 0.99 0.83 0.95 0.99 0.92 0.99 0.99 0.89 0.09 k k k k k k k k k ax ax ax ah ah aa ax aa ax 0.72 k ax 1 ah m b iy ax columbia k k k k k k k k k come k ah m 0.87 complete concord k ax m k ax m p 1 iy td k aa ng k er dd 0.12 0.99 0.79 conditions k aa ng k ao r dd k ax n d ih sh ax n z 0.16 0.99 k ax n eh kd td k ax I k uw ax 1 k uw 1 k ao r p ax s k ow s t ax k aw n td k ah n tr r iy k ah p ax 1 k ah v r ax jh k r iy ey tf ax dd k r iy td k r oy k r uw s ax z 0.99 0.56 0.17 0.09 0.94 0.96 0.97 0.99 0.99 0.99 0.89 0.97 0.98 0.38 city clara clearwater 0.58 k r ih s m ax s s ih n s ax n ae tf iy s ih tf iy k l ae r ax k I ih r w ( aa ao uh ) tf er climate clothes clouds coast coat cod colder collection collins color ( k r iy kd | k r ih kd) k r ow ey sh ax k r uw z k uw p er t iy n ow k er ax n td 1 iy s ay pd r ax s d iy d aa k aa r Weight I Expert Baseform 0.42 ch ae tf ax n uw g ax 0.99 ch eh k ih ng 0.95 k ay 0.46 0.36 0.56 ch ( iy 1ey I ih I ey I ih I iy ) 0.24 0.13 ch ay n ax 0.96 k r ay s ch er ch 0.94 0.95 0.84 0.10 0.99 0.93 connect cool k r ao s as r 1 ax t aw n ae tf ax n uw g ax eh k ih ng iy ay ih I iy k r ih s m ax s s ax n s ax n ae tf iy s ih n s ax n ae tf iy s ih tf iy k I ae r ax k 1 ih r w aa tf er christmas cincinnati colombia k ax k ah k ax k ax k aa [PMM Baseforms corpus costa count country couple coverage created crete croix cruces 1 ah m 1 ah m 1 ah m I er 1 er z l er 1 er 1 r td 1ah m b iy ax bd iy ax bd y ax b iy ax 1 ay m ax td 1ow ( dh z z) l aw dd z ow s td ow td aa dd ow 1df er ax 1 eh kd sh ax n aa 1 ax n z 0.14 0.09 0.31 0.20 0.16 0.12 0.10 0.93 k ah 1er k ax 1 ah m b iy ax k ah m k ax m p 1 iy td k aa ng k ( er I ao r dd k ax n d ih sh ax n z k ax n eh kd td k uw I k k k k k k k k k k ao r p ax s ow s t ax aw n td ah n tr r iy ah p ax 1 ah v ( er I r ) ax jh r iy ey tf ax dd r iy td r oy r uw s iy z Word database PMM Baseforms d ae k ao r d ae I ax s d ae n b eh r iy d ae m b er iy d ae n b er iy d aa r w eh n d aa r w ax n d ey tf ax b ey s Weight 0.15 0.99 0.49 0.31 0.15 0.89 0.10 0.99 davenport day days dayton de d ae v ax n p ao r td d ey d ey z d ey tq en d iy 0.99 0.96 0.96 0.99 0.47 0.44 0.88 0.10 0.40 0.40 0.11 0.88 0.10 0.83 0.98 0.99 0.70 0.24 0.50 0.38 0.09 0.96 0.99 0.99 0.89 0.09 0.96 dallas danbury darwin [ Expert Baseform d ae I ax s d ae n b eh r iy Word cuba current cyclones d aa r w ax n d ( ey Iae tf ax b ey czech daily [ PMM Baseforms k r uw s ih z k r uw s ax s k y uw b ax k er ax n td s ay kd Iaa n z s ay k 1 aa n z ch eh kd d ey 1iy Weight 0.19 0.17 0.97 0.96 0.81 0.97 0.99 ch eh kd d ey 1 iy d ax k ow tf ax d ax m a~es k ax s d ae n v ih 1 d ey tf ax d ae tf ax d ey td 0.94 0.99 0.99 d d d d ax k ow tf ax ax m ae s k ax s ae n v ih 1 ( ey Iae ) tf ax 0.25 0.99 0.99 0.99 0.99 0.97 0.98 0.65 0.31 d d d d d d d ey td ey v ax s ey 1 ay td ey t ay m ey t ow n ax ih r b ao r n (ax iy) s eh m b er 0.84 d (ax 0.14 0.97 0.98 0.94 0.89 0.43 d d d d d Expert Baseform k y uw b ax k er ax n td s ay k 1 ow n z 0.18 s d ey death d eh th t eh th deep degrees delaware delmar denver describe d iy p ax 1 aa d iy pd d iy pd d iy g r iy z d ax g r iy z d eh I ax w eh r d eh 1m aa r d eh n v er d ax s k r ay bd dd d ax s k- r ay bd dd detailed d iy t eh 1 dd d iy t ey I dd d iy t eh 1 dew dhabi diego different d d d d d d direction distance djibouti dodge doing dominican dorado down driving dublin dude duluth durham uw aa b iy iy ey g ow ih f r ax n td ih f er ax ax r eh kd sh ax n jh ax b uw tf iy d ih jh ax b uw tf iy 0.99 0.50 0.42 d aa jh d uw ih ng d ax m ih n ax k ax n 0.93 0.82 0.89 d ax m ih n ax k ax d ax r aa df ow d ax r ey d ow d ax r ey df ow d aw n dr r ay v ih ng d ah bd 1ax n d uw dd d uh dd d ax 1 uw th d ao r ax m 0.10 0.56 0.19 d ih s t ax n s d ae v ax n p ao r td d ey d ey z d ey tq en d ( eh I ey iy d eh th d iy pd d ( ax | iy) g r iy z d eh 1ax w eh r d eh 1m aa r d eh n v er d ( ax | iy) s k- r ay bd d ( ax iy ) t ey 1dd d d d d uw aa b iy iy ey g ow ih f ( er I r ) ax n td date davis daylight daytime daytona dearborn december define del delhi denmark des detail detroit d ey v ax s d ey 1 ay td d ey t ay m d ey t ow n ax d ih r b ao r n d ax s eh m b er d iy s eh m b er d ax f ay n d ax s f ay n d eh 1 d eh 1 iy d eh n m aa r kd d ax d iy t eh 1 d iy t ey 1 0.74 0.23 d iy tr r oy td 0.77 d ax tr r oy td 0.20 0.69 0.30 0.93 d ( ay r I ax r ) eh kd sh ax n d ih s t ax n s jh ax b uw tf iy did difference digital directions d ih f r ax n s d ih jh ax tf ax 1 d ax r eh kd sh ax n z 0.99 0.85 0.99 d d d k district do d ih s tr r ax kd td d d uw 0.99 d ah z d ax m ih ng g ow d ah n 0.97 0.96 0.91 0.99 0.99 aajh uw ih ng ( ax I ow ) m ih n ax ax n d ax r aa df ow d aw n dr r ay v ih ng d ah bd Iax n d uw dd d ax 1 uw th d ( er Iao r ) ax m dewey does dorningo done dover drive dry d ow v er dr r ay v dr r ay t r ay dubuque dulles durango d d d d ax b y uw kd ah I ax s ow I ax s ax r ae ng g ow iy) f ay n eh 1 eh 1 iy eh n m aa r kd (ax eh ) ( ax iy) t ey 1 0.27 d iy t ae 1 d uw iy d uw r iy d ih dd 0.17 0.99 0.99 0.97 0.86 0.13 0.91 0.64 dakota damascus danville data 0.71 d (ax iy) tr r oy td d uw iy d ih dd d ih f ( er r ) ax n s d ih jh ax t ax 1 d ( ay r I ax r ) eh kd sh ax n z d ih s tr r ax kd td d uw 0.26 0.85 0.09 0.99 0.85 0.14 0.39 d ah z d ow m ih ng g ow d ah n d ow v er dr r ay v dr r ay d ax b y uw kd d ah 1 ax s d ( ax r Iao r ) ae ng g ow dusseldorf d er ax m jh ax r ax m d uw s ax I d ao r f d ax 1 s ax 1 d ao r f 0.13 0.10 0.88 0.11 ax r ey n g ow y ao r ae ng g ow d uw s ax 1 d ao r f during ao r ih ng er ih ng 0.34 0.10 0.51 0.27 d ( er Iao r ) ih ng Word earlier early earthquake east eat edinburgh edward egypt eighteenth eighty el eleventh elmira emergency end PMM Baseformis er I iy er eh r 1iy er er 1iy er th k w ey kd iy s td iy td eh dd en b er ow eh df ax n b er ow eh dd en b er gd eh dd w er dd ey dd w er dd iy jh ax pd td ey t iy n th ey tf iy eh I ax 1eh v ax n th iy 1eh v ax n th eh 1 m ay r ax ax iy m er jh ax n s iy eh n dd Weight I Expert Baseform 0.72 er I iy er 0.21 er 1 iy 0.99 0.89 er th k w ey kd iy s td 0.96 iy td 0.99 eh df ax n b er ( ax 0.52 ow I gd ) 0.26 0.13 0.76 eh dd w er dd 0.11 iy jh ax pd td 0.97 ey t iy n th 0.99 ey tf iy 0.99 0.94 eh 1 0.72 ax 1eh v ax n th 0.21 eh I m ay r ax 0.88 ax m er jh ax n s iy 0.90 eh n dd 0.84 Word e earliest earth earthquakes eastern ecuador edmonton effect eight eighth either eleven elkins else emirates PMM Baseforms jh er ih ng iy er 1iy ax s td er th er th k w ey kd s iy s t er n eh kd w ax d ao r Weight I Expert Baseform 0.09 0.95 er 1 iy ax s td 0.99 0.96 er th 0.97 er th k w ey kd s 0.99 iy s t er n eh kd w ax d ao r 0.61 eh kd w ax df ao r eh dd m ax n t ax n ax f eh kd td ey td ey td th ey th ay dh er iy dh er ax 1eh v ax n 1 eh v ax n eh I k ax n z eh 1 s eh m ax r ax td s 0.28 ih ng g 1 ax n dd iy n ih dd iy n ax dd ih n t ay r eh n t ay r ax 0.98 0.70 0.96 0.91 0.99 0.72 0.21 0.47 0.34 0.88 0.11 0.99 0.99 0.99 eh dd m ax n t ax n ( ax | iy ) f eh kd td ey td ey ( td th Ith) ( iy I ay ) dh er ax I eh v ax n eh 1 k ax n z eh Is eh m ax r ( ey I ax) td s english enough equator essen estonia eugene evansville evening ever every exact except exit expanded expected expects extended f factor fair fairfax fallen falls far ih n dd ih ng g I ax sh ax n ah f iy k w ey tf er ih z eh n eh s t ow n ly ax eh s t ow n iy y ax y uw jh iy n eh v ax n z v ih 1 iy v ax n ih jg iy v n ih ng eh v er eh v r iy aa f ax r iy eh gd z ae kd td 0.37 ey gd z ae kd td ax gd z ae kd td 0.34 0.27 eh eh eh eh 0.90 0.33 kd gd kd kd s eh pd z ih td z ih td s ax s 0.15 0.98 0.91 0.99 0.95 0.44 0.99 0.99 0.78 0.20 0.92 0.80 y uw jh iy n eh v ax n z v ih 1 iy v ( ax n I n ) ih ng enid entire ii nt ay er eh ( gd td kd ) z ae kd eh kd s eh pd td eh ( gd z | kd s ) ax td f eh r f ae kd s 0.80 I ax n 0.99 0.99 0.98 0.69 0.32 eh n t ay r 0.29 0.12 0.11 0.65 ih r iy 0.27 0.99 0.42 0.19 0.18 0.74 0.99 0.66 everett everything exactly ax eh eh eh 0.33 0.99 0.99 0.99 v eh n td s v r ax td v r iy th ih ng gd z ae kd td 1 iy eh s t ax m ey tf ax dd iy th iy ow p iy ax y ( er I ao r ) ax pd 0.11 iy v ax n iy v eh n td s iy v ax n ( ax | eh iy)vehntd I eh eh eh kd eh v ( ax r I r ) ax td v r iy th ih ng ( gd z kd s ) ae td I kd) 1iy kd s k- y uw z excuse kd s k- y uw z eh kd s k- y uw z 0.63 0.36 eh kd s p- eh kd t ax dd expand 0.56 0.43 eh kd s p- ae n dd eh kd s p- eh kd td s eh kd s t- eh n d ax dd expect expecting eh kd s p- ae n dd ax kd s p- ae n dd eh kd s p- eh kd td eh kd s p- eh kd t ih ng eh kd s p- 1ey n ax kd s p- 1 ey n eh kd s p- 1ey n dd eh kd s tr r iy n f ae bd y ax l ax s 0.94 eh kd s p- eh kd td eh kd s p- eh kd t ih ng eh kd s p- I ey n f ae f eh f eh f ao 0.80 0.17 explain eh f 0.13 0.97 eh n t ay r iy ih r iy y ih r iy eh s t ax m ey tf ax dd iy th iy ow p iy ax iy th iy ow p iy y ax iy t iy ow p iy ax y ao r ax pd y er ax pd ih ng g 1ax n dd iy n ax dd 0.25 even events eh kd s p- ae n d ax dd 0.19 0.79 estimated ethiopia europe 0.27 0.17 0.71 0.27 0.98 0.75 erie eh v er eh v r iy 0.19 0.16 0.99 f ao I z f aa r ih ng g l ax sh (ax iy) n ah f (ax iy) k w ey tf er eh s ax n eh s t ow n iy ax 0.38 eh kd s ih td iy eh kd s p- ae n d ax dd eh kd s p- eh kd t ax dd ax kd s p- eh kd t ax dd eh kd s p- eh kd s eh kd s t eh n d ax dd eh kd s t- eh n d ax dd eh f eh ax f f ae kd t er f eh r f ao england f ae kd t er f eh r extreme f eh f ao f ao f aa fahrenheit r f ae kd s 1ax n 1z r fabulous fairbanks fall r ax n hh ay td r ax n hh ay td r b ae ng kd s 1 0.89 0.57 0.16 0.12 0.99 0.95 0.99 0.96 eh kd s tr r iy m f ae bd ( y ax | y uw) l ax s f ae r ax n hh ay td f eh r b ae ng kd s f ao I Word fast fayetteville february PMM Baseforms f ao r f ae s td f ey y ax td v ih 1 f ey ax td v ih 1 f eh bd y uw eh r iy Weight 0.28 0.92 0.77 Expert Baseform f ae s td f ey ax td v ih Word failing falmouth PMM Baseforms f ao 1 ih ng f aw 1m ax th fargo f aa 1 m ax th f ae I m ax th f aa r g ow 1 0.11 0.44 f eh bd ( r uw | y uw) Weight 0.99 0.66 0.16 0.13 Expert Baseform f ao I ih ng f ( aw 1 aa I ) m ax th 0.97 f aa r g ow 0.63 0.36 0.96 0.99 0.99 0.99 0.99 0.90 f ae kd s ( eh r I er ) iy feet fergus few fifth fiji find finished first flagstaff flight flint flooding florence flurries flying foggy football forecast forecasts forget form forty fourteenth framingham f eh bd r uw eh r iy f iy td f er g ax s f er g ax z f y uw f ih th f ih f th f iy jh iy f ay n dd f ih n ax sh td f ih n iy sh td f er s td f I ae gd s t- ae f f 1 ay td f 1 ih n td f 1eh n td f 1 ah df ih ng f I ao r ax n s f 1 ao r ax n td s f 1 ao r iy z f 1 ay ih ng 0.32 0.99 f ao g iy f aa g iy f uh td b ao 1 l ax f uh td b ao l f ao r k ae s td f ao r k ae s td s f er g eh td f ao r m f ao r tf iy f ao r t iy n th f r ey m ih ng hh ae m fax f ae kd s f f ae kd s fe feeling fell fernando fifteenth fifty f ey f ay n dd f ih n ax sh td finally f ay n 1 iy f ay 1 iy f ay n dd 1 iy 0.95 f er s td f 1 ae gd s t- ae f 0.99 0.69 f I ay td f 1 ih n td fine finland five flash f 1 ah df ih ng f 1 ao r ax n s flights 0.99 0.99 f I er iy z f l ay ih ng flood floods florida 0.66 f ( aa I ao ) g iy 0.56 0.42 0.97 0.51 0.46 0.99 0.99 0.80 0.12 0.96 f iy td f er g ax s f y uw f ih ( f th I th) f iy jh iy 0.27 0.90 0.84 0.13 f iy 1 ih ng f eh 1 f er n ae n d ow f ih f t iy n th f ih f t iy f ay n f f f f f f f f f ih n 1ax n dd ay v 1 ae sh I ae sh s 1 ae sh f 1 ay td s I ah dd 1 ah dd z I ao r ax df ax 0.51 0.30 0.13 0.99 0.85 0.99 0.68 0.17 0.14 0.98 0.99 0.97 0.53 f f f f f f f ey iy 1 ih ng eh 1 er n ae n d ow ih f t iy n th ih f t iy ay n ax 1 iy f ay n f ih n 1ax n dd f ay v f1 ae sh f l ay td s f I ah dd f 1 ah dd z f 1 ( ao r ax I ao r ) df ax 0.38 0.97 0.65 0.28 forks f I ao r df ax f I ay f ao gd f aa gd f aa 1 ow ih ng f ao r f er f ao r k ae s t ax dd f ao r ax n s f ao r ax n f ao r kd s 0.99 f ao r kd s ( ax I ae ) n s ih s k fort f ao r td 0.98 f ao r td eh dr r ax kd s b er four f ao r 0.97 f ao r iy z eh n ch fourth france f ao r th f r ae n s f r ae n td s f r ae ng kd f er td. f r iy f r iy z ih ng f r eh z n ow f r ax m f r ah m f ah kd f aa kd f ao ng kd sh ax n f ah ng kd sh ax n f er dh er jh iy jh iy iy g ey n z v ih 1 0.98 0.55 0.42 f ao r th f r ae n s 0.90 0.99 0.99 0.99 0.66 0.31 f f f f 0.46 0.37 0.53 0.41 0.98 0.83 0.15 0.89 f ah kd 0.32 0.79 fly f uh td b aol fog 0.10 0.99 0.94 f ao r k ae s td f ao r k ae s td s following 0.83 f ( ao r I er ) g eh td 0.99 0.99 0.99 f ao r m f ao r tf iy f ao r t iy n th forecasted 0.97 f r ey m ih ng ( hh ae | for foreign 0.99 0.53 0.42 0.99 0.85 0.14 f I ay f ( aa Iao ) gd f aa 1 ow ih ng f ( ao r I er Iax r) f ao r k ae s t ax dd f ao r ax n ax) m francisco f r ax n s ih s k ow 0.98 fredericksburg f r eh dr r ax kd s b er gd f r iy z f r eh n ch f r ae n ch f r ay df ey f r ao s td f r aa s td f ax 1 f ao 1 f ah n iy f ah n iy ey tf ax dd f y uw ch er g ae dd z d ax n g ae 1 ax kd s iy g ae 1ax kd s iy iy g ey m g ae td w ih kd g ae td w iy kd 0.98 freeze french friday frost full funny future gadsden galaxy game gatwick 0.86 fr ow fr gd fr fr 0.12 0.99 f r ay df ey 0.99 0.67 0.26 0.64 0.34 f r ao s td frankfurt free f uh 1 freezing fresno 0.66 f ah n iy from 0.32 fuck 0.99 0.99 f y uw ch er g ae dd z d ax n 0.70 0.29 g ae 1 ax kd s iy 0.99 0.61 g ey m 0.37 g ae td w ( ax I ih) kd function further g gainesville r ae ng kd f er td r iy r iy z ih ng r eh z n ow f r ( ah l ax ) m f ah ng kd sh ax n f er dh er jh iy g ey n z v ih 1 Word geneva georgetown I PMM Baseforns jh ax n iy v ax jh ih n iy v ax jh ao r jh t aw n jh ao r zh t aw n Weight I Expert Baseform jh ax n iy v ax 0.60 0.30 jh ao r jh t aw n 0.82 0.16 great green greensboro jh er m ax n jh er m ax n iy g eh tf ih ng g ih tf ih ng g ih v g I ey sh er g 1 eh n w uh dd g 1 ax n w uh dd g ow ih ng g uh g aa td g aa g r ey td g r iy n g r iy n z b er ow 0.99 0.96 0.48 0.38 0.98 0.86 0.81 0.17 0.73 0.21 0.65 0.22 0.98 0.99 0.63 grenoble g r iy n z b ao r ow g r iy n z b aa r ow g r ax n ow b ax I 0.17 0.10 0.99 ground guadaloupe guangzhou g r aw n dd g w aa df ax I uw p ey g w aa ng z ow 0.99 0.99 0.66 guatemala g w aa ng zh uw g w aa tf ax m aa 1 ax 0.25 0.75 g w aa tf ax m aa I aa 0.14 w aa tf ax m aa 1ax 0.10 german germany getting give glacier glenwood going got ax guiana g y iy ae n gulfport guyana g iy ae n ax g aa iy ae n ax iy ae n ax g y aa n ax g ah 1f p ao r td g y ax n ax 0.21 0.15 0.09 0.09 0.80 0.49 g ay ae n ax ey ch 0.49 0.91 0.84 0.99 0.95 0.99 0.74 0.99 0.61 0.18 0.12 0.92 0.99 0.96 h hagerstown haiti hamburg hampton hanoi happen happening harbor harrisburg has hh ey g er at 0.30 aw n hh hh hh hh hh hh hh hh hh hh hh ey ae ae ae ae ae ae ae aa ae ae tf iy m b er gd m pd t ax n n oy p ax n p ax n ih ng pd n ih ng p ax n iy ng r b er r ax s b er gd z hh hh hh hh hh hh ax v ae p ax ax v aa n ax ey v ax n ax w ay iy ax w ay iy iy ih r jh er jh er g eh Word PMM Baseforms g ey n z v ih 1 ax 1 g ae 1v ax s t ax n g aa r b ax jh jh eh n ax r ax 1 Weight 0.09 0.95 0.99 0.51 go good jh eh n r ax 1 jh ao r jh jh ao r jh ax jh er m ax n t aw n g eh td jh ax b r ao I tf er g ih v ih ng jh ih v ih ng g 1 ae z g ow g Iae s k- ow g 1ae s k- aw g I aa s k- ow g 1aa s g ow g ow g uh dd 0.34 0.99 0.98 0.99 0.97 0.99 0.85 0.14 0.27 0.26 0.11 0.10 0.09 0.98 0.92 grand greece greenland g r ae n dd g r iy s g r iy n 1 ax n dd 0.98 0.95 0.48 g r ae n dd g r iy s g r iy n I ax n dd greenville g r iy n 1 ae n dd g r iy n 1 ay n dd g r iy n v ih 1 0.26 0.09 0.91 g r iy n v ih 1 g r aa tq en g w aa df ax l ax hh aa r ax g w aa df ax aa hh aa r ax w aa df ax 1 ax hh aa r ax g w aa df ax l ax hh ae r ax g w aa m g w aa r df iy ax g eh s g ah I f g ay g ay z 0.93 0.36 g r aa tq en g w aa df ax r ax 1 ax hh aa 0.81 0.92 0.99 0.98 0.99 0.74 g ay tf ax z hh ae dd hh ey I hh ae l ax f ae kd s hh ae m pd sh er hh ae ng hh ae n ow v er hh ae p ax n dd hh ax r aa r ey hh ax r aa r ax hh ax r aa r iy hh ae r ax m ih n hh aa r td f er dd hh ae tf ax r ax s 0.25 0.95 0.83 0.93 0.99 0.80 0.99 0.99 0.72 0.13 0.13 0.99 0.98 0.86 hh hh hh hh hh hh 0.12 0.99 0.91 0.94 0.88 0.11 galveston garbage general m ax n m ax n iy ( tf ih ng I tq en) g ih v g 1ey sh er g 1eh n w uh dd george georgia germantown get gibraltar giving g ow ih ng glasgow g aa td g r ey td g r iy n g r iy n z b ( er ow ao r) g r ax n ( ow aa ) b ax 1 g r aw n dd g w aa df ax 1 uw pd g w aa ng ( z | zh Ijh) (aw I ow) g w aa tf ax m aa l ax g ( ay I iy ) ( ae n ax I aa) g ah If p ao r td g ( ay I iy) ( aa n ax ey ch hh ey hh ey hh ae hh ae groton guadalajara ae) guam guardia guess gulf guy guys hh ae p ax n hh ae p ax n ih ng had hail halifax hampshire hang hanover happened harare hh aa r b er hh ae r ax s b er gd hh ae z harriman hartford hatteras g er z t aw n tf iy m b er gd m pd t ax n hh ( ae I ax ) n oy Expert Baseform g ae Iv ax s t ax n g aa r b ax jh jh eh n ( ax r I er r ) ax 1 jh ao r jh jh ao r jh ax jh er m ax n t aw n g eh td jh ax b r ao 1 tf er g ih v ih ng g 1 se z g ow g ow g uh dd 0.31 0.10 0.10 g g g g g g w aa m w ? aa r df iy ax eh s ah I f ay ay z hh hh hh hh hh ae dd ey 1 ae I ax f ae kd s ae m pd sh er ae ng hh ae n ow v er hh ae p ax n dd hh ax r aa r ( ey I ax) hh ae r ax m ax n hh aa r td f er dd hh ae tf ( er I ax r ) ax s havana haven hawaii hear 0.75 0.11 0.98 0.84 0.10 0.96 hh ax v ae n ax hh ey v ax n hh ax w ay iy hh ih r have having head heard ae tf er s ae v ae v ih ng eh dd er dd ey dd hh hh hh hh , ae v ae v ih ng eh dd er dd Word heat heavy heights help helsinki here hi high highest hill hilo hingham, hit hobart hole hollywood hong hope hotter hours how humidity hundred huntington hurricane hutchinson i iceland if image in inclement independence PMM Baseforms hh iy td hh eh v iy hh ae v iy hh ay td s hh eh 1 pd hh eh 1 s ih ng k iy hh ih r hh y er y ih r hh ay hh ay hh ay ax s td hh ih 1 hh iy 1ow hh ih ng ax m hh ih td hh ow ax b aa r td hh ow b aa r td hh ow b er td hh ax b aa r td Weight hh hh hh hh hh hh hh ow 1 ow ax 1 aa 1iy w uh dd ao 1 iy w uh dd aa ng ao ng ow pd 0.66 0.31 hh ow l 0.88 hh ao 1 iy w uh dd hh aw pd hh aa tf er aw er z hh aw hh y uw m ih df ax tf iy hh ah n d er dd 0.24 0.99 0.66 0.99 0.99 hh ah n dr r ax dd hh ah nt ih ng t ax n hh er ax k ey n hh ah ch ax n s ax n ay ay s 1 ax n dd ih f ax f ih m ax jh ih n ih n k 1 ax m ax n td 0.23 0.91 0.93 0.99 0.99 0.92 ih n d ax p eh n d ax td s ih n d ax p eh n d ax n 0.74 0.95 0.63 0.33 0.99 0.99 0.86 0.37 0.32 0.11 0.95 0.98 0.97 0.89 0.93 0.95 0.99 0.38 0.25 Expert Baseform hh iy td hh eh v iy Word heathrow hh hh hh hh ay td s eh 1 pd eh 1 s ih ng k iy ih r heidelberg hello hh hh hh hh hh hh hh hh ay ay ay ax s td ih 1 iy 1ow ih ng ax m ih td ow b ( aa r I er) td helpful hemisphere hey hidden higher highs hills hilton history 0.24 0.11 0.29 0.22 0.97 0.37 0.34 0.16 0.09 0.99 0.34 holland honduras honolulu hot hottest houston hh hh hh hh hh 0.20 hh aa tf er aw er z hh aw hh y uw m ih df ax tf iy hh ah n (d er dr r ax ) dd humid hh ah nt ih ng t ax n hh er ax k ey n hh ah ch ax n s ax n ay ay s 1ax n dd ( ih I ax ) f hummer hungary huntsville hurricanes hyannis ice 0.75 hh ow pd 0.25 0.99 0.91 0.99 Weight I Expert Baseform 0.52 hh iy th r ow 0.18 0.16 hh ay df ax 1 b er gd 0.99 0.77 hh ( eh I ax ) 1ow 0.21 hh eh 1 pd f ax 1 0.99 hh eh m ax s f ih r 0.99 0.97 hh ey hh ih dd en 0.99 0.42 hh ay r 0.33 hh ( aa I ao ) ng 0.72 hh iy th r ow hh iy td r ow hh iy tf r ow hh ay df ax l b er gd hh ax 1ow hh eh 1ow hh eh 1pd f ax 1 hh eh m ax s f ih r hh ey hh ih dd en hh ay er hh ay td er hh ay eh r hh ay z iy 1 s hh ih I z hh ih I s iy 1 z hh ih I tq en hh ih s t ax r iy hh ih s tr r iy hh ih s t er iy hh ow hh ow 1dd td hh ow 1dd hh aa 1ax n dd hh aa n d ao r ax s 0.10 0.76 0.22 0.66 [PMM Baseforms ho hold ih m ax jh (ih ax) n ih | ax) n k I ( ax eh ) m ( ax I eh ) n td ih n d ax p eh n d ax n aa n d er ax s aa n ax 1 uw 1uw aa td aa tf ax s td y uw s t ax n hh ay z hh ih I z hh ih 1 tq en hh ih s ( t ax r iy tr r 0.20 0.99 0.87 0.12 0.92 0.79 hh ow hh ow 1 dd hh aa 1 ax n dd hh aa n d ( er ao r) ax s 0.87 hh hh hh hh hh y uw m ax dd 0.55 hh y uw m ax dd 0.43 0.98 0.84 0.87 0.98 0.97 0.48 0.41 hh ah m er hh ah ng g ax r iy hh ah n td s v ih 1 hh er ax k ey n z hh ay ae n ax s ay s idaho illinois hh y uw m ih dd hh ah m er hh ah ng g ax r iy hh ah n td s v ih 1 hh er ax k ey n z hh ay ae n ax s ay s ay iy s ay ay ay df ax hh ow ih 1 ax n oy impressive ax m p r eh s ax v 0.99 0.97 0.99 0.96 0.09 0.95 0.81 aa n ax 1 uw 1 uw aa td aa tf ax s td y uw s t ax n ay df ax hh ow ih 1ax n oy ( ih I ax ) m p r eh s ax v s 0.20 inches ih n ch ax z 0.99 ih n ch ax z s india ih n d iy ax 0.93 ih n d iy ax including ih n k 1 uw df ih ng 0.96 indianapolis ih n d iy ax n ae p ax 1 ax s ih n f er m ey sh ax n 0.96 ih n d iy ax n ae p ax 1 ax s ih n f ( ao r Ier ) m ey sh ax n index ih n d eh kd s 0.99 ( ih I ax ) n k 1 uw df ih ng ih n d eh kd s indiana ih n d iy ae n ax 0.99 ih n d iy ae n ax ax n f er m ey sh ax n 0.09 indonesia ih n d ax n iy zh ax 0.46 ih n d ( ow I ax ) n iy zh ax inquiring ax n k w ay ih ng 0.48 ih n d ow n iy zh ax 0.33 0.24 0.18 0.84 innsbruck instead international ih n z b r ax kd ax n s t- eh dd ih nt er n ae sh ax n ax 0.91 interested iy n k w ay r ih ng ih n k w ay er ih ng ih n tr r ax s t ax dd ih nt ax r ax s t ax dd ih nt er n eh td 0.13 internet iowa ih nt er n ae sh n ax 1 ay ax w ax 0.18 0.62 information 0.87 0.52 (ih ng ih ng ih n) k w ay r ih ( nt ax In tr ) r ( eh | ax ) s t ax dd ih nt er n eh td III . . 0.73 0.78 ih n z b r ( uh I ax ) kd ( ih I ax ) n s t- eh dd ih nt er n ae ( sh ax sh ) n ax 1 ay ax w ax Word iran PMM Baseforms ih nt er n eh th td ih nt er n s td ih nt er n ax td ay ax r ae y ax r aa ireland is n n iy r aa n y ax r ae n ax r aa n ay r ae n ay r I ax n dd ay r ax 1 ax n dd ih z s island isle israel it ithaca ivory jacket jacksonville jamaica january jeopardy jerusalem johannesburg johns, jordan juan june jupiter k kahului kalispell katmandu kentucky key kilimanjaro kind kingdom kinshasa kittyhawk knowledge knoxville ay I ax n dd ay 1 ax ay ax 1 ih z r iy ax I ih z r ey ax 1 ih z r ey 1 ih z r iy 1 ih td ih th ax k ax ay v r iy jh ae k ax td jh ae kd s ax n v ih 1 jh ax m ey k ax jh ae n y uw eh r iy jh ae n y so r iy jh eh p er df iy jh ax p er df ly jh ax r uw s ax 1ax m jh ax r uw s ax 1 aa m jh ow hh ae n ax s b er gd jh aa n z jh ao n z jh ao r dd en w aa n jh uw n jh uw p ax tf er k ey k aa hh ax 1 uw iy k ae 1 ax s p- eh 1 k ae 1 ax s p eh I k ae td m ae n d uw k ax n t ah k iy k iy k ih I ax m ax n jh aa r ow k ay n dd k ih ng d ax mn k ih n sh aa s ax k ih n sh aa sh sh ax k ih n sh aa s aa td k ih tf iy hh ao kd k ih tf iy hh aa n aa l ax jb n aa kd s v ib 1 Weight I Expert Baseform 0.17 0.17 0.10 0.17 0.16 0.15 0.13 0.12 0.09 0.81 0.11 0.72 0.25 0.99 0.72 0.22 0.31 0.21 0.21 0.12 0.99 0.95 0.95 0.99 0.95 0.94 0.78 iraq (ih r I ay r I ax r) (aa, | ae ) n irvine islamabad ay r 1 ax n dd ( ih Iax )z ay 1ax n dd ay 1 ih z r ( iy 1 Iiy ax 1I ey ax l ) ih td ih th ax k ax ay v ax ? r iy jh ae k ax td jh ae kd s ax n v ih 1 jh ax m ey k ax jh ae n y uw eh r iy 0.72 0.15 0.84 0.85 0.62 0.99 0.95 0.99 0.89 0.99 0.60 0.19 0.19 0.75 0.23 0.90 0.93 k ae td m ase n d uw k ( eh I ax ) n t ah k iy k iy k ih 1ax m ax n jh aa r ow k ay n dd k ih ng d ax m k ih n sh aa s ax iy r ae kd 0.15 ax r aa kd 0.14 ay r aa kd er v ay n ih z 1aa m ax b aa dd ih z 1 aa m ax b ay dd ax z 1aa m ax b aa dd ay I ax n dd z ay 1z ih s t ae n b ax 1 ih s t- ae n b ax 1 ih s t ax n b ax 1 0.13 0.92 1 0.48 (ih r I ay r I ax r) (aa I ae) kd er v (ay I ax ) n ax z 1 aa m ax b aa dd 0.31 0.18 0.97 0.99 0.35 ay I ax n dd z ay 1z ih s t aa n b uw I 0.20 0.17 0.14 ax td s 0.54 ( ih ih td s 0.30 ih z 0.14 0.98 0.98 jackson jakarta janeiro jh ey jh ae kd s ax n jh ax k aa r tf ax zh ax n eh r ow jh ax n eh r ow jh ax p ae n jh er z iy jh aa bd y aa n jh r aa n jh ow n z b r ow jh ow n z b er ow hh ow z ey 0.91 0.70 0.28 I ax ) td s jh ey jh ae kd s ax n jh ax k aa r tf ax ( zh I jh ) ax n eh r ow 0.94 0.99 ax p ae n er z iy 0.80 aa bd aa n 0.66 0.33 0.81 0.18 0.98 jh ow n z b er ow 0.47 jh ( uh I | ax juneau just jh uh I ay jh ax 1ay jh uw n ow jh ah s td 0.57 jh uw n ow jh ah s td kabul jh ax s td k aa b ax 1 0.34 0.30 k aa b uh 1 k ax b uw ax I k ae b uw I 0.17 july k ax b uw kalamazoo kansas kenosha kenya kiev killington kinds k ih tf iy hh ao kd kingston n aa 1 ax jh n aa kd s v ih 1 0.42 its 0.32 0.79 0.09 ih tf ax 1 iy jose jh uw n jh uw p ax tf er k ey k aa hh ax 1 uw iy k ae I ax s p- eh I Expert Baseform 0.93 jh ow hh ae n ax s b er gd jh aa n z 0.98 Weight 0.17 ih tf ax 1 iy jonesboro w aa n Baseforms ay ax w aa ay ow w ax ay r ae kd italy jh ax r uw s ax 1ax m jh ao r dd en. [ PMM ih s t aa n b ax jh eh p er df iy 0.99 0.99 0.99 islands isles istanbul japan jersey job john 0.10 0.67 0.28 0.85 0.12 0.85 Word kitts k k k k k ax b uw 1 ae 1 ax m ax z uw ae n z ax s ax n ow sh ax ax n ow sh s ax 0.46 0.99 hh ? ow z ey 1 ) ay 0.14 0.12 0.10 0.98 0.97 0.89 k ae 1 ax m ax z uw k ae n z ax s k ax n ow sh ax 0.10 k ( eh I iy ) n y ax k iy eh v k eh n y ax k iy eh v k iy ae f k ih 1 ih ng t ax n k ay n dd z k ay n dd 0.96 k ih ng s t- ax n 0.89 k ih ng s t ax n k ih ng s t ax n 0.09 0.99 k ih td s k ih td s 0.65 0.14 0.98 0.67 k ih 1 ih ng t ax n k ay n dd z 0.32 Word kong kosovo, kuala kuwait labor laconia lagos lake lancaster lanka large las late latest lauderdale lebanon less level lewiston lhasa life lihue likelihood lima lisbon listening little liverpool located locations london look looks lottery louisiana low PMM Baseforms k aa ng k ao ng k ow s ax v ow k ao s ow v ow k ow s ow v ow k w aa 1 ax k w aa 1 aa k uw w ey td eh 1 1 ey b ax r 1 ax k ow n iy ax 1 ax k ow n iy y ax 1aa g ow s 1 ey g ow s 1 ey kd 1 ae ng k ax s t er Weight I Expert Baseform 0.70 k ( aa I ao ) ng 1 ae 1 aa 1 aa 1 aa ng k ae s t er ng k ax ng k aa, r jh 0.40 aa r jh 1 aa s 1 ey td 1 ay tf ax s td 1 ey tf ax s td 1ao df er d ey 1 1 aa df er d ey 1 1 eh b ax n aa n 1 eh b ax n ax n 1 eh s 1 eh v ax 1 I ax v ax l 1 uw ax s t ax n 1uw ax s t- ax n 1 aa s ax I ay f 1 ax hh uw ey 1 ax hh uw iy l ay kd 1 iy hh uh dd 0.32 0.94 1 iy m ax 1 iy in aa 1 ih z b ax n 0.59 1 ih s ax n ih ng 1 ax s t ax n ih ng 1ih tf ax 1 1 ih v er p ax 1 1 ih v er p ao 1 1 ow k ey tf ax dd 1 ow k ey sh ax n z 1 ah n d ax n 1 uh kd I uh kd s 1uw kd s I aa tf er iy 1 uw iy z iy ae n ax 1 ow 0.19 0.53 0.13 k ow s ax v ow "' Word know knox kodiak Weight _]Expert Baseform n ow 0.99 0.94 n aa kd s k ow df iy ae kd 0.39 laboratory PMM Baseforms n ow n aa kd s k ow df iy ae kd k ow df iy y ae kd td ow df iy ae kd k ax r iy ax k ao r iy ax k ax r uw k ao r uw k ow ao r uw k uh n m ih ng k ah n m ih ng k y ow tf ow 1ax I aa 1ae bd r ax t er 0.27 lafayette 1ae 1ae 1ae 1aa 0.13 0.83 0.16 0.92 0.91 0.99 0.68 0.10 0.59 0.39 0.90 0.49 0.65 0.18 0.67 0.99 0.50 k w aa 1 ax korea k uw w ey td eh 1 1ey b er 1ax k ow n iy ax kourou 1aa g ow s kyoto kunming la 1ey kd 1 ae ( ng [ n ) k ( ax ae ) s t er 1 aa ng k ax 1 aa r jh lahaina 1 ao df er d ey 1 lakes 0.21 0.84 1eh b ax n ( aa I ax) n 1eh s 1eh v ax 1 languages lansing largest 0.75 0.24 1 uw ax s t ax n last 0.92 1 ( aa I ae ) s ax I ay f 1 ax hh uw iy 0.15 0.99 0.83 0.10 0.98 0.62 0.09 0.99 later latvia lawrence l ay kd 1iy hh uh dd 1 iy m ax leesburg let 1 ih z b ax n 1 ih s ( ax n n ) ih ng levels 1 ih tf ax 1 1 ih v er p uw 1 lexington libya 1 ow k ey tf ax dd 1 ow k ey sh ax n z 1 ah n d ax n lightning like likely 0.98 0.59 0.35 1 uh kd 1 uh kd s lincoln 0.99 0.92 1 aa tf er iy 1 uw iy z iy ae n ax l ow 0.29 0.79 0.71 0.19 0.94 0.59 0.11 0.94 0.99 0.96 0.98 bd r ax tf er bd r ax t er iy bd r ax tf er iy f iy eh td 1aa 1 aa s 1ey td 1ey tf ax s td 0.49 0.72 ' list lithuania 0.35 0.22 0.85 0.12 k ( ax I ao ) r iy ax 0.52 k uw r uw 0.26 0.17 0.64 0.35 0.96 k uh n m ih ng 0.51 k ( iy y ) ow tf ow I (ax I aa ) 0.32 0.30 1ae bd r ax t ao r iy 0.23 0.14 0.42 f ey y eh td 1ae f iy eh td Iaa f iy y eh td 1aa hh ay n ax 1ax hh ay n ay iy 1aa hh ey n ax 1ey kd s l ay kd s 1 ae ng g w ax jh ax z I ae n s ih ng I aa r jh ax s td 1 aa r jh ax td I ae s td 1ey tf er 1 ae td v iy ax 1 aa td v iy ax 1 ao r ax n s 1 aa r ax n s 1 iy z b er gd 1eh td 1 ax 1eh 1eh v ax 1 z 1eh v w ax 1z 1eh kd s ih ng t ax n I ih b iy ax 0.14 0.10 1 ih bd y ax 1ay td n ih ng l ay kd 1 ay kd 1 iy 1 ih ng k ax n 1 ih s td 1 ih th ax w ey n iy ax 1 ih th ax w ih n iy ax 1 ih td th w ey n iy ax 1 ih tf ax w ey n iy ax 0.36 0.99 0.99 0.09 0.69 0.16 0.14 0.77 1 ( aa I ae) f ( ay I iy eh td 1ax hh ay n ax 1ey kd s 0.22 0.93 0.86 0.77 1 ae ng g w ax jh ax z 1 ae n s ih ng 1 aa r jh ax s td 0.13 0.51 1 ae s td 1 ey tf er 1 ( aa I ae ) td v iy ax 0.48 0.64 I ao r ax n s 0.95 0.94 0.29 0.99 0.46 0.35 0.09 0.77 0.15 0.78 0.63 0.98 0.99 0.99 0.44 0.17 0.16 0.09 I iy z b er gd 1 eh td 1 eh v ax 1 z 1 eh kd s ih ng t ax n 1 ih ( b iy ax I bd y ax I ay td n ih ng l ay kd I ay kd 1 iy 1 ih ng k ax n I ih s td I ih th uw ey n iy ax Word lower lows lucia lumpur luxembourg lyon macon madison madrid major malaysia malta managua manhattan ]PMM Baseforms 1 ow ao r 1 ow er 1 aw 1ao r 1 ow z 1 ax s y ax 1 uw sh ax I ah m p ao r I ah kd s ax m b er gd 1 ah kd s ax n b ao r gd. 1 ah kd s ax m b ao r gd 1 iy aa n 1 y ao, Weight 0.33 0.25 0.13 0.11 0.99 0.57 0.36 0.94 0.63 0.24 0.11 m m m m m m m m m m m m m m ey ay ae ax ey ax aa aa aa aa ax ax aa ae k ax ni k ax n df ax s ax n dr r ih dd jh er 1ey zh ax 1tf ax 1t aa 1t ax 1tf aa n aa gd w ax n aa gd w ah n aa g w aa n hh ae tq en Expert Baseform ' Word '' I ow er live 1ow z 1uw sh ax local location logan 1ah m p ao r l ah kd s ax m b er gd long looking los 0.74 0.24 1 ay ax n louis louisville 0.81 0.18 0.99 0.76 0.99 0.94 0.50 0.23 0.16 0.09 0.64 0.11 0.09 0.87 m ey k ax n m m m m ae df ax s ax n ax dr r ih dd ey jh er ax 1ey zh ax m ao 1tf lowell lowest lubbock luis ax m ax n aa g w ax lusaka, lynn m made m ( ax I ae ) n hh ae tq PMM Baseforms 1 ih td ax w ey n iy ax 1 ih v I ax v 1ow k ax 1 1ow k ey sh ax n 1ow g ax n 1ow g ax n z 1ao ng I uh k ih ng 1aa s l ao s Weight 0.09 0.82 0.14 0.96 0.89 0.86 0.13 0.96 0.99 0.57 0.40 1uw ax s 0.99 0.55 1 uw iy v ih 1 I uw ax v ax l I uw ax v ih 1 1ow ax I 1ow ax s td I ah b ax kd Iuw ax s 1uw iy z 1uw iy s 1uw iy v z I uw s aa k ax 1ih n eh m m ey dd m ey td 0.28 0.10 0.97 0.99 0.96 0.36 0.33 0.10 0.10 0.99 0.99 0.97 0.88 0.09 Expert Baseform 1 ih v 1ow k ax 1 1ow k ey sh ax n 1ow g ax n l ao ng 1 uh k ih ng 1 ( ow I ao I aa) s 1 uw (iy ax s) 1uw ( ax iy ) v ( ax ih ) 1 l ow 1ow 1 ah I uw ax 1 ax s td b ax kd ( ax s iy) 1uw s ( aa| ae) k ax 1ih n eh m m ey dd en manitoba m ae n ax t ow b ax 0.96 m ae n ax t ow b ax madras m ax dr r aa s 0.76 map marine m ae pd 0.99 0.97 0.99 0.42 0.30 0.27 0.63 0.80 0.14 0.99 0.84 0.97 0.99 0.91 0.89 0.96 m ae pd maine m er iy n make mali m m m m m m m m m m m m m m m m ey n ey kd aa1 l iy ae n ax n ae n ch eh s t er ax n ih I ax ax n ax I ax eh n iy aa r ch eh r ax t ay m aa r 1 b r ow aa r 1 b er ow aa r z aa r tq en eh r ax 1 ax n dd 0.98 0.99 0.82 0.70 0.15 0.93 0.72 0.16 0.94 0.91 0.99 0.65 0.23 0.86 0.99 0.99 m ax r iy n market marquette m m m m marseille martinique m aa r s ey mass maui may mckinley mean meant mediterranean m ae s aa aa aa aa r r r r k ax td k eh td k ae td kd ae td m aa r tq en iy kd m aa r tq en ey iy kd memorial m m m m m m iy m aw iy ey ax k ih n 1iy iy n eh n td eh df ax tf ax r ey n ax n ax m ao r ly ax 1 menlo menu m ax m ao r ax 1 m eh n 1 ow m eh n y uw m ey n y uw message mexico michigan middle midwest m m m m m m ih n y uw eh s axjh eh kd s ax k ow ih sh ax g ax n ih df ax I ih d w eh 6 td m ih dd w eh s td m aa r k ax td m aa r k eh td m aa r s ey m aa r tq en iy kd m m m m ae s aw iy ey ax k ih n I iy man manchester manila many march maritime marlborough 0.33 0.91 0.73 0.15 0.11 0.99 0.97 0.99 0.99 0.53 0.38 aa) s m ey n m ey kd m aa 1 iy m ae n m ae n ch eh s t er m ax n ih I ax m eh n iy m aa r ch m ae r ax t ay m m aa r I b ( r er ) ow m eh n td m eh df ax tf ax r ey n mars martin maryland ax n m ax m ao r iy ax I massachusetts m ae s ax ch uw s ax td s 0.97 maximum maybe m ae kd s ax m ax m m ey b iy m ae kd s ax m ax m m ey b iy me meaning m iy iy n ih ng eh dd f er dd eh 1b er n eh Ib ao r n eh m pd f ax s eh n sh ax n dd eh n t ax n dd 0.99 0.91 0.99 0.98 0.89 0.62 0.36 0.94 0.88 0.09 m ax r ih df iy ax n 0.39 m ax r ih df iy ax n m iy n iy 0.66 m ( aa dr r ax | ax dr r m eh n 1 ow m eh n y uw m m m m m td eh s ax jh eh kd s ax k ow ih sh ax g ax n ih df ax 1 ( ih I ax ) d w eb s medford melbourne memphis mentioned meridian m m m m m m m m aa r z m aa r tq en m eh r ax I ax n dd m ae s ax ch uw s ax td s m iy m iy n ih ng m eh dd f er dd m eh 1 b er n m eh m (f I pd f) ax s m eh n sh ax n dd Word PMM Baseforms Weight might m ay td mild mind m ay 1 dd m ay n dd 0.99 0.93 0.95 Expert Baseform m ay td m ay 1dd m ay n dd minneapolis m ih n iy ae p ax I ax s 0.99 m ih n iy ae p ax 1ax s minot m ih n ow 0.39 m ( ih n ow I ay n aa td m ay n aa td m ay n ae td m ay n ow 0.28 0.17 0.13 mississippi m ih n ax td m ih s ax s ih p iy 0.93 0.90 m ih n ax td m ih s ax s ih p iy missouri m ax z er iy 0.73 m ax z ( er I ao r ) ( iy m ax z ao r iy m ow g ax d iy sh uw m ow m ax n td m aa m ax n td m ah n d ey m aa n ax k ax m aa n s uw n z m aa nt iy m ah nt ax m aa nt ax r ey m aa nt ax v ax d ey eh 0.20 0.99 0.69 0.26 0.98 0.99 0.96 0.79 0.18 0.99 0.35 minute mogadishu moment monday monica monsoons monte monterey montevideo, ow m aa nt ax v ih df iy uw m aa nt ax v ax d ey Word 0.37 0.22 0.58 0.99 m ay ae m iy mid midland midwestern m ih dd m ih dd I ax n dd m ih d w eh s t ax n 0.99 0.88 0.86 milan milwaukee m ax 1 aa n m ax 1w ao k iy 0.86 0.84 m m m er m m m ih 1 w ao k iy 0.09 mobile m ih n m ih n ax s ow tf ax m ax n ax s ow tf ax m ih ng td s kd m ih n s kd m iy n td s kd ax m ih s td m ih s td m ax z uw I ax m ax z ax 1ax m ow b iy 1 0.99 0.84 0.13 0.56 0.25 0.14 0.52 0.47 0.75 0.20 0.94 moines monaco m oy n m aa n ax k ax 0.95 0.60 m aa n ax k ow 0.35 miami minnesota minsk m ah n d ey m aa n ax k ax missed m aa n s uw n z m aa nt iy missoula m aa nt ax r ey m aa nt ax v ( ax d ey ow I ih df iy ow) 0.28 0.12 m eh tr r ax p aa 1 ax tf ax n 0.41 ih dd ih dd I ax n dd ( ih I ax) d w eh s t n ax 1 (aa |ae ) n ( ih 1 | ax 1 ) w ao k 1y minh m ow g ax d iy sh uw n ow m ax n td Weight I Expert Baseform m ax r ih df iy ae n m ax r ih df iy eh n m eh tr r ax p aa 1 ax tq en m eh tr r ax p aa 1 a x tf ax n m ay ae m iy metropolitan lax ) JPMM Baseforms m ih n m ih n ax s ow tf ax m ih n s kd m ih s td m ax z uw 1ax m ow b ( ax 1 ay 1 | iy 1) m oy n m aa n ax k ow ow m aa nt ax v ih df ey 0.11 ow month months montpelier m aa nt ax v ih d ey ow 0.10 m m m m m m 0.54 0.44 0.34 0.30 0.20 0.62 aa ah ah aa aa aa n th n th n th s n n th s n p iy 1 y er mongolia m ah n th m ah n th s m aa n p ( iy 1 y er 1iy ax 0.43 ow 1 iy ax ow 1y ax ax n y ax n ax ow n ae n ax 0.36 0.10 0.48 0.47 0.99 0.82 montana montego m m m m m m monterrey montgomery m ah n t iy g ow m aa nt ax r ey m aa n td g ah m r iy 0.10 0.99 0.35 0.23 0.15 0.62 0.18 0.09 0.97 monsoon Iax m aa ng g ow aa aa aa aa aa ng ng ns ns nt aa n t iy g ow 1 y ey ) m aa n td p iy 1 y er m aa n y ax m ax n td m aa n y uw m eh td 0.22 0.45 0.31 most m aa n y uw m ax n td m ao r hh eh dd m ao r n ih ng m ao r ax s t aw n m aa r ax s t aw n m ow s td 0.21 0.99 0.99 0.72 0.24 0.89 mountain m aw n tq en 0.81 m aa n t ax n m uw v m aw ey m ax v ax m uw v iy z m ax 1v iy z m ah ch m y uw z ax kd m ay m ay r z m ay er z m ay r ax sh eh n 0.16 0.33 0.28 0.25 0.80 0.10 0.98 0.97 0.99 0.50 0.29 0.09 0.74 0.24 monument moorhead morning morristown move movies much music my myers n ih n m aa n y ax m ax n td m ao r hh eh dd m ao r n ih ng m ao r ax s t aw n monthly m ow s td monticello m m m m m m m aw n tq en montreal m aa n tr r iy ao 1 0.51 m uw v moon more morocco m m m m m m m m m m m m eh m 0.40 0.76 0.99 0.74 0.14 0.11 0.50 0.42 0.89 0.79 0.12 0.71 0.28 0.99 m uw v iy z m m m m ah ch y uw z ax kd ay ay r z moscow mount mountains movie eh n .rnozambique ah td g ah m r iy aa nt g ah m r iy ah n th 1 iy ah n th ax I iy aa n t 1iy aa nt ax s eh 1ow ah n tr r iy ao 1 uw n ao r ax r aa k ow aa r aa k ow aa ax r aa k ow aa s k ow aa s k aw aw n td aw n tq en z aw n tq en s uw v iy m eh uw v iy ow z ae m b iy kd m ( aa I ax) (n g ow 1 iy ax ng ) m aa n s uw n m aa n t ae n ax m aa n t iy g ow m aa nt ax r ey m aa n td g ah m (r ax r ) iy m ah n th 1 iy m aa nt ax ( ch I s) eh 1 ow m ( aa I ah ) n tr r iy ao 1 m uw n m ao r m ( ax ao) r aa k ow m aa s k (ow I aw) m aw n td m aw n tq en z m uw v iy m ow z ae m b iy kd Word nairobi named nanchang nantucket naples nashville natchez near neat need nepal nevada new newfoundland ' [ Expert Baseform Word munich muskogee myanmar [ PMM Baseforms Weight I Expert Baseform m y uw n ax kd 0.90 0.88 m ax s k- ow g iy m ( iy Iay) (aa ax) 0.66 n m aa r 0.16 0.16 0.98 m er tf ax 1 n aa g aa n ow 0.47 PMM Baseforms n ay r ow b ly n ey m dd n ae n dd ch ae ng Weight n ae n t al k ax td n ae n t al k eh td n ey p ax 1z n ae sh v ili I n ae ch iy z n ae ch ax z n ae ch ax z s n ih r n y ih r n iy td n iy dd n ax p ao 1 n ey p ao 1 n ax p aa 1 n eh p aa 1 n ax v ae df ax n ax v aa df ax n uw n uw f ao n dd 1ae n dd 0.78 n uw f ax n dd I ax n dd n uw f ax n dd I ae n dd n uw f ax n dd I aa n 0.28 n ay tf er 0.17 0.24 n ay td er 0.13 0.09 n iy dh eh r 0.10 n n n n ay dh er eh dh er 1ax n dd z eh v er uw ao r kd 0.09 0.99 n uw er kd n uw kd n uw z n ay ae gd r ax n ay ae g r ax n ay s n iy s n ay td n ay n n ay n t iy n th n iy n y ow n ow n ah n n aa n n ow pd n ao r m ax 1 n ao r th n ao r th hh iy s t er n n ao r th iy s t er n n ao r th w eh s td n aa td ow n ow n ow v ax n aw n ah m b er z n ah m b er z s 0.14 1 0.90 0.99 0.92 0.12 0.99 0.97 0.66 ' n ay r ow b iy n ey m dd n ae n ch ae ng n ae n t ah k ax td n ey p ax 1z n ae sh v ih 1 n ae ch ax z 0.21 0.11 0.44 '" n ih r 0.27 myrtle nagano name names nanjing naperville nashua nassau 0.99 0.99 0.36 0.23 0.19 0.09 n iy td n iy dd n ax p (aa 1 I ao 1) 0.67 n ax v ( ae [ aa ) df ax nearest 0.32 0.99 0.30 n uw n uw f ax n dd 1 ax n nebraska neither national ' m y uw n ax kd m ax s k- ow g iy m ay ae n m aa r m iy ax n m aa r m y aa n m aa r m er tf ax 1 n aa g ax n ow n aa g aa n ow n ey m n ey m z n ae n jh ih ng n ey p er v ih 1 n ae sh uw ax n ae s ao n ae s ao 1 n ae s s ao n ae sh ax n ax 1 n ae sh n ax 1 n ih r ax s td n iy ax s td n ax b r ae s k ax n iy dh er ' 0.36 0.99 1 0.98 0.89 0.89 0.39 0.16 0.13 n n n n n n ey ey ae ey ae m mz n jh ih ng p er v ih I sh uw ax ae s ao 0.80 0.14 0.77 0.11 n ae sh ( ax n n) ax 1 0.99 n ax b r ae s k ax n ( ay I iy ) dh er 0.45 n ih r ax s td dd dd newport next nicaragua nigeria nighttime nineteen ninety ninth nome noon norfolk normally northeast northern norway nothing november number 0 obispo occured n n n n uw p ao r td eh kd s td ih k ax r aa gd w ax ih k ax r aa gd w aa n n n n n n ay ay ay ay ay ay jh ih r iy ax td t ay m n t iy n n d iy nt iy n th n ow in n uw n n ao r f ax d n ao r f ao 1kd n ao r f ax 1 kd n ao r f ow k n ao r ax I iy n ao r th iy s td n ao r t iy s td n ao r dh er n n ao r w eh ey n ao r w ey n ah th ih rg n ow v eh m b)er n ah m b er ow ow ax 1 ax b ih s p ow ax b ih z p ow ow b ih s p ow ax k er dd 0.89 0.91 0.56 n uw p ao r td n eh kd s td n ih k ax r aa g w ax 0.35 0.94 0.99 0.99 0.81 0.18 0.91 0.93 0.99 0.46 0.20 n n n n ay jh ih r iy ax ay td t ay in ay n t iy n ay ( nt I n d ) iy n ay n th 0.74 0.23 0.96 0.52 0.21 0.99 0.92 0.93 0.88 0.10 0.26 0.24 0.13 0.99 newk news niagara nice n ow m n uw n n ao r f ax kd 0.17 0.11 0.99 netherlands never newark n ao r m ax 1 iy n ao r th iy s td night nine nineteenth nino, no none n ao r dh er n n ao r w ey nope normal north northeastern n ah th ih ng n ow v eh m b er n ah m b er ow northwest not ax b ih z p ow ax k er dd nova now numbers 0.98 0.68 0.99 0.59 0.25 0.76 0.23 0.99 0.96 0.99 n eh dh er 1 ax n dd z n eh v er n (uw er I uw aa r ao r) kd n uw z n ay ae g r ax n ( iy | ay ) s n n n n n n ay td ay n ay n t iy n th iy n y ? ow ow ah n 0.99 0.99 0.99 0.90 0.09 n n n n ow pd ao r m ax 1 ao r th ao r th iy s t er n 0.92 0.58 n ao r th w eh s td n aa td 0.92 0.99 0.81 0.11 0.16 0.14 0.99 0.95 0.46 0.25 n ow v ax n aw n ah m b er z Word PMM Baseforms Weight Expert Baseform october off aa kd t ow b er 0.97 0.88 0.11 0.95 0.92 0.93 0.84 0.12 0.45 0.33 aa kd t ow b er ogden okay old omaha once only ao f aa f aa gd d ax n ow k ey ow 1 dd ow m ax hh aa ow m ax hh ao w ah n s w aa n s w ah n s eh ow n 1 iy ao n 1 iy ow n iy ow 1 iy ow ax 1 iy ao 1 iy options aa pd sh ax n z orange ao r n jh oregon ao r ax n jh ao r ax g ax n ao r ax g aa n Word ao f aa gd d ax n ow k ey ow 1 dd ow m ax hh ( aa I ao) PMM Baseforms Weight Expert Baseform n ow m b er z 0.15 0.93 0.99 ow kd 1 ax n dd ax k er ow kd ocean of ow sh ax n ax v ao f er aa f er ow hh ay ow ow kd I ax hh ow m ax ax 1 ih m p iy ax offer w ah n s oklahoma olympia ow n 1 iy on one ow 1ih m p iy ax aa n w ah n w aa n aa n t eh r iy ow ao r ao r df er ow r df er ao r iy ax n td ao r iy eh n ao r iy eh nt ao r 1 ax n z ontario or order aa pd sh ax n z ao r ax n jh n dd ax k er ohio 0.20 0.31 0.18 0.14 0.12 0.11 0.09 0.89 0.56 0.43 0.88 0.10 1ax oakland occur orient ao r ax g ( aa Iax) n orleans 0.86 0.76 0.73 ow sh ax n ( ah I ax ) v ao f er 0.25 0.98 0.98 0.48 0.36 0.94 0.90 0.09 0.96 0.93 0.80 0.18 0.66 ow hh ay ow ow kd 1 ax hh ow m ax (ow l ax) l ih m p iy ax ( ah I aa ) n w ah n aa n t eh r iy ow ao r ao r df er ao r iy ( eh I ax ) n td 0.22 0.11 0.71 ao r 1 ( ax | iy I iy ax ) nz orlando orly oslo ottawa out outside overcast oxford p pakistan palo paraguay ao r 1ae n d ow w ao r 1 iy ao r 1 iy aa z 1 ow aa s 1 ow aa tf ax w aa aa tf ax w ax aw td aw td s ay dd ow v er k ae s td aa kd s f er dd aa kd s f er td p iy p iy ey p ae k ax s t ae n p ae k ax s t ax n p aa k ax s t aa n p aa 1 ow p ae 1 ow 0.98 0.73 0.26 0.47 0.46 0.57 0.34 0.98 0.91 0.99 0.62 0.35 0.67 0.26 0.62 0.11 0.10 0.63 0.33 ao r 1 ae n d ow p ae r ax g w ay 0.46 p ae r ax g w ( ay ao r I iy osaka aa ( z | s ) 1ow other our outlook over overnight ozone aw td s ay dd ow v er k ae s td aa kd s f er dd pacific palm p iy p ae k ax s t ae n aa r 0.43 0.31 0.23 0.92 ( aw er I aa r) 0.98 ow v er 0.99 ow v er n ay td ow z ow n p aa m p aa 1 m p ae n ax m aa p aa r dd en p aa r df ax n p aa r kd p ax r tf ax s ax p ey tf ih ng p ax r t ih s ax p ey tf panama pardon park participating p ( ae I aa) 1 ow 0.20 0.99 ax ey aa r ey aw td 1 uh kd ow v er ow v er n ay td ow z ow n ow z ow n ax td p ax s ih f ax kd aa tf ax w ax aw td ao r 1 iy n z ow s aa k ax ah dh er ey 0.94 0.79 0.19 0.92 0.73 0.23 0.91 0.74 0.17 0.99 0.77 ow s aa k ax ah dh er aw td 1 uh kd p ax s ih f ax kd p aa 1 ? m p ae n ax m aa p aa r dd en p aa r kd p aa r t ih s ax p ey tf ih ng 0.17 ih ng paris part partly paso paul paz pegasus pendleton p p p p p p p p p p p p p p ae ae ae aa aa ae ao aa aa eh ey eh eh eh r ax g w eh ey r ax g w ey r ax s r td r td 1 iy s ow 1 z ay z g ax s ax s g ax s ax s n d ax 1 tf ax n n dd ax 1 tf ax n n ax 1tf ax n 0.22 0.22 0.97 0.99 0.94 0.85 0.97 0.78 0.11 0.79 0.16 0.34 0.33 0.20 pasadena past p ae s ax d iy n ax p ae r ax s paulo p aa r td p aa r td 1 iy p ae s ow p aa 1 ow p ao I ow pearson k er s ax n 0.99 peking p iy k ih ng p ey k ih ng 0.65 penh p ae n pensacola p eh n s ax k ow I ax p eh g ax s ax s peoria p y ao r iy ax p iy ao r iy ax p eh n d ax 1 tf ax n percentage peru p er s eh nt ax jh p er uw p ax r uw 0.97 0.83 0.71 0.24 0.85 p ao I p aa z , p ae s td , 0.99 0.99 0.49 p ae s ax d iy n ax p ae s td p ( aw I ao) 1 ow 0.37 p ih r s ax n p iy k ih ng 0.33 0.36 0.25 p eh n p eh n s ax k ow 1 ax p iy ao r iy ax p er s eh nt ax jh p er uw Word pennsylvania people percent perth petersburg philippines phoenix picture pine pittsburgh. pizza places plains planning play plus pocatello poland PMM Baseforms p eh n s ax I v ey n y ax p iy p ax 1 p er s eh n td p er th p iy tf er z b er gd f ih I ax p iy n z f iy n ax kd s p ih kd ch er p ay n p ih td s b er gd p iy td s ax p 1ey s ax a p 1ey n z p 1ay ae n z p1 ae n i ng p 1ey p 1ay p 1ah s p 1ax s p ow k ax t eh 1ow p aa k ax t eh I ax p ow I ax n dd Weight I Expert Baseform 0.99 | p eh n s ax 1v ey n y ax 0.92 p iy p ax 1 0.81 p er s eh n td 0.89 p er th 0.98 p iy tf er z b er gd 0.99 f ih I ax p iy n z f iy n ax kd s 0.99 0.99 p ih kd ( ch sh ) er 0.99 p ay n 0.99 p ih td s b er gd. 0.99 p iy td s ax 0.95 p 1 ey s ax z 0.90 p 1 ey n z p aa 1 ax n p ao r td p ao r td s rn ax th p aa s ax b ih 1ax tf iy p ax k ih pd s iy p r aa gd p r aa kd p r ax d ih kd td 0.09 1 0.84 0.11 0.77 0.15 0.90 0.09 0.97 " Word philadelphia PMM Baseforms f ih 1 ax d eh If iy ax Weight [_Expert Baseform 0.97 f ih 1 ax d eh 1f iy ax 0.99 0.91 price n aa m f ow n p iy eh r p ih s k ae tf ax w ey p ax s k ae tf ax w ey p ax s k ae tf ax w ay p ih td s f iy I dd p 1ey s p 1ey n p 1ae n ax td p 1ae n ey p 1 ae td p 1ax td p 1iy z p 1ih m ax th p 1 ih m ax th td p oy n td p ao I p ow 1 p ow ax 1 p aa pd y ax 1 ey sh ax n p ao r td I ax n dd p ao r ch ax g ax I p aa s ax b ax 1 p aa s ax b ax 1 z p ao I p aa ax 1 p aw ax 1 p r ax s ih p ax t ey sh ax n p r ax d ih kd t ax dd p r ax d ih kd sh ax n p r eh z ax n td p r eh s kd p r eh sh kd p r ax t ao r iy ax p r ax t ao r y ax p r ay s princeton probably project p r ih n s t ax n p r aa b ax 1 iy p r aajh eh kd td 0.90 0.99 p p p p p 0.16 projection providence provo puerto pyongyang p ao r tf ow p y aa ng y ae ng 0.37 0.47 qingdao p y ah ng y ae ng p y aa ng y ax ih ng ch iy ng d aa ow 0.25 0.11 0.70 jh ih ng d aw k ax b eh kd k w ax b eh kd k w ax b ae kd k r iy 0.16 phnom phone pierre piscataway pittsfield place plain planet platte p I ae n ih ng p Iey please plymouth p 1ah s point pole p ow k ax t eh I ow p ow I ax n dd population 0.92 0.98 0.96 0.99 0.95 0.73 0.14 0.99 p aa I ax n p ao r td p ao r td s m ax th p aa s ax b ih 1ax tf iy p ax k ih pd s iy p r aa gd portland portugal possible p r ax d ih kd td precipitation p r ax d ih kd t ih ng p r ax d ib kd sh ax n z p r eh z ax df ax n td p r eh sh er p r ih tf iy p r ih n s p r ax n td s p r aa b ax b ih I ax tf iy p r ow gd r ae m p r axjh eh kd t ax dd p r ax v ay dd 0.99 0.99 0.96 0.99 0.99 0.73 0.25 0.94 p p p p p p predicted prediction present presque p p p p p 0.87 0.12 0.43 0.40 0.99 p r aa v ax n s t- aw n k y uw z k y uw 0.50 0.49 k y uw quality queens question k w aa 1ax tf iy k w iy n z k w eh s ch ax n 0.97 0.90 0.52 k w aa I ax tf iy k w iy n z k w eh ( s ch I sh ch ? ) ax n quincy k w eh sh ax n k w eh sh ch ax n k w ih n s ly k w ih n z ly k ih n iy 0.29 0.10 0.48 0.32 0.14 pollen port portsmouth possibility poughkeepsie prague predict predicting predictions president pressure pretty prince probability program projected provide provincetown pueblo puson q r aa v ax n s t aw n r aa v ax n s t ax v n w eh bd Iow w eh b 1ow uw s aa n 0.99 0.99 0.99 r ax d ih kd t ih ng r ax d ih kd sh ax n z r eh z ax df ax n td r eh sh er r ih tf iy r ih n s pr iy pr pr pr aa b ax b ih Iax tf ow gd r ae m ax jh eh kd t ax dd ax v ay dd p w eh b 1ow p uw s aa n powell pretoria quebec k w ih n ( s I z ) iy query ' r aajh ax kd td r ax jh eh kd sh ax n r aa v ax df ax n s r ow v ow ao r tf ax 0.86 0.38 0.37 ( f n I p n I n ) aa m f ow n p iy eh r p ax s k ae tf ax w ey 0.10 0.97 0.99 0.98 0.76 p ih td s f iy Idd p 1ey s p 1 ey n p 1 ae n ax td 0.09 0.86 0.12 0.99 0.71 0.09 0.93 0.39 0.37 0.13 0.85 0.99 0.99 0.67 0.32 0.49 0.39 0.11 0.95 0.99 0.99 0.99 0.60 0.31 p 1 ae td p 1iy z p 1 ih m ax th p oy n td p ow 1 p n p p p aa pd y ax 1ey sh ax ao r td I ax n dd ao r ch ax g ax 1 aa s ax b ax 1 p aw ax 1 p r ax ax n p r ax p r ax p r eh p r eh s ih p ax t ey sh d ih kd t ax dd d ih kd sh ax n z ax n td s kd 0.87 p r ax t ao r iy ax 0.09 0.99 p r ay s 0.79 0.99 0.93 0.99 0.59 0.56 p r ih n s t ax n p r aa b ax b 1iy p r ( ax I aa ) jh eh kd td p r ax jh eh kd sh ax n p r aa v ax df ax n s p r ow v ow p ( w eh ao) r tf ( ow I ax ) p y ( aa I ah) ng y (aa I ae ) ng ch iy ng d aw k ( ax w ax) b eh kd 0.32 0.09 0.34 k w ( ih r I eh r ) iy Word quite r rain [ PMM Baseforms Weight k w ay td aa r r ey n 0.99 0.99 record r ey n dd r ey n ih ng r ey n iy r ey n jh r ae p ax dd r ae p ax td r ax t ow n r eh df ih ng r iy df ih ng r iy 1iy r ax s iy v dd r eh k ax gd n ih sh ax n r eh k ax gd n dd ih sh ax n r eh k er dd regarding r iy g aa r df ih ng 0.99 region relative r iy jh ax n r eh 1 ax tf ih v r eh 1 ax tf ax v r ax p iy td r iy p iy td r iy p ao r td r ax p ao r td r iy p ao r td s r ax p ao r td s r ax k w eh s td r iy k w eh s td r iy s t- aa r td r eh s tr r aa n td s 0.95 r eh s t ax r aa n td s 0.21 r ey kd y ax v ih kd 0.46 r ey kd jh ax v ih kd r ey kd y ax v iy kd 0.19 0.13 rhodes richmond right rise r ow dd z r ih ch m ax n dd r ay td r ay z 0.91 river road rochester r ih v er r ow dd r aa ch eh s t er 0.91 0.99 rockford rocky rome, r aa kd f er dd r aa k iy r ow m 0.93 0.99 rosa rouge r ow z ax r uw zh r uw jh r ah td 1 ax n dd s ae kd r ax m eh nt ow s eh dd s ey 1 ax m s ao 1 z b eh r iy s ae 1v ax d ao r s ey m s ae n 0.91 rained raining rainy range rapid raton. reading really received recognition repeat report reports request restart restaurants reykjavik rutland sacramento said salem salisbury salvador same san. 0.85 0.99 0.98 0.99 0.97 0.72 0.17 0.79 0.86 Expert Baseform k w ay td aa r r ey n r r r r r ey ey ey ey ae n dd n ih ng n iy n jh p ax dd r ax t ow n. r ( eh I iy ) df ih ng 0.12 0.83 0.99 0.87 r iy 1iy r ax s iy v dd r eh k ax gd n ih sh ax n Word questions quit quito rabat raincoat rainfall rains raleigh 0.12 0.99 0.66 0.25 0.50 0.48 0.55 0.44 0.50 r ( eh k er ax k ao r) dd r ( iy I ax) g aa r df ih ng r iy jh ax n r eh I ax tf ( ih I ax ) v ranges r (iy readings ax ) p iy td 0.59 0.94 0.99 0.99 0.97 0.98 0.80 0.18 0.96 0.98 0.81 0.99 0.84 0.95 0.99 0.99 k w eh sh ax n z k w ih td k w eh td k iy tf ow k iy t ow k w iy tf ow r aa b ao td r ey n k ow td r ey n f ao 1 r ey n z r ey n s r aa 1 iy 0.35 r ao 1 iy 0.10 r ey n jh ax z 0.68 r ey n jh iy z s 0.27 r r r r r 0.98 0.80 r ae p ax dd z r iy ch 0.10 0.49 r iy df ih ng z ae p ax dd z iy ch iy ch s iy df ih ng z iy ih ng z 0.33 0.31 0.62 0.75 0.48 0.28 0.15 0.94 0.96 0.98 0.88 0.83 0.44 0.61 0.38 r ( iy I ax ) p ao r td s recently recognize 0.93 redmond r iy s ax n td 1 iy r eh k ax gd n ay z ax r eh k ax gd n ay z r eh k ax gd n ay z z r eh k ax gd n ay s ax r eh dd m ax n dd regina r ax jh ay n ax 0.53 r iy jh ay n ax 0.28 related r ax jh ax n ax r iy 1ey tf ax dd 0.10 0.75 r r r r ax 1ey dd iy n ow iy f r ey z ax p ao r tf ax dd 0.23 reno rephrase reported r ih v er r ow dd r aa ch eh s t er 0.29 0.10 republic r iy p ao r tf ax dd r iy p ao r df ax dd r ax p ah bd l ax kd r aa kd f er dd r aa k iy r ow m r iy p ah bd I ax kd r eh s td r eh s tr r aa n td 0.42 rest restaurant r eh s tr r ax n td r ax s t ax r ax n td r eh s t er aa n td r eh s t ax n r ow dd r iy k ax r iy k ow r iy ow r ay z ih ng ax r ay z ih ng r iy aa dd 0.18 ax ) k w eh s td r iy s t- aa r td r eh s ( tr r It ax r ) ax j aa) n td s r ey kd (iy ax I y ax) v ih kd r ow dd z r ih ch m ax n dd r ay td r ay z r ow z ax r uw zh r ah td 1 ax n dd s ae kd r ax m eh nt ow s eh dd s ey I ax m s ao 1z b ( er I eh r ) iy s ae 1 v ax d ao r s ey m s ae n reston rhode rica rico rio rising riyadh k w ih td k ( iy t I w iy tf ) ow r ax r ey r ey r ey b aa td n k ow td n f ao 1 n z 0.11 r iy s iy v r ax s iy v r (iy k w eh ( s ch I sh ch ? ) ax n z 0.23 receive 0.45 0.99 Weight r ( iy Iax ) p ao r td 0.48 0.49 rapids reach Expert Baseform PMM Baseforms k w eh r iy k w ih r iy k w eh s ch ax n z 0.46 r ( aa 1 I ao 1 ) iy r ey n jh ax z r ax s iy v r iy s ax n td 1 iy r eh k ax gd n ay z 0.22 0.13 0.12 0.99 0.89 0.99 0.59 0.57 0.93 0.40 0.15 0.09 0.99 0.99 0.98 0.99 0.94 0.90 0.09 0.41 r eh dd m ax n dd r ax ( jh ) n ax g) ( iy I ay r ( iy I ax ) 1 ey tf ax dd r iy n ow i ( iy ax ) f r ey z r ( iy I ax ) p ao r tf ax dd r ( iy ax kd ax ) p ah bd 1 r eh s td r eh s ( tr r I t ax r ax I aa ) n td eh s t ax r ow dd iy k ax iy k ow iy ow ay z ih ng iy aa dd Word sanibel PMM Baseforms s ae n ax b eh I Weight 0.83 II s aa nt iy aa g ow s ah nt iy aa g ow s ae nt iy aa g ow 0.16 0.40 0.29 0.27 0.83 0.09 0.35 s ae n ax b eh santiago sao s aw S ow sarajevo s ae r ax y ey v ow Expert Baseform s ae n ax b ( ax I Ieh I Word Weight 0.24 r iy y ae dd r iy ax dd r ow ax n ow kd r aa kd r aa kd v ih 1 r ow m ey n y ax r ah m ey n y ax 0.17 0.09 0.90 0.98 0.93 0.45 0.17 root r ah m ey n iy ax r ow m iy n y ax r uw td 0.13 0.09 0.65 r ( uw I uh ) td rotterdam russia r uh td ax r aa tf er d ae m r ah sh ax 0.34 0.93 0.95 r aa tf er d ae m r ah sh ax s eh s s ae g ax n aa s ey g ax n aa 0.93 0.64 0.25 0.09 0.97 eh s saginaw 0.49 0.15 0.14 0.97 0.61 0.38 0.85 0.10 0.51 0.42 0.99 0.64 0.18 s ax 1 iy n ax s aa nt iy aa g ow roanoke rock rockville romania s aw | PMM Baseforms r iy ae dd s ( aa r I ae r ) ax y ey Expert Baseform ow ax n ow kd aa kd aa kd v ih 1 ow m ey n iy ax v ow s eh r ax y ey v ow s aa sardinia r ax y ax v ow s eh r d iy n y ax 0.25 0.19 0.47 s aa r d iy n ( y ax I iy ax ) saskatoon saturday saudi savannah scandinavia science scores scotland scottsdale scratch sea s aa r d iy n y ax s aa r d iy n iy ax s ae s k ax t uw n 0.26 0.24 0.99 s ae tf er df ey s aw df iy s ao df iy s ax v ae n ax s k- ih n d ax n ey v iy ax 0.99 0.71 0.25 0.87 0.99 s ay ax n s s s s s s s s k- ao r z k- ao r s k- ao r ax s k- aa td 1 ax n dd k- aa td s d ey 1 k- r ae ch r ch s iy S iy S y uw s iy uw S iy s k- ae n d ax n ey v iy ax saint s ae g ax n ao ow s ey n td 1 0.43 0.38 0.11 0.98 0.83 0.89 0.10 0.45 0.22 0.20 0.11 0.99 0.96 0.99 0.61 0.30 s ay ax n s salina s ax s eh pd t eh m b er season second select senegal s iy z ax n september s eh pd t eh m b er 0.87 s ax pd t eh m b er 0.10 s er v ax s s er v ay s 0.64 0.16 0.11 0.99 0.94 0.89 service s s s s eh ax eh eh k ax n dd 1eh kd td n ax g ao 1 n ax g aa 1 s s er v ax s setting seventeenth seventy s eh tf ih ng s eh v ax n t jy n th s eh v ax n d iy seville s ax v ih 1 shanghai shining shore should showers shut s ax v ih I1 sh ae ng hh ay sh ay n ih ng sh ao r sh uh dd sh ax dd ch uh dd sh aw er z sh ah td sh aw td sh ah td ax s ( ae Iaa ) s k ax t uw n s ae tf er df ey s ( ao I aw ) df iy s ax v ae n ax s k- ao r z s k- aa td 1 ax n dd s k- aa td s d ey I s k- r ae ch santa santo s iy z ax n s eh k ax n dd s ax 1eh kd td s eh n ax g ao 1 s er v ax s s eh tf ih ng s eh v ax n t iy n th s eh v ax ( nt I n d ) iy s ax v ih 0.82 sh sh sh sh 0.87 0.22 0.21 0.15 samoa sa.ndusky S iy 0.86 0.12 0.91 0.99 0.66 0.20 0.11 salt salzburg I 1ay n ax s ae s ao I ax n ax 1ay n ax s s s s s s s 1td Itd s b er gd 1z b er gd m ow ax b ow ax n d ah s k iy n d ah s k iy ao ao ao ax ax ae ax s ae nt ax s ae nt ax s ey nt ax 0.15 say schedules s ae nt ow s ax p ao r ow s ae r ax s ow tf ax s ae s k ae ch ax w aa n s ax s k ae ch ax w aa n s ax s k- ae ch ax w aa n s ae tf ax 1 ay td s ae tf er n s ao s ax I iy tf ow s aw td s ax 1 iy tf ow s ey s k- eh jh ax 1 z score s k- ow 0.38 0.33 0.27 0.93 0.99 0.92 0.99 0.48 0.30 0.12 0.99 0.93 sapporo sarasota saskatchewan satellite saturn sausalito s k- ao r ae ng hh ay ay n ih ng ao r uh dd scotia scotts scranton screwed seal sh aw er z sh ah td s s s s s kkkkk- w ao r ow sh ax aa td s r ae n tq en r uw dd S iy 1 s y ax 1 s iy ax 1 seattle s iy ae tf ax 1 see S iy 0.99 0.99 0.46 s ae g ax n ao s ey n td s ao 1 td s ao I td s b er gd s ax m ow ax s ae n d ah s k iy s ae nt ax s ae nt ow s ( aa I ax ) p ao r ow s ae r ax s ow tf ax s ae s k ae ch ax w aa n 0.38 0.09 0.99 0.99 0.82 0.09 0.90 0.99 s ae tf ax 1 ay td s ae tf er n s ao s ax 1 iy tf ow s ey s k- ehjh ( uh z s k- ao r s s s s s kkkkiy ow sh ax aa td s r ae n tq en r uw dd I s iy ae tf ax 1 S iy 1 ax 1) Word sicily significant PMM Baseforms sh ax td sh ah tf ax s ih s ax 1iy s ax gd n ih f ax kd eh n td s ax gd n ih f ax kd ae n td s ax gd n ih f ax k ax n Weight I Expert Baseform 0.14 0.13 0.99 0.42 '" ' s ih s ax 1 iy s ax gd n ih f ax k ax n td Word Seneca seoul s er b iy ax Weight _[Expert Baseform s eh n ax k ax 0.99 s ow 1 0.73 0.11 s er b iy ax 0.88 s er b iy ae m 0.10 set s eh td 0.98 s eh td seven seventh severe sex shine shit s eh v ax n s eh v ax n th 0.99 0.99 0.94 0.93 0.99 0.51 0.36 0.99 0.94 0.96 0.94 0.99 0.51 0.44 0.99 0.86 0.67 0.14 0.99 0.62 0.37 0.46 0.22 0.18 0.11 0.92 s eh v ax n s eh v ax n th serbia 0.23 0.22 PMM Baseforms ' s eh n ax k ax S ow I s ow ax 1 t ax sioux six sixth ski sky sleet slow smith snow snowed snowing snowstorm snowy sofia some s uw s s uw s n ow dd s n ow ih ng s n ow s t ao r m s n ow iy 0.88 s n ow iy s ow s ow s ow s ow s ah 0.33 0.30 0.19 0.14 0.70 0.29 0.96 0.84 0.62 0.20 0.10 0.99 0.95 0.94 0.95 0.49 0.27 0.20 0.56 0.18 s ow f iy ax s ih kd s s ih kd s th s k- iy s k- ay s 1iy td s 1iy iy td S 1ow s m ih th s m ih td s m eh th s n ow f iy ao f iy ax f iy ax b y ax ax f y ax m s ah m th ih ng s ah m w eh r s aa r iy s ao r iy s aa r ay sounds southeast southern southwest space spanish speak specific speed spell sports spring springs squall sri s uw 0.09 0.96 0.96 0.89 0.94 0.86 0.13 0.99 0.55 0.31 0.12 0.99 0.99 0.99 0.97 s ax m something somewhere sorry 0.84 s s s s s s s s s aw n dd z aw th iy s td ah dh er n aw th w eh s td p- iy s p- ey sh y ax p- ey sh p- ae n ax sh p ae n ax sh s p- ae n ih z s p- ae n ax s s p- iy kd s p- ax s ih f ax kd s p- iy dd s p- eh 1 s p- ao r td s s p- r ih ng s p- r ih ng z s k- w ao 1 s k- w aa s r iy sh r iy s ih kd s s ih kd s th s k- iy s k- ay s 1 iy td S 1 ow s m ih th S n ow s n ow dd s n ow ih ng s n ow s t ao r m 0.47 0.72 0.21 s eh kd s sh ay n sh ih td sh ax t s short show shreveport siberia side singapore situation sixteenth sixty skiing sleep sh r sh ow sh r iy v p ao r td s ay b ih r iy ax s ay dd s ih ng gd ax p ao r s ih ng ax p ao r s ih ch uw ey sh ax n s ih kd s t iy n th s ih kd s t iy s ih kd s t ax s k- iy ih ng s 1 iy pd s 1 iy s ah m s ah m th ih ng s ah m w eh r s ( aa r I ao r ) iy s aw n dd z s aw th iy s td s ah dh er n s aw th w eh s td s p- ey s slovakia s s s s smart smog s m aa r td s m aa gd s m ao gd eh s n ow b ey s snowfall snowmass snowstorms s s s s s so s ow s p- ae n ax sh someone s p- iy kd sometime s p- iy dd s p- eh 1 s p- ao r td s s p- r ih ng s p- r ih ng z s k- w ao 1 n n n n ow ow ow ow b ey s f ao 1 m ae s s t ao r m z n ow ax S t- ao r m z s ax somalia s p- ax s ih f ax kd I ow v aa kd y ax 1 ow v aa kd y uw 1 ax v aa k iy ax 1 ow v aa k y ax snowbase 0.11 0.09 0.99 0.99 0.99 0.99 0.93 0.99 0.99 0.52 s ax v ih r soon sound south southeastern ( s I sh ) r iy southernmost s ax m aa 1 y ax s ax s ah s ah s ah ax s m aa 1 iy ax m w ah n m w aa n m t ay m ah m t ay m s uw n s ax n s aw n dd s aw th s aw th iy s aw th iy s aw th iy s ah dh er s t er n z t er n sh t er n n m ow s td 0.77 0.22 0.53 0.46 0.97 0.99 0.78 0.11 0.86 0.09 0.65 0.28 0.55 s ax v ih r s eh kd s sh ay n sh ih td sh ao r td sh ow sh r iy v p ao r td s ay b ih r iy ax s ay dd s ih ng ax p ao r s ih ch uw ey sh ax n s ih kd s t iy n th s ih kd s t iy s k- iy ih ng s 1 iy pd s I ow v aa k iy ax s m aa r td s m aa gd s n ow b ey s s n ow f ao 1 s n ow m ae S s n ow s t ao r m z s ow s ( ow Iax ) m aa 1 (iy ax | y ax ) s ah m w ( ah Iax) n 0.44 0.66 0.33 0.64 0.27 0.99 0.99 0.40 0.35 0.21 0.99 s ah m t ay m s uw n s aw n dd s aw th s aw th iy s t er n s ah dh er n m ow s td Word stafford stapleton PMM Baseforms s t- ae f er dd s t- ey p ax 1tf ax n s t- ey p ax 1tf s ax n s t- aa r tf ih ng s t- ao r tf ih ng s t- ey td rn eh n td s t- ey td n ax n td s t- ae tf ax s s t- aa kd s t- aa kd v ih 1 s t- ao r m s t- ao r m iy s t- uw p ax dd s t- uw p ih td s t- uw p ih dd s ax b er bd s ah b er bd s ah m ax r iy s ah n s ah n 1 ay td s ah n iy v ey 1 Weight 0.99 0.51 0.33 0.65 0.33 0.52 0.47 0.98 0.94 0.99 0.99 0.99 0.68 0.12 0.09 0.86 0.13 1 0.99 0.96 0.98 Expert Baseform s t- ae f er dd s t- ey p ax 1 tf ax n sh er sh ao r s er f iy ng s er f iy ng gd 0.69 0.21 0.73 0.25 sh ( er I ao r ) sweden sydney syria systems s s s s s w iy dd en tacoma s t 0.93 0.98 0.95 0.77 0.11 0.86 0.99 0.99 0.53 0.46 0.84 0.57 0.14 0.13 0.57 0.11 starting statement status stock stockville storm stormy stupid suburb summary sun sunlight sunnyvale sure surfing tahoe taiwan talk t t t t tallahassee tanzania t t t t tasmania w iy dd en ih dd n iy ih r iy ax ih s t ax rri z s t ax m z ax k ow rmax aa hh ow ay w aa n ao kd aa kd. ae I ax hh ae s iy ae n z ey n iy ax ae n z ax n iy y ax ae n z ax n iy ax ae z m ey n iy ax t t ae z m iy n y ax s t- aa r tf ih ng Word southwestern spain spartanburg special s t- ey td m ax n td specifically s ts ts ts ts ts t- ( ae I ey ) tf ax s aa kd aa kd v ih 1 ao r m ao r m iy uw p ax dd s ah b er bd s ah s ah s ah s ah m ax r iy n n 1 ay td n iy v ey I s er f ih ng s p- ey n 0.96 0.99 square st stamford start state states still stockholm stop storms strasbourg s s s s speeds spokane spot springfield sprinkles stuttgart t t t t suburbs summer sunday sunny supposed k ow m ax hh ow w aa n kd Weight I Expert Baseform s aw th w eh s t er n 0.99 s p- aa r tq en b er gd s p- eh sh ax 1 s p- ey sh eh I s p- ax s ih f ax k I iy s p- ax s ih f ax kd 1 iy s p- iy dd z s p- ow k ae n s p- aa td s p- r ih ng f iy 1 dd s p- r ih ng k ax 1 z s p- r ih ng kd ax 1z s k- w eh r s ey n td s t- ae m f er dd s t- aa r td s t- ey td s t- ey td s s t- ih 1 s t- aa kd hh ow I m s ih dd n iy s ih r iy ax s ih s t ax m z ax aa ay ao PMM Baseforms s aw th w eh s t er n t ae 1 ax hh ae s iy t ae n z ax n iy ax surf sweater t ae z m ey n iy ax switzerland t- aa k ow I m t- aa pd t- ao r m z tr r ae s b er gd. s tr r aa s b er gd sh t uh td g aa r td s t- uh td g aa r td s t- ax td g aa r td sh t uw td g aa r td s ah b er bd z s ah m er s ah n d ey s ah n iy s ax p ow z dd s ax p ax s td s ax p ow s td s er f s w ah tf er s w eh tf er s w ih td s er I ax n dd 0.68 0.22 s p- ey n s p- aa r tq en b er gd s p- eh sh ax 1 0.82 0.17 s p- ax s ih f ax k 1iy 0.99 0.92 0.99 s p- iy dd z 0.81 0.72 0.27 s p- r ih ng f iy I dd s p- r ih ng k ax 1 z 0.99 0.99 s k- w eh r 0.78 0.85 s t- ae m f er dd 0.99 0.98 0.98 0.58 s p- ow k ( ae I ey )n s p- aa td s ey n td s t- aa r td s t- ey td s t- ey td s s t- ih 1 s t- aa ( k I kd hh) ow Im 0.39 0.91 0.98 0.62 s t- aa pd s t- ao r m z s tr r ( aa I ae ) ( z ) b ( er I ao r ) gd s 0.17 0.48 0.18 0.09 0.09 0.99 0.90 0.98 0.99 0.69 sh t uh td g aa r td s ah b er bd z s ah m er s ah n d ey s ah n iy s ax p ow z dd 0.15 0.15 0.91 s er f 0.51 0.48 0.84 s w eh tf er s w ih td ( z | s ) er I ax n dd tehran telephone t ae z m ey n iy y ax t ae z m ey i y ax 0.11 0.11 t t t t t 0.49 0.25 0.24 0.49 0.35 0.14 0.28 0.24 0.13 0.10 0.55 t temecula eh r aa n ey r aa n iy r aa n eh 1 ax f ao n eh I ax f ow n eh 1 ax f ao ax m eh kd y uw I ax ax m eh k uw I ax ax m eh k uw I aa ax m eh kd y uw I aa, eh m p ax ch er z temperatures t t t t t tennessee t eh m p r ax ch er z t eh n ax s iy 0.43 0.75 syracuse s w ih td s ax 1 ax n dd s ih r ax k y uw z 0.14 0.49 system s ih r ax k y uw s s ih s t ax m 0.36 0.96 t t iy 0.97 t eh 1 ax f ow n tahiti t ey hh iy tf iy 0.62 t t t t t t t 0.27 0.98 t eh m ax k uw I ax taipei take talking t eh r aa n t eh m p ( er ax I r ax I ax ) ch er z t eh n ax s iy tampa taos ax ay ey ao aa ae aa hh iy tf iy p ey kd k ih ng k ih ng m p ax ow s t aw s t aw z r ax k y uw ( z | s s ih s t ax m t iy t ax hh iy tf iy 0.95 0.66 t ay p ey t ey kd t ao k ih ng 0.27 0.95 0.38 t ae m p ax t aa ow s 0.30 0.16 Word PMM Baseforms term territories t er m t eh r ax t ao r iy z Weight 0.99 0.99 t er m t eh r ax t ao r iy z texarkana t eh kd s er k ae n ax 0.35 t eh kd s aa r k ae n ax t eh kd s aa r k ae n ax 0.24 t eh kd s ax k ae n ax 0.19 t ay 1 ae n dd 0.64 t ay 1 ax n dd 0.34 th ae ng kd dh ae td ax m 0.99 0.97 0.35 dh ax m dh eh m dh eh r dh ey th ih ng z th er dd th er tf iy ax th dh ih s dh ow z th aw z ax n dd th r iy th r ax aw td th r uw aw td th ah n d er s t- ao r m th er z d ey t ax b eh td t ax b ae td t ay dd z t ay m tf ax tf uw t uw t ow kd y ow t ow I t ax n ay td t ax p iy k ax t ax p iy k ae s 0.23 0.13 0.96 0.95 0.99 0.95 0.95 0.98 0.91 0.99 0.94 0.65 0.34 0.98 0.99 0.76 0.22 0.99 0.99 0.40 0.24 0.23 0.83 0.98 0.99 0.90 0.09 tornados t ao r n ey df ow z 0.93 torrence toulouse t ao r ax n s t ax 1 uw z town t aw n thailand thank that them there they things third thirtieth this those thousand three throughout thunderstorm thursday tibet tides time to tokyo toll tonight topeka Expert Baseform t ay 1 ( ae I ax Word tegucigalpa ) n dd th ae ng kd dh ae td dh eh m tel tell temperature 1p 0.10 0.23 1p 0.19 1p 0.18 1p 0.10 1p 0.09 1p 0.09 0.99 0.99 0.60 think thirteenth thirty thomas thought threatening 0.37 0.76 0.10 0.95 0.99 0.72 0.27 0.97 0.43 0.25 0.21 0.99 0.58 0.18 0.11 0.10 0.69 0.17 0.97 0.90 0.09 0.99 0.99 0.99 0.89 0.99 0.95 tenth terrific test texas than thanks the th ah n d er s t- ao r m th er z d ey t ax b eh td then t ay dd z t ay m tf ( ax uw) Weight t ay ow z t eh g uw s iy g aa ax t ax g uw s ax g ae ao t ey g uw s iy g aa ax t ax g uw s ax g aa aa m t ax g uw s iy g aa aa n t ax g uw s iy g aa ax t eh 1 t eh 1 t eh m p ax ch er t eh m p r ax ch er t eh n ae n t eh n th t ax r ih f ax kd t eh s td t ax s td t eh kd s ax s dh ae n dh en dh ax n th ae ng kd s dh ax dh ah dh iy dh ih dh eh n dh ax n dh iy z th ih ng th th ih ng kd th er t iy n th th er tf iy t aa m ax s th ao td th r eh tq en ih ng ten dh eh r dh ey th ih ng z th er dd th er tf iy ax th dh ih s dh ow z th aw z ax n dd th r iy th r uw aw td PMM Baseforms these thing Expert Baseform t ax g uw s ( aa 1 p ax ly| ax) g t eh 1 t eh 1 t eh m p ( er ax I ax ) ch er r ax t eh n t eh n th t ax r ih f ax kd t eh s td t eh kd s ax s dh ae n th ae ng kd s dh ( ax I ah I ih I iy) dh eh n dh iy z th ih ng through th r uw 0.92 0.99 0.95 t ( ao r I er ) n ey df ow z t ao r ax n s t uw I uw z th ih ng kd th er t iy n th th er tf iy t aa m ax s th ao td th r eh ( tf ax n I tq en I td n ) ih ng th r uw thunder thunderstorms 0.99 0.99 th ah n d er th ah n d er s t- ao r m 0.99 t aw n tianjin th ah n d er th ah n d er s t- ao r m z t ih ae n jh ih n t ow k ( iy y ) ow t ow 1 t ax n ay td t ax p iy k ax z 0.42 t ( iy I y) (eh I aa )n jh ih n traffic travelling trenton trip tropical try tucson tulsa tunisia tuscaloosa twelve twenty tr r ae f ax kd tr r ae v ax I ih ng tr r eh n tq en tr r ih pd tr r aa p ax k ax 1 tr r ay t uw s aa n t ah 1 s ax t ax n iy zh ax 0.94 0.99 0.97 0.99 0.94 0.96 0.94 0.95 0.62 t uw n iy zh ax t ah s k ax I uw s ax t ah s k ax I uh z ax t w eh 1 v ey t w eh 1 v t w eh nt iy 0.28 0.66 0.20 0.80 0.19 0.98 tr r ae f ax kd tr r ae v ax 1 ih ng tr r eh n tq en tr r ih pd tr r aa p ax k ax 1 tr r ay t uw s aa n t ah 1 s ax t ( ax I uw ) n iy zh ax I iy ax ) tide timbuktu times today toledo t ah s k ax 1 uw s ax t w eh 1 v t w eh nt iy tomorrow too tornado toronto t t t t t t t t t ih njh ih n ih ih n jh ih n ay dd ih m b ax kd t uw ih m b ah kd t uw ax m b ax kd t uw ay m z ax df ey ax d ey t ax 1 iy df ow tf ax 1 iy df ow t ax m aa r ow t uw t ao r n ey df ow t ax r aa nt ow 0.38 0.16 0.99 0.39 0.21 0.13 0.98 0.84 0.14 0.87 0.10 0.99 0.95 0.90 0.98 t ay dd t ih m b ax kd t uw t ay m z t ax df ey t ax 1 iy df ow t t t t ax m aa r ow uw ( ao r I er ) n ey df ow ax r aa nt ow Word two typhoons PMM Baseforns Weight Expert Baseform t uw 0.99 t uw t ay f ax 1 ax n z 0.27 t ay f uw n z 0.21 0.16 0.14 0.12 0.96 0.99 uruguay t ay f ao n z t ay f uw m t ay f uw n s ax z t ay f uw n z y uw y uw k r ey n ah n d er s t- ae n dd y uw n ax v er s ah pd ah p er ow p er y ao r ax g w ay use usually y er ax g w ey y uw z y uw zh ax I iy 0.49 0.90 0.44 u ukraine understand universe up upper utica v vail valley vegas venezuela vermont vero vicinity vienna view virgin visibility voice w wait wales walnut wanted warmer warning warsaw y y y y v v v v v v v v v v v v v v v uw zh uw ax 1iy uw zh 1iy uw zh ax 1ay uw tf ax k ax iy y ax ey 1 ae 1iy ey g ax s eh n ax z w ey I ax eh n ax z w eh 1ax er m aa n td ih r ow y ih r ow ax s ih n ax tf iy iy eh n ax y uw er jh ax n ih z ax b ih I ax tf iy 0.81 0.99 0.81 0.88 0.11 0.50 0.22 0.16 0.15 0.99 0.86 0.12 0.89 0.97 0.96 0.53 0.42 0.97 0.69 0.29 0.97 y uw y uw k r ey n ah n d er s t- ae n dd y uw n ax v er s ah pd ah p er y uw ( s Iz ) y uw zh ( uw ax I | ax I w ax 1 ) iy 1 y uw tf ax k ax v iy v ey v ae v ey v eh 1 1iy g ax s n ax z w ey I ax united until update urbana us v er m aa n td v ( ih I eh ) r ow 0.99 0.59 v oy s d ah b ( ax 11 ax ? ) y ( uw I ax) d ah b y uw w ey td w ah iy w ey w ey 1 w ah tf ax I z w ey I ax s w ey 1 z w ao 1 n ah d w aa n ax td w ao 1 n ax w ao 1 m ax td w ao 1 n ax td w aa nt ax dd w ah tf ax dd 0.38 0.73 0.11 w ey td w ao r m er w ao r n ih ng 0.80 w ao r s ao turkey twelfth twentieth twin typhoon typically uganda umbrella v ih z ax b ih 0.10 0.29 0.25 0.23 0.21 0.25 0.21 t aw n z tr r ae v ax 1 tr r ae v er s tr r ih n ax d ae dd tr r ih p ax 1w iy tr r oy ch er oy tr r ay ih ng tr r ay t uw z d ey t uw n ax s t uw n ih s y ( er ao r ) ax g w ( ey ay ) 0.99 d ah b ax 1y uw towns travel traverse trinidad tripoli troy tuesday tunis 0.98 v oy 6 PMM Baseforms t ow tf ax 1 trying v ax s ih n ax tf iy v iy eh n ax v y uw v er jh ax n 0.87 0.99 Word total towards 1ax tf iy useful utah uzbekistan vacation vallarta vancouver w ey 1 z velocity w ao 1 n ah td venice vernal 0.17 0.16 0.13 0.83 very victoria w aa nt ax dd vietnam 0.09 0.99 0.71 w ao r m er w ao r n ih ng w ao r s ao vilnius t ao r dd s iy Weight T Expert Baseform 0.96 t ow tf ax I t ( ao r [ ax w ao r I w 0.99 ao r ) dd z t aw n z 0.98 0.96 tr r ae v ax 1 tr r ( ae I ax ) v er s 0.95 tr r ih n ax d ae dd 0.99 tr r ih p ax I iy 0.99 tr r oy 0.82 0.15 tr r ay ih ng 0.83 0.13 0.95 0.54 t uw z d ey t uw n ax s 0.33 ch uw n ih s t er k iy t w eh I f th 0.12 0.99 0.75 t er k iy t w eh 1f th t w eh 1 th t w eh nt iy ax th t w ih n t ay f uw n t ih p ax k ax 1 iy y uw g aa n d ax ax m b r eh Iax ah m bd r eh 1 ax ax m bd r eh 1 ax ah m b r eh I ax y uw n ay tf ax dd ax n t ih 1 ah pd d ey td er b ae n ax ah s 0.16 0.99 0.99 0.99 0.99 0.99 0.36 t w eh nt iy ax th t w ih n t ay f uw n t ih p ax k ( ax 1I ) iy y uw g aa n d ax ( ax I ah ) m b r eh 1 ax ax s 0.38 0.98 uw s f ax 1 y uw t aa y uw z b eh k ax s t ae n v ey k ey sh ax n v ey k y ey sh ax n v v v v v v v v v v v v v v v ay aa r tf ax ay y aa r tf ax ey y aa r tf ax ae n k uw v er ae ng k uw v er ax 1aa s ax tf iy ax 1ao s ax tf iy eh n ax s er n ax 1 er n ow I er t n ao I eh r iy ax kd t ao r iy ax ih kd t ao r iy ax iy eh td n aa m v iy eh td n ae m v ih I ax n ax s v ih n iy ax s 0.21 0.10 0.10 0.99 0.99 0.99 0.99 0.54 y uw n ay tf ax dd ax n t ih 1 ah pd d ey td er b ae n ax ah s 0.97 y uw s f ax 1 y uw t ( ao I aa) uh z b eh k ax s t aa n 0.89 v ey k ey sh ax n 0.78 0.10 0.60 v ay ( aa I eh ) r tf ax 0.25 0.12 0.73 0.24 0.82 0.17 0.88 0.69 v ae n k uw v er v ax 1 aa s ax tf iy v eh n ( ax I iy ) s v er n ax 1 0.19 0.10 0.99 0.86 v eh r iy v ax kd t ao r iy ax 0.10 0.76 0.17 0.48 0.44 v iy ( ax | eh ) td n ( ae l aa) m v ih 1 n iy ax s Word PMM Baseforms [ Expert Baseform w ah z w uh z Weight 0.59 0.36 watch w aa ch 0.93 w aa ch water w aa tf er 0.63 w ( aa df) er waterloo w ao tf er w aa df er w aa tf er 1 uw 0.19 0.12 0.78 waterville w ao tf er 1 uw w aa df er v ih 1 0.21 0.78 wellington w w w w w w w w w w w w w aa tf er v ih 1 ey v z ey n eh r ih r er eh dh er iy kd iy kd 1 iy eh 1 k ax m ax 1 k ax m ax 1 k ow m ax k ah m eh 1ih ng t ax n 0.21 0.98 0.91 0.89 0.10 0.99 0.97 0.99 0.50 0.18 0.15 0.10 0.88 west westlake v eh 1 ih ng t ax n w eh s td w eh s I ey kd 0.11 0.99 0.58 w eh s td 1 ey kd w ah td w eh n w eh dh er w ay td hh uw w ay w ay 1dd w uh dd w aa 1 dd w uh dd w ih 1 y ax m z b er gd w ih 1 iy ax m z b er gd w ax 1m ih ng t ax n w ih 1 m ih ng t ax n w ih n w ih n d iy ax s td w ih n d iy z s td w ih n z er w ax n z er w iy n z er w ih n z w ih n z ao r w ih n s er f ih ng w ih n dd s er f ih ng w ih n ax p ey gd w ih n ax p eh gd w ih n ax p ey kd w ih nt er w ih sh w ih th ih n 0.41 0.99 0.93 0.99 0.99 0.86 0.81 0.77 0.21 0.76 0.12 0.67 0.30 0.98 0.86 0.12 0.36 0.20 0.19 0.10 0.09 0.53 0.46 0.33 0.33 0.20 0.97 0.99 0.38 w ax th ih n w ax th ax n 0.19 0.18 was waves wayne wear weather week weekly welcome what when whether white who why wildwood williamsburg wilmington win windiest windsor windsurfing winnipeg winter wish within ' ' '" w ah z ao uh ) (tf Word virginia vladivostok PMM Baseforms Weight Expert Baseform v er jh ih n y ax v 1 aa df ax v aa s t aa kd v 1 ae df ax v aa s t aa kd v oy jh er 0.99 0.32 v er jh ih n y ax v 1 aa df ax v aa s t aa kd 0.59 v ( oy ax ow y ax J ao y ax | oy) jh er v oy ax jh er w ey k ow w ah ey k ow 0.33 0.70 0.19 w ey k ow w ey k ow ax 1 w ey kd. 0.09 0.99 waterbury w w w w w w w w w w w w w 0.90 0.60 0.16 0.12 0.99 0.99 0.99 0.46 0.41 0.98 0.82 0.09 0.39 watertown w aa tf er b er iy w aa df er b er iy w aa df er t aw n 0.21 0.09 0.35 w ah tf er t aw n w aa tf er t aw n w ao tf er t aw n w ey v w ey w ey ey w ah oy w ah ey w iy w eh r ih ng w ih r ey ih ng er ih ng w eh n z d ey w iy kd eh n dd w iy kd s w eh 1 w ax 1 w er w eh r w eh s t er n w eh td w iy 1 ih ng w eh r w ih ch w ay td hh ao r s w ay td hh ao r s s hh ow l w ih ch ax t ao w ih ch ax t aa 0.22 0.15 0.09 0.92 0.52 0.20 0.16 0.09 0.97 0.68 0.10 0.10 0.93 0.97 0.96 0.66 0.29 0.70 0.22 0.94 0.99 0.99 0.99 0.95 0.66 0.12 0.99 0.59 0.28 w ax 1 w ih 1 y ax m z p ao r td 0.92 0.72 voyager waco w ( aa ao uh df) er 1 uw ) w ( aa ao uh) df) er v ih I (tf (tf walla want w ey v z w ey n w eh r w w w w wake warm warmest warnings warwick eh dh er iy kd iy kd 1 iy eh 1 k ax m washington watches w eh 1 ih ng t ax n w eh s td w eh s ( td 1 | 1 ) ey kd w ah td w eh n w eh dh er w ay td hh uw w ay w ay I dd w uh dd w ih 1y ax wave way m z b er gd we wearing w ih 1 m ih ng t ax n wednesday weekend weeks well w ih n w ih n d iy ax s td w ih n z er were w ih n dd s er f ih ng w ih n ax p eh gd w w w ih ih nt er ih sh ( ax I ih) (th n western wet wheeling where which whitehorse whole wichita dh) will williamsport ' aa aa ah aa ao ao ao ao ao aa aa aa aa I ax n td n rm r m ax s td r n ih ng z r w eh kd r w ih kd sh ih ng t ax n ch ax z ch ax z s tf er b eh r iy 0.29 w ey kd w aa I ax w aa n td w w w w ao ao ao ao r r r r m m ax s td n ih ng z w ax kd w aa sh ih ng t ax n w aa ch ax z w (aa I ao I uh df) er b er iy ) ( tf w (aa ao I uh df) er t aw n ) ( tf w ey v w ey w iy w eh r ih ng w w w w eh n z d ey iy k eh n dd iy kd s eh I w er w w w w w w eh s t er n eh td iy 1 ih ng eh r ih ch ay td hh ao r s hh ow l w ih ch ax t ao w ( ax I ih ) I w ih 1 y ax m z p ao r td Word PMM Baseforms w ax td ih p Weight_[ Expert Baseform 0.11 wonderful w ah n d er f ax 1 w aa n d er f ax 1 w ah dd I ax n dd 0.55 woodland worcester world would wrong xian yakima year years yellowstone yesterday york yosemite your yukon zaire zealand zimbabwe w uh dd 1 ax n dd w uh s t er w ah s t er w ao r I dd w er Idd w uh dd r ao ng r aa ng sh iy aa n ch y aa n sh ae n sh iy ae n y ae k ax m aa y ih r y ih r z y ih r ax z y er z y eh 1ow s t- ow n y eh s t er d ey y ao r kd y ow s eh m ax tf iy y ow s ax m ax tf iy y er y ao r y uw k aa n z ay ih r z ay y er z ay iy er z ay er z iy I ax n dd z ax m b aa bd w ey z ax m b aa bd w eh ey z ih m b aa bd w ey Word PMM Baseforms w ih I iy ax m z p ao r td w ih m b ax 1df ax n w ih m b ax 1df en w ih n dd Weight 0.18 Expert Baseform 0.85 0.12 w ih m b ax 1 df ax n 0.98 w ih n dd winds windsurf windy winston w ih n dd z w ih n dd s er f w ih n d iy 0.93 0.99 0.99 0.61 0.30 w w w w wisconsin w ax s k- aa n s ax n w ax s k aa n s ax n w ih th w ax th w ih dh w ah n d er w ah n d er ax w ah n d er ih ng w uh dd z w ao r kd w er kd w ao r th w er th w ah aw w aa ow w ah ow w ay ow m ih ng I ay w ay y eh y ae y ey ax y ih r 1iy y eh 1ow n ay f y eh s y ow k ax hh aa m ax y ow k ow hh aa m ax y ow k ow hh aa m aa y ao r kd t av n y uv y uw g ow s 1aa v iy ax y uw m ay y uw m aa iy y uw m ax y uw m ax z ae m b iy ax z iy ae m b iy ax aa m b iy ax z ih r ow z iy ow s ih r ow z er ax kd z ao r ax kd 0.74 w ax s k aa n s ax n w ah n d er f ax I wimbledon w uh dd 1 ( ax I ae ) n dd wind 0.44 0.51 0.42 0.66 0.30 w uh s t er 0.59 w er 1 dd 0.24 0.98 0.78 w ( uh I ax ) dd r ao ng 0.19 0.45 0.18 with sh ( ax y I iy ) aa n 0.11 0.09 0.82 0.96 0.43 wonder y ae k ax m ax y ih r y ih r z 0.38 0.15 0.91 0.95 0.99 0.52 0.19 0.79 0.19 0.92 0.31 0.29 0.23 0.11 0.97 0.52 0.31 0.11 wondering woods work worth y y y y eh I ow s t- ow n eh s t er d ey ao r kd ow s eh m ax tf iy ( y uw r I y ao r y er) y uw k aa n z ay ( ih r I eh r z iy I ax n dd z ax m b aa bd w ey wow wyoming y yeah yearly yellowknife yes yokohama yorktown you yugoslavia yuma zambia zero zurich w ih n s t- ax n w ih n s t ax n ih ih ih ih n n n n dd z dd s er f d iy s t- ax n 0.23 0.38 0.24 w ( ax I ih ) (th I dh) 0.23 0.72 0.27 w ah n d er 0.96 0.99 0.65 0.29 0.54 0.45 0.32 0.23 w ah n d er ih ng w uh dd z w er kd w er th w aw 0.21 0.95 0.68 0.31 0.56 0.20 w ay ow m ih ng w ay y ( ae I eh I ey ax) 0.10 0.99 0.99 0.99 0.48 0.24 0.19 0.99 0.98 0.96 0.33 0.33 0.16 0.16 0.68 0.17 0.14 0.35 0.33 0.30 0.66 0.15 y y y y ih r Iiy eh I ow n ay f eh s ow k ax hh aa m ax y y y y ao r kd t aw n ( uw I ax ) uw g ow s 1aa v iy ax uw m ax z ae m b iy ax z ih r ow z ( er Iao r ) ax kd Bibliography [1] 6.345 Automatic Speech Recognition, MIT. available from http://ocw.mit.edu/courses/electrical-engineering-and-computer-science/ 6-345-automatic-speech-recognition-spring-2003/. [2] Phonebook, NYNEX isolated words. available from http: //www. ldc upenn. edu. [3] PronLex, COMLEX English http://www.ldc.upenn.edu. pronunciation dictionary. available from [4] L. R. Bahl, S. Das, P. V. deSouza, M. Epstein, R. L. Mercer, B. Merialdo, D. Nahamoo, M. A. Picheny, and J. Powell. Automatic phonetic baseform determination. In Proc. ICASSP, Washington, DC, USA, 1991. [5] J.R Bellegarda. Unsupervised, language-independant grapheme-to-phoneme conversion by latent analogy. In Speech Comm, 2005. [6] Maximilian Bisani and Hermann Ney. Open vocabulary speech recognition with flat hybrid models. In Proc. Europeen Conf. on Speech Communication and Technology, 2003. [7] Maximilian Bisani and Hermann Ney. Joint-sequence models for grapheme-to-phoneme conversion. Speech Commun., 50(5):434-451, 2008. [8] Hung-An Chang and James R. Glass. A back-off discriminative acoustic model for automatic speech recognition. In Proc. INTERSPEECH, 2009. [9] Stanley F. Chen. Conditional and joint models for grapheme-to-phoneme conversion. In Proc. European Conf. on Speech Communication and Technology, 2003. [10] Ghinwa Choueiter. Linguistically-Motivated Sub-word Modeling with Applications to Speech Recognition. PhD thesis, MIT, 2009. [11] Ghinwa F. Choueiter, Mesrob I. Ohannessian, Stephanie Seneff, and James R. Glass. A turbo-style algorithm for lexical baseforms estimation. In Proc. ICASSP, 2008. [12] Michael Collins. Discriminative training methods for hidden markov models: Theory and experiments with perceptron algorithms. In Proc. EMNLP, 2002. [13] V. Zue et al. Jupiter: A telephone-based conversational interface for weather information. IEEE Transactions on Speech and Audio Processing, 8(1):85-96, 2000. [14] K. Macherey F. Wessel, R. Schluter and H. Ney. Confidence measures for large vocabulary continuous speech recognition. IEEE transactions on Speech and Audio Processing, 2001. [15] James Glass. A probabilistic framework for segment-based speech recognition. Computer Speech and Language, 17:137-152, 2003. [16] A. Gruenstein, I. McGraw, and I. Badr. The wami tollkit for developing, deploying, and evaluating web-accessible multimodal interfaces. In Proc. ICMI, 2008. [17] J. Hakkinen, J. Suontausta, S. Riis, and K.J Jensen. Assessing test-to-phoneme mapping strategies in speaker independant isolated word recognition. In Speech Comm, 2003. [18] Timothy Hazen, Lee Hetherington, Han Shu, and Karen Livescu. Pronunciation modeling using a finite-state transducer representation. Speech Communication, 46(2):189-201, 2005. [19] I. Lee Hetherington. An efficient implementation of phonological rules using finite-state transducers. In Proc. EuroSpeech, 2001. [20] Lee Hetherington. The mit finite-state transducer toolkit for speech and language processing. In Proc. ICSLP, 2004. [21] Sittichai Jiampojamarn and Grzegorz Kondrak. Online discriminative training for grapheme-to-phoneme conversion. In In Proc. INTERSPEECH, 2009. [22] R.M Kaplan and M. Kay. put.Linguist, 1994. Regular models of phonological rule systems. In Com- [23] Antoine Laurent, Paul Deleglise, and Sylvain Meigner. Grapheme to phoneme conversion using an smt system. In In Proc. INTERSPEECH,2009. [24] Xiao Li, Asela Guanawardana, and Alex Acero. Adapting grapheme-to-phoneme conversion for name recognition,. In Proc. ASRU, 2007. [25] et al. M. Riley. Stochastic pronunciation modelling form hand-labelled phonetic corpora, 1999. [26] D. MacAllester, L. Gillick, F. Scattone, and M. Newman. Fabricating conversational speech data with acoustic models: A program to examine model-data mismatch. In Proc. Spoken Language Processing, 1998. [27] Benoit Maison. Automatic baseform generation from acoustic data. In Proc. Eurospeech 2003, 2003. [28] Y. Marchand and R.I Damper. Multistrategy appraoch to improving pronunciation by analogy. In Comput. Linguist., 2000. [29] Ian McGraw, Chia-ying Lee, Lee Hetherington, Stephanie Seneff, and James R. Glass. Collecting voices from the cloud. In Proc. LREC, 2010. 100 [30] V. Pagel, K. Lenzo, and A.W Black. Letter-to-sound rules for accented lexicon compression. In In Proc International Conference on Spoken Language Processing, 1998. [31] A. Park, T.J Hazen, and J. Glass. Automatic processing of audio lectures for infonation retrieval: vocabulary selection and language modeling. In Proc. ICASSP, 2005. [32] Brian Roark, Murat Saraclar, and Michael Collins. Corrective language modeling for large vocabulary asr with the perce ptron algorithm. In Proc. ICASSP, 2004. [33] Stephanie Seneff. Reversible sound-to-letter/letter-to-sound modeling based on syllable structure. In Proc. HLT-NAACL, 2007. [34] Tilo Sloboda and Alex Waibel. Dictionary learning for spontaneous speech recognition. In Proc. ICSLP, Philadelphia, PA, USA, 1996. [35] Oriol Vinyals, Li Deng, Dong Yu, and Alex Acero. Discriminative pronunciation learning using phoentic decoder and minimim-classification-error criterion. In Proc. ICASSP, 2009. [36] Stanley Wang. Using graphone models in automatic speech recognition. Master's thesis, MIT, 2009. 101