|
{ |
|
"paper_id": "O05-4004", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T07:58:29.346291Z" |
|
}, |
|
"title": "Modeling Pronunciation Variation for Bi-Lingual Mandarin/Taiwanese Speech Recognition", |
|
"authors": [ |
|
{ |
|
"first": "Dau-Cheng", |
|
"middle": [], |
|
"last": "Lyu", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Gung University", |
|
"location": { |
|
"country": "Taiwan" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Ren-Yuan", |
|
"middle": [], |
|
"last": "Lyu", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Gung University", |
|
"location": { |
|
"country": "Taiwan" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Yuang-Chin", |
|
"middle": [], |
|
"last": "Chiang", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Gung University", |
|
"location": { |
|
"country": "Taiwan" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Chun-Nan", |
|
"middle": [], |
|
"last": "Hsu", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Chang Gung University", |
|
"location": { |
|
"country": "Taiwan" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "In this paper, a bilingual large vocaburary speech recognition experiment based on the idea of modeling pronunciation variations is described. The two languages under study are Mandarin Chinese and Taiwanese (Min-nan). These two languages are basically mutually unintelligible, and they have many words with the same Chinese characters and the same meanings, although they are pronounced differently. Observing the bilingual corpus, we found five types of pronunciation variations for Chinese characters. A one-pass, three-layer recognizer was developed that includes a combination of bilingual acoustic models, an integrated pronunciation model, and a tree-structure based searching net. The recognizer's performance was evaluated under three different pronunciation models. The results showed that the character error rate with integrated pronunciation models was better than that with pronunciation models, using either the knowledge-based or the data-driven approach. The relative frequency ratio was also used as a measure to choose the best number of pronunciation variations for each Chinese character. Finally, the best character error rates in Mandarin and Taiwanese testing sets were found to be 16.2% and 15.0%, respectively, when the average number of pronunciations for one Chinese character was 3.9.", |
|
"pdf_parse": { |
|
"paper_id": "O05-4004", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "In this paper, a bilingual large vocaburary speech recognition experiment based on the idea of modeling pronunciation variations is described. The two languages under study are Mandarin Chinese and Taiwanese (Min-nan). These two languages are basically mutually unintelligible, and they have many words with the same Chinese characters and the same meanings, although they are pronounced differently. Observing the bilingual corpus, we found five types of pronunciation variations for Chinese characters. A one-pass, three-layer recognizer was developed that includes a combination of bilingual acoustic models, an integrated pronunciation model, and a tree-structure based searching net. The recognizer's performance was evaluated under three different pronunciation models. The results showed that the character error rate with integrated pronunciation models was better than that with pronunciation models, using either the knowledge-based or the data-driven approach. The relative frequency ratio was also used as a measure to choose the best number of pronunciation variations for each Chinese character. Finally, the best character error rates in Mandarin and Taiwanese testing sets were found to be 16.2% and 15.0%, respectively, when the average number of pronunciations for one Chinese character was 3.9.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Words can be pronounced in more than one ways according to a lexicon; i.e., they usually have multiple pronunciations. Words are also pronounced differently by different people, a phenomenon called \"pronunciation variation.\" Pronunciation variation has been studied in the speech recognition field [Chen 1996; Cremelie 1996] , and reports show that pronunciation variation can cause the performance of automatic speech recognizers to deteriorate if it is not well accounted for. A common approach to solving the pronunciation variation problem is to use pronunciation modeling; where multiple pronunciations are added to each lexeme in a lexicon in order to fit the acoustic data better.", |
|
"cite_spans": [ |
|
{ |
|
"start": 298, |
|
"end": 309, |
|
"text": "[Chen 1996;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 310, |
|
"end": 324, |
|
"text": "Cremelie 1996]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "A Chinese character is pronounced differently in different languages which use that Chinese character in their writing systems. The same character may or may not have the same meaning in such languages. For instance, the Chinese character \"\u7a97\"(window) is pronounced \"chuang 11 \" in Mandarin and \"tang 11 \" in Taiwanese, and these are considered to be multiple pronunciations in a Mandarin/Taiwanese bi-lingual lexicon. \"Chuang 11 \" is often mistakenly pronounced \"cuang 11 \" (the un-retroflex of \"chuang 11 \") by native Taiwanese speakers, who do not have un-retroflex consonants in their language. This is a common cause of pronunciation variation. In the case of English, which has a more complex vowel inventory than the Han language family, the words \"ear\" and \"year\" are difficult for Mandarin speakers to tell apart. In other words, pronunciation variation is a natural and unavoidable phenomenon in a multi-lingual environment.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "In this world of people who are well-connected by various types of communication devices, multi-lingual communication is necessary, and multi-lingual speech recognition is a must. This paper focuses on Mandarin-Taiwanese bi-lingual large vocabulary speech recognition, and the framework studied here is applicable to other language combinations as well.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "Studies on the pronunciation variation problem have focused on two basic approaches, which are based on acoustic modeling or pronunciation modeling. For acoustic modeling, reports [Jurafsky et al. 2001] show that the triphone model can well capture variation resulting from phone substitution or phone reduction; other reports Kam et al. 2003] show that well-trained triphone acoustic models can handle partial change of the pronunciation variation which depends on the context.", |
|
"cite_spans": [ |
|
{ |
|
"start": 180, |
|
"end": 202, |
|
"text": "[Jurafsky et al. 2001]", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 327, |
|
"end": 343, |
|
"text": "Kam et al. 2003]", |
|
"ref_id": "BIBREF12" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "In pronunciation modeling, entries in the pronunciation dictionary include alternative pronunciation variations and associated probabilities, determined through either knowledge-based or data-driven approaches [Kipp et al. 1996; Zeppenfeld et al. 1997; Wiseman et. al. 1998; Wester 2003; Polzin et al. 1998; Peters et al. 1998; Bacchiani et al. 1999; Singh et al. 2002; Kessens 2003; Strik 2003.] . With the knowledge-based approach, variation information is obtained from research reports or pronunciation dictionaries. Techniques for obtaining the probabilities of possible pronunciation variations of a word in the data-driven approach include training decision trees, training an artificial network, using entropy, using the maximum likelihood criterion, and using the calculated phone confusion Mandarin/Taiwanese Speech Recognition matrix [Cremelie and Martens 1998; Riley et al. 1999; Kam et al. 2003; Fukada et al. 1997; Yang et al. 2000; Holter et al. 1999; Torre et al. 1997] . Techniques that achieve higher scores are chosen to serve as pronunciation variation rules.", |
|
"cite_spans": [ |
|
{ |
|
"start": 210, |
|
"end": 228, |
|
"text": "[Kipp et al. 1996;", |
|
"ref_id": "BIBREF16" |
|
}, |
|
{ |
|
"start": 229, |
|
"end": 252, |
|
"text": "Zeppenfeld et al. 1997;", |
|
"ref_id": "BIBREF47" |
|
}, |
|
{ |
|
"start": 253, |
|
"end": 274, |
|
"text": "Wiseman et. al. 1998;", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 275, |
|
"end": 287, |
|
"text": "Wester 2003;", |
|
"ref_id": "BIBREF44" |
|
}, |
|
{ |
|
"start": 288, |
|
"end": 307, |
|
"text": "Polzin et al. 1998;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 308, |
|
"end": 327, |
|
"text": "Peters et al. 1998;", |
|
"ref_id": "BIBREF34" |
|
}, |
|
{ |
|
"start": 328, |
|
"end": 350, |
|
"text": "Bacchiani et al. 1999;", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 351, |
|
"end": 369, |
|
"text": "Singh et al. 2002;", |
|
"ref_id": "BIBREF38" |
|
}, |
|
{ |
|
"start": 370, |
|
"end": 383, |
|
"text": "Kessens 2003;", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 384, |
|
"end": 396, |
|
"text": "Strik 2003.]", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 845, |
|
"end": 872, |
|
"text": "[Cremelie and Martens 1998;", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 873, |
|
"end": 891, |
|
"text": "Riley et al. 1999;", |
|
"ref_id": "BIBREF37" |
|
}, |
|
{ |
|
"start": 892, |
|
"end": 908, |
|
"text": "Kam et al. 2003;", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 909, |
|
"end": 928, |
|
"text": "Fukada et al. 1997;", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 929, |
|
"end": 946, |
|
"text": "Yang et al. 2000;", |
|
"ref_id": "BIBREF46" |
|
}, |
|
{ |
|
"start": 947, |
|
"end": 966, |
|
"text": "Holter et al. 1999;", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 967, |
|
"end": 985, |
|
"text": "Torre et al. 1997]", |
|
"ref_id": "BIBREF42" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "In addition to the pronunciation variation within a word, substantial variation occurs across word boundaries [Finke et al. 1997; Fukada et al. 1998; Kessens et al. 1999.] . Due to the mono-syllabic nature of Mandarin and Taiwanese, pronunciation variation is complex, and we can identify five types of variation: (1) one orthography with pronunciation variation; (2) colloquial/literate switching; (3) tone sandhi; (4) one orthography with multiple pronunciations; (5) one pronunciation with multiple orthography. The first three types of variation occur in mono-lingual environment, while the last two occur in bi-lingual environments. Details will be given in Section 3.", |
|
"cite_spans": [ |
|
{ |
|
"start": 110, |
|
"end": 129, |
|
"text": "[Finke et al. 1997;", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 130, |
|
"end": 149, |
|
"text": "Fukada et al. 1998;", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 150, |
|
"end": 171, |
|
"text": "Kessens et al. 1999.]", |
|
"ref_id": "BIBREF15" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "The goal of this study was to construct a Mandarin/Taiwanese bi-lingual large vocabulary speech recognizer. We implemented a one-pass recognizer based on a bi-lingual acoustic model, an integrated pronunciation model, and a word searching net with tree-structured nodes. Most of the state-of-the-art speech recognizers, for either Western or Oriental languages, are implemented with the one-pass search strategy [Odell 1994; Aubert 1999; Hagen 2001] . In the acoustic modeling, one phonemic inventory called ForPA (Formosa Phonetic Alphabet) is used to transcribe bi-lingual corpora. According to this inventory, the acoustic models for similar sounds across languages are shared. In addition, we use an algorithm based on a decision tree to cluster similar acoustic models by means of the maximum likelihood criterion. In the pronunciation modeling, we integrate knowledge-based and data-driven approaches. If only the knowledge-based approach is adopted, some variation in the speech corpus can not be covered at all, while if only the data-driven approach is employed, the variation for each new corpus has to be determined. However, the more variations for each word there are in the searching net, the more the recognition time and confusability will increase. To limit the number of pronunciation variations for each Chinese character, we adopt a score based on the relative frequency ratio and choose the best average number of pronunciation variations. Furthermore, the tree-structured net directly uses each Chinese character as a searching node, which is also a new trial in the ASR field of Chinese languages.", |
|
"cite_spans": [ |
|
{ |
|
"start": 412, |
|
"end": 424, |
|
"text": "[Odell 1994;", |
|
"ref_id": "BIBREF33" |
|
}, |
|
{ |
|
"start": 425, |
|
"end": 437, |
|
"text": "Aubert 1999;", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 438, |
|
"end": 449, |
|
"text": "Hagen 2001]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "This paper is organized as follows. Section 2 states the problem. Section 3 represents the proposed framework, which includes acoustic modeling, pronunciation modeling, and a searching net. In section 4, we report experimental results and analyze three different pronunciation models using a bi-lingual testing set. The final section is a summary.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "In recent decades, most of the speech recognition research related to the Chinese (also called the \"\u6f22\" Han) language family has focused on Mandarin speech [Lee 1998; Liao et al. 2000] . Relatively few studies have focused on other languages [Lyu et al. 2000; Gau et al. 2000] . In this paper, we consider two languages in this language family, i.e., Mandarin and Taiwanese, simultaneously within the same framework of speech recognition. In Taiwan, Mandarin Chinese is the official language, and Taiwanese is the mother tongue of about three quarters of the population. Quite a few people speak Mandarin with an accent that is strongly influenced by Taiwanese, and when they speak Taiwanese, they mix in words from Mandarin. It appears that people in Southern China do much the same. If successful, we expect that this framework will work well for other combinations of Chinese languages.", |
|
"cite_spans": [ |
|
{ |
|
"start": 155, |
|
"end": 165, |
|
"text": "[Lee 1998;", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 166, |
|
"end": 183, |
|
"text": "Liao et al. 2000]", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 241, |
|
"end": 258, |
|
"text": "[Lyu et al. 2000;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 259, |
|
"end": 275, |
|
"text": "Gau et al. 2000]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Problem Statements", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "In the Mandarin Chinese speech recognition system, a typical syllable decoder is implemented by searching a 3-layer network consisting of an acoustic model layer, a lexical layer, and a grammar layer, as shown in Figure 1 . After the optimal syllable sequence or the syllable lattice is determined by the decoder, a syllable-to-character converter is applied to handle the homonym issue for the final text output, as shown in Figure 2 . This framework works well and has long been used by the speech communication community. To generalize the system so as to incorporate more than one language, a straightforward approach is to extend the system with more acoustic models, more entries in the pronunciation dictionary, and more paths in the searching net. However, this will lead to the following difficulties: Mandarin/Taiwanese Speech Recognition character in a word for a recognition network. Doing so will not only unnecessarily enlarge the searching space but also increase the time spent on decoding.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 213, |
|
"end": 221, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
}, |
|
{ |
|
"start": 426, |
|
"end": 434, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Problem Statements", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "Generating multiple pronunciation lexicons efficiently is not a trivial task.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "2.", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "3. The language model for mixed languages is hard to estimate.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "2.", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "When new acoustic features like tones are added to the system, all 3 layers in syllable decoding and in the syllable-to-character converter should be modified. This also is not a trivial task.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "4.", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Unlike some conventional approaches, which divide the recognition task into syllable decoding and character decoding, our proposed approach adopts a one-stage searching strategy, as shown in Figure 3 , which decodes the acoustic feature sequence X directly to obtain the desired character sequence C*, no matter what languages are spoken. The decoding equation is, thus, as follows:", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 191, |
|
"end": 199, |
|
"text": "Figure 3", |
|
"ref_id": "FIGREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Our Approach", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "* ( ) arg max ( | ). C C X P C X = (1)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Our Approach", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "In this framework, character decoding can be implemented by searching in a three-layer network composed of an acoustic model layer, a lexical layer, and a grammar layer, as shown in Figure 4 . There are at least 2 critical differences between our framework and the conventional one. 1) In the lexicon layer, character-to-pronunciation mapping can easily incorporate multiple pronunciations caused by multiple languages, including Japanese, Korean, and even Vietnamese, which also use Chinese characters. 2) In the grammar layer, characters instead of syllables are used as nodes in the searching net. Under this ASR structure, we do not care which language the user speaks. No matter whether the language is Taiwanese, Mandarin or a mixture of them in one sentence, the ASR outputs the Chinese character only. This makes it language independent! As in other multi-lingual researches [Young et al. 1997; Waibel 2000] , determining how to efficiently and easily combine two languages in the acoustic and pronunciation models is very important. In the following two subsections, we will describe various approaches to integrating these two models in order to improve the recognition performance of ASR systems. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 883, |
|
"end": 902, |
|
"text": "[Young et al. 1997;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 903, |
|
"end": 915, |
|
"text": "Waibel 2000]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 182, |
|
"end": 190, |
|
"text": "Figure 4", |
|
"ref_id": "FIGREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Our Approach", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "It has been shown that the performance of acoustic models trained by combined speech database from multiple languages is better than that of models trained with speech data from a single language Lyu et al. 2002] . For this reason, we use ForPA, which is an inventory of phoneme symbols, to transcribe the corpus of the two languages discussed here. Table 1 shows the statistical information of the phonemic inventory in different phonetic levels. Sounds in different languages that are transcribed using the same phonemic symbols in ForPA share the same speech material. Combining two languages in this manner reduces the number of syllables by 21%. In order to easily integrate tone information, we used the context-dependent Initial and tonal Final as acoustic units, and trained these models by sharing the data which belonged to the same acoustic unit. Then, a divisive clustering algorithm was used to create context querying decision trees using four question sets, including an Initial set, a tonal Final set, the set of language properties, and a tonal information set. The above clustering approach could achieve significant improvement compared to previous results ].", |
|
"cite_spans": [ |
|
{ |
|
"start": 196, |
|
"end": 212, |
|
"text": "Lyu et al. 2002]", |
|
"ref_id": "BIBREF26" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 350, |
|
"end": 357, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Unified Bi-lingual Acoustic Modeling", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Furthermore, in order to more efficiently merge the similar part of the sound for one phoneme or triphone model in both languages, we used a tying algorithm based on a decision tree to cluster the HMM models by using the maximum likelihood criterion [Liang et al. 1998; Lyu et al. 2002] . For the question sets, we used phonetic knowledge to design a total of 63 questions, including 10 language-dependent questions, 11 common questions, 28 Initial questions, and 14 Final questions. Then, the tree grew and split as we chose the optimal one among all the questions to maximize the increase in the likelihood scores or the decrease in uncertainty. Finally, the convergence condition was set to halt the growth of the decision tree. The acoustic model used in the experiment depended on the different splitting and convergence criteria adopted.", |
|
"cite_spans": [ |
|
{ |
|
"start": 250, |
|
"end": 269, |
|
"text": "[Liang et al. 1998;", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 270, |
|
"end": 286, |
|
"text": "Lyu et al. 2002]", |
|
"ref_id": "BIBREF26" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Mandarin/Taiwanese Speech Recognition", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "The pronunciation model plays an important role in the Chinese character-based ASR engine Huang et al. 2000] . It not only provides more choices during decoding if the speaker exhibits variations in pronunciation but also handles various speaking styles . As mentioned above, one Chinese character has more than two pronunciations in the combined phonetic inventory of Mandarin and Taiwanese. The factors of accent and regional migration can influence the pronunciation or speaking style of speakers too. Therefore, we identify the most common pronunciation variations in Taiwan in Table 2. In Table 2 , we list the five pronunciation variations that the Mandarin-Taiwanese bi-lingual recognizer can handle. Take the Chinese character \"\u8d70\" as an example. It is pronounced as \"zau 51 \" in Taiwanese and means \"to run\" but is pronounced \"zou 21 \" in Mandarin and means \"to walk.\"", |
|
"cite_spans": [ |
|
{ |
|
"start": 90, |
|
"end": 108, |
|
"text": "Huang et al. 2000]", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 582, |
|
"end": 590, |
|
"text": "Table 2.", |
|
"ref_id": "TABREF1" |
|
}, |
|
{ |
|
"start": 594, |
|
"end": 601, |
|
"text": "Table 2", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Pronunciation Modeling", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "On the other hand, the total number of pronunciations in the pronunciation model for the decoding process is also important, because the more pronunciations are included in the lexicon, the more time the decoding process will take, and the less accurate of the ASR results will be [Strik et al. 1999] . The pronunciation variations will generate both improvements and deterioration in the ASR system, so previous research tried to find the optimal method to efficiently control the average pronunciation variations for one word in one language [Kesssens et al. 2003 ]. Our task is harder than that which deals with only one language. The reason is that one Chinese character must be mapped to at least two pronunciation variations, so cross-language confusion increases. In the following sections, we will propose two different methods, knowledge-based and data-driven methods, for obtaining rules of pronunciation variation. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 281, |
|
"end": 300, |
|
"text": "[Strik et al. 1999]", |
|
"ref_id": "BIBREF40" |
|
}, |
|
{ |
|
"start": 544, |
|
"end": 565, |
|
"text": "[Kesssens et al. 2003", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Pronunciation Modeling", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "As in [Wester et al. 2003 ], information about pronunciation can be derived from knowledge sources, such as pronunciation dictionaries hand-crafted by linguistic experts or extracted from the literature. In this approach, a pronunciation variation rule is simply the multiple pronunciations that appear in the lexicon for the same character. Associated probabilities can be calculated as follows. 1) the character-pronunciation pairs are derived; 2) the frequencies of the pairs are counted, and the relative frequency with respect to the total frequency of the", |
|
"cite_spans": [ |
|
{ |
|
"start": 6, |
|
"end": 25, |
|
"text": "[Wester et al. 2003", |
|
"ref_id": "BIBREF44" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Knowledge-Based Method", |
|
"sec_num": "3.2.1" |
|
}, |
|
{ |
|
"text": "same Chinese character is calculated; 3) the pairs with high relative frequencies are kept as multiple pronunciation rules.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Mandarin/Taiwanese Speech Recognition", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "As our Mandarin knowledge source, we adopted the CKIP lexicon (http://ckip.iis.sinica.edu.tw/CKIP/) as our pronunciation lexicon source; it contains about 78,410 words. The length of one word in the lexicon varies from one Chinese character to ten, and the average of the length is 2.4 Chinese characters per word. As our Taiwanese knowledge resource, we adopted the Formosa lexicon (ForLex) [Lyu et al. 2000] , which contains 104,179 words. The average length of one word in it is 2.8 Chinese characters. The pronunciation variation for each Chinese character was assigned a probability, which was estimated based on the frequency count of the pronunciations observed in both lexicons. The number of pronunciation variations for one Chinese character was 1.2 in the CKIP lexicon, and 2.1 in the Formosa lexicon. The number of pronunciation variations for Taiwanese was larger than that for Mandarin. The reasons are that most of the Chinese characters used in Taiwanese carry a classic literature pronunciation and a daily life pronunciation and that Taiwanese has much richer tone sandhi rules. Thus, the average number of pronunciation variations for one Chinese character is increased.", |
|
"cite_spans": [ |
|
{ |
|
"start": 392, |
|
"end": 409, |
|
"text": "[Lyu et al. 2000]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Mandarin/Taiwanese Speech Recognition", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Although the regular pronunciation variations can be obtained from linguistic and phonological information, such as a dictionary, this information is not exhaustive; many phenomena in real speech have not yet been described. Therefore, another approach to deriving pronunciation variations from acoustic clues is presented below. All of the steps are also shown in Figure 5 . First of all, the canonical transcription (T can ) is generated for each Chinese character in the phonetic levels of Initials and tonal Finals. Secondly, for each word in the utterance, a baseline recognition engine based on the Initial/tonal Final acoustic models is used to perform forced recognition, which adopts Viterbi search with an optional phonetic network [Strik 2003 ]. In this way, data-driven transcriptions (T dd ) of all the utterances in the training corpus can be obtained. Then, a dynamic programming algorithm is used to align T can with T dd . With this alignment, we can obtain a confusion table, which consists of pairs of easily confused phonetic units along with their likelihood scores.", |
|
"cite_spans": [ |
|
{ |
|
"start": 742, |
|
"end": 753, |
|
"text": "[Strik 2003", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 365, |
|
"end": 373, |
|
"text": "Figure 5", |
|
"ref_id": "FIGREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Data-Driven Approach", |
|
"sec_num": "3.2.2" |
|
}, |
|
{ |
|
"text": "A partial list of confusing phonetic units is shown as in Table 3 . Using the above approach, we found that the major variation part in a syllable is Initial for both languages, especially in the retroflexion/un-retroflexion set. One of the possible reasons is that retroflex phonetic units exist only in Mandarin and most speakers usually do not accurately pronounce those retroflex units if their mother tongue is Taiwanese. These speakers tend to replace retroflex units with their un-retroflex counter parts.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 58, |
|
"end": 65, |
|
"text": "Table 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Data-Driven Approach", |
|
"sec_num": "3.2.2" |
|
}, |
|
{ |
|
"text": "where T can and T dd represent canonical transcription and data-driven transcription, respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Table 3. Some pronunciation variations obtained with the data-driven approach,", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Mandarin Taiwanese T can T dd T can T dd T can T dd T can T dd zh z s sh gh g p t sh s c ch g d r l ch c n l bh l h t z zh f b k t u3 u4", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Table 3. Some pronunciation variations obtained with the data-driven approach,", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In the searching net, we use a large-vocabulary tree structured word net, because the perplexity can be reduced in the tree-structured searching net compare to the linear searching net. Figure 6 and Figure 7 show examples for a linear searching net and a tree-structured searching net, respectively. There were 5 words as searching paths in the linear net, and the equal probability of each path was set to be 1/5. We used equation 2 to calculate the entropy value based on the number of branches in each path, and we then used equation 3 to calculate the entropy from the perplexity. The perplexity of the linear searching net was found to be 5. This means that the perplexity in the linear searching net equals the number of distinct words. On the other hand, the procedure for determining the perplexity of the tree-structured searching net is described as follows. First, the Chinese characters are aligned according to their locations in multi-character words; characters that are in the same location in each word are considered to be redundant and, thus, eliminated. Finally, the entropy value is also calculated based on the number of branches for each node, using equation (2). In the case shown in Figure 7 , the entropy is 2.29, and the perplexity is 4.89, which is smaller than that of the linear searching net shown in Figure 6 .", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 186, |
|
"end": 194, |
|
"text": "Figure 6", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 199, |
|
"end": 207, |
|
"text": "Figure 7", |
|
"ref_id": "FIGREF5" |
|
}, |
|
{ |
|
"start": 1208, |
|
"end": 1216, |
|
"text": "Figure 7", |
|
"ref_id": "FIGREF5" |
|
}, |
|
{ |
|
"start": 1332, |
|
"end": 1340, |
|
"text": "Figure 6", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Searching Net", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "2 log i i i entropy p p = \u2212\u2211 ,", |
|
"eq_num": "(2)" |
|
} |
|
], |
|
"section": "Searching Net", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "2 entropy perplexity = .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Searching Net", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "(3) Figure 6 . An example of an isolated linear searching net with its probability value. ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 4, |
|
"end": 12, |
|
"text": "Figure 6", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Searching Net", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "All of the experiments employed a bi-lingual corpus, called ForSDa (Formosa Speech Database) . Both the training and testing data were read speech, which was recorded in the 16 kHz/16-bit wave-format in a normal office environment. The training set included a total of 89,164 utterances from 100 speakers, including 50 males and 50 females. Every speaker recorded speech in both languages. The utterances were phonetically balanced words, which were selected from a lexicon of about 40,000 words, using the phonetic abundant algorithm ]. The length of the word varied from 1 to 6. The testing set included 2,000 utterances from 20 speakers; 10 speakers recorded speech in Taiwanese, and the other 10 speakers recorded speech in Mandarin. The statistics of the corpus employed here are listed in Table 4 . ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 795, |
|
"end": 802, |
|
"text": "Table 4", |
|
"ref_id": "TABREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Corpus", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "The experiment setup can be described as follows. Firstly, we used context dependent Initials and tonal Finals with 16 Gaussian mixtures in HMM modeling. The feature vectors used in the HMM included 42 components, with 12 mel-frequency cepstral coefficients (MFCCs), normalized log energy, and pitch with their first and second order derivatives. Secondly, in pronunciation modeling, we used three models, which included knowledge-based, data-driven, and combined approaches, called P KW , P DD and P KW+DD , respectively. The average number of pronunciations for one Chinese character for each pronunciation lexicon was 3.2, 2.7 and 3.9 for P KW , P DD and P KW+DD , respectively. Finally, the tree-structured searching net consisted of 30,000 words, and the word perplexity of the net was 15,249. This means that there where almost 15,249 candidates for each input speech utterance in the decoding phase. Additionally, the output of the recognizer was Chinese characters; therefore, we evaluated the performance based on the Chinese character error rate (CER). Table 5 shows the CER results for pronunciation modeling with the Taiwanese and Mandarin testing sets. We can draw two conclusions; firstly, when the pronunciation model P KW+DD was used, the CER was minimal for both languages. The reason is that P KW+DD could capture both within-language and cross-language pronunciation variations. Secondly, the CER of the Test_M set with P DD was better than that with P KW , but the CER of the Test_T set was worse.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 1063, |
|
"end": 1070, |
|
"text": "Table 5", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experimental Setup", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "A possible reason is that most of the pronunciation variations in Taiwanese can be found in the dictionary or lexicon source, such as tone sandhi or colloquial/literate switching. However, in Mandarin, most of the pronunciation variations are due to co-articulation, regional accents, speaking rates, speaking styles, etc. Such types of the variation can only be captured in speech data, not in lexicons. Therefore, the CER of Test_M dropped about 2.2% (17.9%-20.1%) when P DD was used compared to the result obtained with P KW , but the CER of Test_T increased 0.7% (18.3%-17.6%). Table 5 . CER (Character Error Rate) results for three pronunciation models with two testing sets. P KW : pronunciation modeling using the knowledge-based method; P DD : pronunciation modeling using the data-driven approach; P KW+DD : pronunciation modeling using both P KW and P DD . ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 582, |
|
"end": 589, |
|
"text": "Table 5", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experiment Results", |
|
"sec_num": "4.3" |
|
}, |
|
{ |
|
"text": "The addition of pronunciation variants to a lexicon increases the confusability, especially if the lexicon is large. Here, the large increase in confusability was probably the reason why only a small improvement or even deterioration in performance is found. The experimental results represented in Figure 8 show the CER performance as a function of the number of pronunciation variations for each Chinese character. It can be seen that the CER decreased when the average number of pronunciation variations increased. The lowest CER results were obtained when the number of pronunciation averaged 3.9. This was achieved using P KW+DD and by eliminating variants with probabilities smaller than 0.1. Moreover, the error types mentioned above can be classified into the following 3 sets.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 299, |
|
"end": 307, |
|
"text": "Figure 8", |
|
"ref_id": "FIGREF7" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Error Analysis", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "This kind of error is just like the fifth term in Table 1 , and occurs when different Chinese words belonging to different languages have the same or similar pronunciation. Therefore, the confusion of choosing the final Chinese words will occur during the decoding phase. For example, the pronunciation of the Chinese word \"\u661f\u7cfb\" in Mandarin, that is, /sing 55 -si 51 /, is similar with that of \"\u5148\u6b7b\" in Taiwanese, that is, /sing 33 -si 51 /. The same is true of \"\u9ad8\u7b49\" in Mandarin, pronounced /gau 55 -dng 13 /, and \"\u6559\u5802\" in Taiwanese, pronounced, /gau 51 -dng 13 /.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 50, |
|
"end": 57, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "A. Cross-language homophonic confusion", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "This type of error is similar to the first error type, but it only occurs within one language. For example, the Chinese words \"\u7a62\uf91b\" and \"\u6703\uf91b\" have the same pronunciation, that is, /huei 51 -luan 51 /, in Mandarin, and \"\u4ea4\u5f85\" and \"\u4ea4\u4ee3\" both have the same pronunciation in Mandarin, that is, /jiau 55 -dai 51 /, and in Taiwanese, that is, /gau 55 -dai 55 /.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "B. Within-language homophonic confusion", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "This kind of error occurs due to mismatch between the tone pattern and speech features. We add the tone vectors to the feature parameters, the words, \"\u6c34\u9903\" and \"\u7761\u89ba\" can be easily discriminated a tonal phase. However, there is also a side effect if the acoustic model in the tone aspect is not robust enough. A major tone error may be due to confusion between a high-level (55) tone and a mid-level (35) tone. Another major error may due to the confusion between a mid-falling (31) tone and a high-falling tone. Following are some tone confusion examples: 1) \"\u7e2b\u88dc\" /fng 35 -bu 31 / and \"\u8702\u821e\" /fng 55 -u 31 /.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "C. Tone confusion:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "2) \"\u80a1\u7968\" /gu 31 -piau 51 / and \"\u9867\u7968\" /gu 51 -piau 51 /.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "C. Tone confusion:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Most of the performance deterioration observed in this experiment was caused by the above error types; however, the performances of deterioration are smaller than that of improvements by adding pronunciation variations to the lexicon. Therefore, finally, we got an improvement in CER result.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "C. Tone confusion:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "As mentioned in the introduction, the goal of this study was to convert both Taiwanese and Mandarin speech into Chinese characters. In order to deal with the issues of multiple pronunciations and pronunciation variations for each Chinese character in these two languages in the ASR system, we developed a one-pass, three-layer recognizer, which includes combined bi-lingual acoustic models, an integrated pronunciation model and a tree-structure-based searching net. In the pronunciation model, an integrated method is used to combine the knowledge-based and data-driven approaches. Since the knowledge-based approach is used, homophony in Chinese characters can be addressed, and since the data-driven approach is employed, speakers' accents or styles can also be dealt with.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "The experimental results showed that the CER could be improved by using the three different pronunciation models. The best performance was 16.2% and 15.0% for the testing sets Test_M and Test_T, respectively, where the perplexity was 15,249 for 30,000 words, and the P KW+DD pronunciation model was used. In addition, in order to limit the side effect where in the increase in the size of the pronunciation lexicon causes the performance to deteriorate, the average number of pronunciations for both languages was 3.9.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "The method proposed in this paper has been applied to two languages in the Chinese language family, but it can be easily extended to other languages or dialects. We have also discussed the major five pronunciation variations found in Taiwan. This is the first work, to the best of our knowledge, that has systemically investigated pronunciation variations in Mandarin and Taiwanese speech conversion to Chinese characters using ASR technology.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5." |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "One pass cross word decoding for large vocabularies based on a lexical tree search organization", |
|
"authors": [ |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Aubert", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "Proceedings of the European Conference on Speech Communication and Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1559--1562", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Aubert, X., \"One pass cross word decoding for large vocabularies based on a lexical tree search organization,\" In Proceedings of the European Conference on Speech Communication and Technology, 1999, Budapest, Hungary, pp. 1559-1562.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Joint lexicon, acoustic unit inventory and model design", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Bacchiani", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Ostendorf", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "International Journal of Speech Communication", |
|
"volume": "29", |
|
"issue": "2-4", |
|
"pages": "99--114", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bacchiani, M., and M. Ostendorf, \"Joint lexicon, acoustic unit inventory and model design,\" International Journal of Speech Communication, 29(2-4), 1999, pp. 99-114.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "In search of pronunciation rules", |
|
"authors": [ |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Cremelie", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J.-P", |
|
"middle": [], |
|
"last": "Martens", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the European Speech Communication Association (ESCA) Workshop on Modeling Pronunciation Variation for Acoustic Speech Recognition", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "103--108", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Cremelie, N., and J.-P. Martens, \"In search of pronunciation rules,\" In Proceedings of the European Speech Communication Association (ESCA) Workshop on Modeling Pronunciation Variation for Acoustic Speech Recognition, 1998, Rolduc, Kerkrade, pp. 103-108.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Dynamic and static improvements to lexical baseforms", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Downey", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Wiseman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the Workshop on Modeling Pronunciation Variations", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "157--162", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Downey, S., and R. Wiseman, \"Dynamic and static improvements to lexical baseforms,\" In Proceedings of the Workshop on Modeling Pronunciation Variations, 1998, Roldue, pp. 157-162.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Speaking mode dependent pronunciation modeling in large vocabulary conversational speech recognition", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Finke", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Waibel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1997, |
|
"venue": "Proceedings of the European Conference on Speech Communication and Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "2379--2382", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Finke, M., and A. Waibel, \"Speaking mode dependent pronunciation modeling in large vocabulary conversational speech recognition,\" In Proceedings of the European Conference on Speech Communication and Technology, 1997, Rhodos, Greece, pp. 2379-2382.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Automatic generation of a pronunciation dictionary based on a pronunciation network", |
|
"authors": [ |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Fukada", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Sagisaka", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1997, |
|
"venue": "Proceedings of the European Conference on Speech Communication and Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "2471--2474", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Fukada, T., and Y. Sagisaka, \"Automatic generation of a pronunciation dictionary based on a pronunciation network,\" In Proceedings of the European Conference on Speech Communication and Technology, 1997, Rhodos, pp. 2471-2474.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Automatic generation of multiple pronunciations based on neural networks and language statistics", |
|
"authors": [ |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Fukada", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Yoshimura", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Sagisaka", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the European Speech Communication Association (ESCA) Workshop on Modeling Pronunciation Variation for Acoustic Speech Recognition", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "103--108", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Fukada, T., T. Yoshimura, and Y. Sagisaka, \"Automatic generation of multiple pronunciations based on neural networks and language statistics,\" In Proceedings of the European Speech Communication Association (ESCA) Workshop on Modeling Pronunciation Variation for Acoustic Speech Recognition, 1998, Rolduc, Kerkrade, pp. 103-108.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Maximum likelihood modelling of pronunciation variation", |
|
"authors": [ |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Holter", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Svendsen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "International Journal of Speech Communication", |
|
"volume": "29", |
|
"issue": "", |
|
"pages": "177--191", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Holter, T., and T. Svendsen, \"Maximum likelihood modelling of pronunciation variation,\" International Journal of Speech Communication, 29, 1999, pp. 177-191.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Accent Modeling Based on Pronunciation Dictionary Adaptation for Large Vocabulary Mandarin Speech recognition", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Huang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "E", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Zhou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [ |
|
"F" |
|
], |
|
"last": "Lee", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of the International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Huang, C., E. Chang, J.L. Zhou, and K.F. Lee, \"Accent Modeling Based on Pronunciation Dictionary Adaptation for Large Vocabulary Mandarin Speech recognition,\" In Proceedings of the International Conference on Spoken Language Processing, 2000, Beijing.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "What kind of pronunciation variation is hard for triphones to model", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Jurafsky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Ward", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Herold", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Yu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Proceedings of the International Conference on Acoustics, Speech, and Signal Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "577--580", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jurafsky, D., W. Ward, J. Zhang, K. Herold, X. Yu, and S. Zhang, \"What kind of pronunciation variation is hard for triphones to model?\" In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, 2001, Salt Lake City, Utah, pp. 577-580.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Modeling pronunciation variation for Cantonese speech recognition", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Kam", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of ISCA ITR-Workshop on Pronunciation Modeling and Lexicon Adaptation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "12--17", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kam, P., and T. Lee, \"Modeling pronunciation variation for Cantonese speech recognition,\" In Proceedings of ISCA ITR-Workshop on Pronunciation Modeling and Lexicon Adaptation, 2002, Colorado, USA, pp.12-17.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Modeling Cantonese pronunciation variation by acoustic model refinement", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Kam", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "F", |
|
"middle": [], |
|
"last": "Soong", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the 8th European Conference on Speech Communication and Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1477--1480", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kam, P., T. Lee, and F. Soong, \"Modeling Cantonese pronunciation variation by acoustic model refinement,\" In Proceedings of the 8th European Conference on Speech Communication and Technology, 2003, Geneva, Switzerland, pp.1477-1480.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "A data-driven method for modeling pronunciation variation", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Kessens", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Cucchiarini", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Strik", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "International Journal of Speech Communication", |
|
"volume": "40", |
|
"issue": "", |
|
"pages": "517--534", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kessens, J.M., C. Cucchiarini, and H. Strik, \"A data-driven method for modeling pronunciation variation,\" International Journal of Speech Communication, 40, 2003, pp. 517-534.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Modeling pronunciation variation for ASR: Comparing criteria for rule selection", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Kessens", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Strik", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Cucchiarini", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of the Workshop on Pronunciation Modeling and Lexicon Adaptation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "18--23", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kessens, J.M., H. Strik, and C. Cucchiarini, \"Modeling pronunciation variation for ASR: Comparing criteria for rule selection,\" In Proceedings of the Workshop on Pronunciation Modeling and Lexicon Adaptation, 2002, Estes Park, USA, pp. 18-23.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Improving the Performance of a Dutch CSR by Modeling Within-word and Cross-word Pronunciation Variation", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Kessens", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Wester", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Strik", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "International Journal of Speech Communication on Special issue", |
|
"volume": "29", |
|
"issue": "2-4", |
|
"pages": "193--207", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kessens, J.M., M. Wester, and H. Strik, \"Improving the Performance of a Dutch CSR by Modeling Within-word and Cross-word Pronunciation Variation,\" International Journal of Speech Communication on Special issue of 'Modeling Pronunciation Variation for Automatic Speech Recognition', 29(2-4), 1999, pp. 193-207.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "Automatic detection and segmentation of pronunciation variants in German speech corpora", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Kipp", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M.-B", |
|
"middle": [], |
|
"last": "Wesenick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "F", |
|
"middle": [], |
|
"last": "Schiel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1996, |
|
"venue": "Proceedings of the International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "106--109", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kipp, A., M.-B. Wesenick, and F. Schiel, \"Automatic detection and segmentation of pronunciation variants in German speech corpora,\" In Proceedings of the International Conference on Spoken Language Processing, 1996, Philadelphia, USA, pp. 106-109.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Using tone Information In Cantonese Continuous Speech Recognition", |
|
"authors": [ |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Lau", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"W" |
|
], |
|
"last": "Wong", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Ching", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "ACM Transactions on Asian Language Information Processing", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "83--102", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lee, T., W. Lau, Y. W. Wong, and P.C. Ching, \"Using tone Information In Cantonese Continuous Speech Recognition,\" ACM Transactions on Asian Language Information Processing, 1, 2002, pp. 83-102.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "An efficient algorithm to select phonetically balanced scripts for constructing corpus", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Chiang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the International Conference on Natural Language Processing and Knowledge Engineering", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Liang, M.S., R.Y. Lyu, and Y.C. Chiang, \"An efficient algorithm to select phonetically balanced scripts for constructing corpus,\" In Proceedings of the International Conference on Natural Language Processing and Knowledge Engineering, 2003, Beijing, China.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Decision Tree Clustering for Acoustic Modeling in Speaker-Independent Mandarin Telephone Speech Recognition", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Shen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Lee", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the International Symposium on Chinese Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "207--211", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Liang, P.Y. , J. L. Shen, and L. S. Lee, \"Decision Tree Clustering for Acoustic Modeling in Speaker-Independent Mandarin Telephone Speech Recognition,\" In Proceedings of the International Symposium on Chinese Spoken Language Processing , 1998, Singapore, pp. 207-211.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Taiwan Mandarin Benchmark System", |
|
"authors": [], |
|
"year": 2000, |
|
"venue": "Proceedings of the International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "298--301", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Taiwan Mandarin Benchmark System,\" In Proceedings of the International Conference on Spoken Language Processing, 2000, Beijing. pp. 298-301.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Mandarin/Taiwanese Speech Recognition", |
|
"authors": [], |
|
"year": null, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mandarin/Taiwanese Speech Recognition", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "Modeling partial pronuncia-tion variations for spontaneous Mandarin speech recog-nition", |
|
"authors": [ |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Fung", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "International Journal of Computer Speech and Language", |
|
"volume": "17", |
|
"issue": "", |
|
"pages": "357--379", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Liu, Y., and P. Fung, \"Modeling partial pronuncia-tion variations for spontaneous Mandarin speech recog-nition,\" International Journal of Computer Speech and Language, 17, 2003, pp. 357-379.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Partial change accent models for accented Mandarin speech recognition", |
|
"authors": [ |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Fung", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the IEEE Workshop on ASRU", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Liu, Y., and P. Fung, \"Partial change accent models for accented Mandarin speech recognition,\" In Proceedings of the IEEE Workshop on ASRU, 2003, St. Thomas, U.S. Virgin Islands.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "State-Dependent Phonetic Tied Mixtures with Pronunciation Modeling for Spontaneous Speech Recognition", |
|
"authors": [ |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Fung", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "IEEE Transactions on Speech and Audio Processing", |
|
"volume": "12", |
|
"issue": "", |
|
"pages": "351--364", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Liu, Y., and P. Fung, \"State-Dependent Phonetic Tied Mixtures with Pronunciation Modeling for Spontaneous Speech Recognition,\" IEEE Transactions on Speech and Audio Processing, 12, 2004, pp. 351-364.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Speaker Independent Acoustic Modeling for Large Vocabulary Bi-lingual Taiwanese/Mandarin Continuous Speech Recognition", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [ |
|
"H" |
|
], |
|
"last": "Yang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Hsu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of the 9th Australian International Conference on Speech Science & Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lyu, D.C., B.H. Yang, M.S. Liang, R.Y. Lyu, and C.N. Hsu, \"Speaker Independent Acoustic Modeling for Large Vocabulary Bi-lingual Taiwanese/Mandarin Continuous Speech Recognition,\" In Proceedings of the 9th Australian International Conference on Speech Science & Technology, 2002, Melbourne, Australia.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Large Vocabulary Taiwanese (Min-nan) Speech Recognition Using Tone Features and Statistical Pronunciation Modeling", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Chiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Hsu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the 8 th European Conference on Speech Communication and Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lyu, D.C., M.S. Liang, Y.C. Chiang, C.N. Hsu, and R.Y. Lyu, \"Large Vocabulary Taiwanese (Min-nan) Speech Recognition Using Tone Features and Statistical Pronunciation Modeling,\" In Proceedings of the 8 th European Conference on Speech Communication and Technology, 2003, Geneva, Switzerland.", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "Large Vocabulary Taiwanese (Min-nan) Speech Recognition Using Tone Features and Statistical Pronunciation Modeling", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Chiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Hsu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the European Conference on Speech Communication and Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lyu, D.C., M.S. Liang, Y.C. Chiang, C.N. Hsu and R.Y. Lyu, \"Large Vocabulary Taiwanese (Min-nan) Speech Recognition Using Tone Features and Statistical Pronunciation Modeling,\" In Proceedings of the European Conference on Speech Communication and Technology, 2003, Geneva, Switzerland.", |
|
"links": null |
|
}, |
|
"BIBREF29": { |
|
"ref_id": "b29", |
|
"title": "Bi-lingual", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Chiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": null, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lyu, R.Y., C.Y. Chen, Y.C. Chiang, and M.S. Liang, \"Bi-lingual", |
|
"links": null |
|
}, |
|
"BIBREF30": { |
|
"ref_id": "b30", |
|
"title": "Large Vocabulary, Continuous Speech Recognition System Based on the Yong-yong Phonetic Alphabet", |
|
"authors": [], |
|
"year": 2000, |
|
"venue": "Proceedings of the International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Manda-rin/Taiwanese(Min-nan), Large Vocabulary, Continuous Speech Recognition System Based on the Yong-yong Phonetic Alphabet,\" In Proceedings of the International Conference on Spoken Language Processing, 2000, Beijing, China.", |
|
"links": null |
|
}, |
|
"BIBREF31": { |
|
"ref_id": "b31", |
|
"title": "A Unified Framework for Large Vocabulary Speech Recognition of Mutually Unintelligible Chinese \"Regionalects", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"H" |
|
], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Chiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Hsu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the 8th International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lyu, R.Y., D.C. Lyu, M.S. Liang, M.H. Wang, Y.C. Chiang, and C.N. Hsu, \"A Unified Framework for Large Vocabulary Speech Recognition of Mutually Unintelligible Chinese \"Regionalects\",\" In Proceedings of the 8th International Conference on Spoken Language Processing, 2004, Jeju Island, Korea.", |
|
"links": null |
|
}, |
|
"BIBREF32": { |
|
"ref_id": "b32", |
|
"title": "Toward Constructing A Multilingual Speech Corpus for Taiwanese (Minnan), Hakka, and Mandarin", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lyu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Liang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Chiang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "International Journal of Computational Linguistics and Chinese Language Processing", |
|
"volume": "9", |
|
"issue": "2", |
|
"pages": "1--12", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lyu, R.Y., M.S. Liang, and Y.C. Chiang, \"Toward Constructing A Multilingual Speech Corpus for Taiwanese (Minnan), Hakka, and Mandarin,\" International Journal of Computational Linguistics and Chinese Language Processing, 9(2), 2004, pp. 1-12.", |
|
"links": null |
|
}, |
|
"BIBREF33": { |
|
"ref_id": "b33", |
|
"title": "A One Pass Decoder Design for Large Vocabulary Recognition", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Odell", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "V", |
|
"middle": [], |
|
"last": "Valtchev", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Woodland", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Young", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Proceedings of Human Language Technology Workshop", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "405--410", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Odell, J.J., V. Valtchev, P.C.Woodland, and S.J.Young, \"A One Pass Decoder Design for Large Vocabulary Recognition,\" In Proceedings of Human Language Technology Workshop, 1994, pp. 405-410.", |
|
"links": null |
|
}, |
|
"BIBREF34": { |
|
"ref_id": "b34", |
|
"title": "Visualizing speech trajectories", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [ |
|
"D" |
|
], |
|
"last": "Peters", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Stubley", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the European Speech Communication Association (ESCA) Workshop on Modeling Pronunciation Variation for Acoustic Speech Recognition", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "103--108", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Peters, S.D., and P. Stubley, \"Visualizing speech trajectories,\" In Proceedings of the European Speech Communication Association (ESCA) Workshop on Modeling Pronunciation Variation for Acoustic Speech Recognition, 1998, Rolduc, Kerkrade, pp. 103-108.", |
|
"links": null |
|
}, |
|
"BIBREF35": { |
|
"ref_id": "b35", |
|
"title": "Pronunciation variations in emotional speech", |
|
"authors": [ |
|
{ |
|
"first": "T", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Polzin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [ |
|
"H" |
|
], |
|
"last": "Waibel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": null, |
|
"venue": "Proceedings of the European Speech Communication Association (ESCA) Workshop on", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Polzin, T.S., and A.H. Waibel, \"Pronunciation variations in emotional speech,\" In Proceedings of the European Speech Communication Association (ESCA) Workshop on", |
|
"links": null |
|
}, |
|
"BIBREF36": { |
|
"ref_id": "b36", |
|
"title": "Modeling Pronunciation Variation for Acoustic Speech Recognition", |
|
"authors": [], |
|
"year": 1998, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "103--108", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Modeling Pronunciation Variation for Acoustic Speech Recognition, 1998, Rolduc, Kerkrade, pp. 103-108.", |
|
"links": null |
|
}, |
|
"BIBREF37": { |
|
"ref_id": "b37", |
|
"title": "Stochastic pronunciation modelling from hand-labelled phonetic corpora", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Riley", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Byrne", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Finke", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Khudanpur", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Ljolje", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Mcdonough", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Nock", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Saraclar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Wooters", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Zavaliagkos", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "International Journal of Speech Communication", |
|
"volume": "29", |
|
"issue": "", |
|
"pages": "209--224", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Riley, M., W. Byrne, M. Finke, S. Khudanpur, A. Ljolje, J. McDonough, H. Nock, M. Saraclar, C. Wooters, and G. Zavaliagkos, \"Stochastic pronunciation modelling from hand-labelled phonetic corpora,\" International Journal of Speech Communication, 29, 1999, pp. 209-224.", |
|
"links": null |
|
}, |
|
"BIBREF38": { |
|
"ref_id": "b38", |
|
"title": "Automatic generation of subword units for speech recognition systems", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Singh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Raj", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Stern", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "IEEE Transactions on Speech and Audio Processing", |
|
"volume": "10", |
|
"issue": "", |
|
"pages": "89--99", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Singh, R., B. Raj, and R. Stern, \"Automatic generation of subword units for speech recognition systems,\" IEEE Transactions on Speech and Audio Processing, 10, 2002, pp. 89-99.", |
|
"links": null |
|
}, |
|
"BIBREF39": { |
|
"ref_id": "b39", |
|
"title": "A One-pass decoder based on polymorphic linguistic context assignment", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Soltau", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "F", |
|
"middle": [], |
|
"last": "Metze", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Fuegen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Waibel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Proceedings of Automctic Speech Recognition and Understanding Workshop", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Soltau, H., F. Metze, C. Fuegen, and A. Waibel, \"A One-pass decoder based on polymorphic linguistic context assignment,\" In Proceedings of Automctic Speech Recognition and Understanding Workshop, 2001, Trento, Italy.", |
|
"links": null |
|
}, |
|
"BIBREF40": { |
|
"ref_id": "b40", |
|
"title": "Modeling Pronunciation Variation for ASR: Overview and Comparison of Method", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Strik", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Cucchiarini", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "International Journal of Speech Communication", |
|
"volume": "29", |
|
"issue": "", |
|
"pages": "225--246", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Strik, H., and C. Cucchiarini, \"Modeling Pronunciation Variation for ASR: Overview and Comparison of Method,\" International Journal of Speech Communication, 29, 1999, pp. 225-246.", |
|
"links": null |
|
}, |
|
"BIBREF41": { |
|
"ref_id": "b41", |
|
"title": "Modeling Pronunciation Variation for Automatic Speech Recognition", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Strik", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Kessens", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Wester", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the European Speech Communication Association (ESCA) workshop", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "137--144", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Strik H., J.M. Kessens, and M. Wester, \"Modeling Pronunciation Variation for Automatic Speech Recognition,\", In Proceedings of the European Speech Communication Association (ESCA) workshop, 1998, Rolduc, Kerkrade, pp. 137-144.", |
|
"links": null |
|
}, |
|
"BIBREF42": { |
|
"ref_id": "b42", |
|
"title": "Automatic Alternative Transcription Generation and Vocabulary Selection for Flexible Word Recognizers", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Torre", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Villarrubia", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Hernandez", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Elvira", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1997, |
|
"venue": "Proceedings of the International Conference on Acoustics, Speech, and Signal Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1463--1466", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Torre, D., L. Villarrubia, L. Hernandez, and J.M. Elvira, \"Automatic Alternative Transcription Generation and Vocabulary Selection for Flexible Word Recognizers,\" In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, 1997, Munich, pp. 1463-1466.", |
|
"links": null |
|
}, |
|
"BIBREF43": { |
|
"ref_id": "b43", |
|
"title": "A comparison of data-derived and knowlegde-based modeling of pronunciation variation", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Wester", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "E", |
|
"middle": [], |
|
"last": "Fosler-Lussier", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "270--273", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wester, M., and E. Fosler-Lussier, \"A comparison of data-derived and knowlegde-based modeling of pronunciation variation,\" In Proceedings of International Conference on Spoken Language Processing, 2000, Beijing, China, pp. 270-273.", |
|
"links": null |
|
}, |
|
"BIBREF44": { |
|
"ref_id": "b44", |
|
"title": "Pronunciation Modeling for ASR knowledge-based, Data-driven Methods", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Wester", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "International Journal of Computer Speech and Language", |
|
"volume": "88", |
|
"issue": "", |
|
"pages": "69--85", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wester, M., \"Pronunciation Modeling for ASR knowledge-based, Data-driven Methods,\" International Journal of Computer Speech and Language, 88, 2003, pp. 69-85.", |
|
"links": null |
|
}, |
|
"BIBREF45": { |
|
"ref_id": "b45", |
|
"title": "Pronunciation Variation in ASR: Which Variation to model", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Wester", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Kessens", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Strik", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of the International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "488--491", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wester, M., J.M. Kessens, and H. Strik, \"Pronunciation Variation in ASR: Which Variation to model?\" In Proceedings of the International Conference on Spoken Language Processing, 2000, Beijing, China, pp. 488-491.", |
|
"links": null |
|
}, |
|
"BIBREF46": { |
|
"ref_id": "b46", |
|
"title": "Data driven lexical modeling of pronunciation variation in ASR", |
|
"authors": [ |
|
{ |
|
"first": "Q", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J.-P", |
|
"middle": [], |
|
"last": "Martens", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of the International Conference on Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "417--420", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yang, Q., and J.-P. Martens, \"Data driven lexical modeling of pronunciation variation in ASR,\" In Proceedings of the International Conference on Spoken Language Processing, 2000, Beijing, China, pp. 417-420.", |
|
"links": null |
|
}, |
|
"BIBREF47": { |
|
"ref_id": "b47", |
|
"title": "Recognition of conversational speech using the JANUS speech engine", |
|
"authors": [ |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Zeppenfeld", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Finke", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Ries", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Westphal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Waibel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1997, |
|
"venue": "Proceedings of the International Conference on Acoustics, Speech, and Signal Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1815--1818", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zeppenfeld, T., M. Finke, K. Ries, M. Westphal, and A. Waibel, \"Recognition of conversational speech using the JANUS speech engine,\" In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, 1997, Munich, pp. 1815-1818.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "A 3-layer grammar searching net for syllable decoding" |
|
}, |
|
"FIGREF1": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "The syllable-to-character converter 1. In the case of multi-syllabic words such as \"\u570b\u5bb6\" (country), people rarely use Mandarin pronunciation for part of the word and Taiwanese pronunciation for the other part. It is, thus, impractical to generate all instances of all possible bi-lingual pronunciation variations of each" |
|
}, |
|
"FIGREF2": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "One-stage searching strategy for Chinese speech recognition" |
|
}, |
|
"FIGREF3": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "A unified 3-layer framework for multi-language Chinese speech recognition" |
|
}, |
|
"FIGREF4": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "Diagram of pronunciation variations obtained with a data-driven approach." |
|
}, |
|
"FIGREF5": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "An example of an isolated tree-structured searching net with its probability value." |
|
}, |
|
"FIGREF7": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "CER performance for P KW+DD with different numbers of pronunciation variations per Chinese character." |
|
}, |
|
"TABREF1": { |
|
"num": null, |
|
"text": "", |
|
"content": "<table><tr><td/><td/><td colspan=\"6\">levels: 1. one orthography with pronunciation variations</td></tr><tr><td/><td colspan=\"7\">(OOPV); 2. colloquial literate switching (CLS) 3. tone sandhi (TS); 4. one</td></tr><tr><td/><td colspan=\"7\">orthography with multiple pronunciation (OOMP); 5. one pronunciation</td></tr><tr><td/><td colspan=\"7\">with multiple orthographies (OPMO). Other symbols and their meanings</td></tr><tr><td/><td colspan=\"7\">are: Chinese character (CC); Taiwanese or Mandarin pronunciations in</td></tr><tr><td/><td colspan=\"7\">literate style (TPL, MPL); Taiwanese or Mandarin Chinese character in</td></tr><tr><td/><td colspan=\"7\">colloquial style (TCC, MCC). The number [Yuen Ren Chao] following each syllable represents the tone patterns. e.g., zong 51 means the syllable</td></tr><tr><td/><td colspan=\"2\">has a high-falling tone.</td><td/><td/><td/><td/></tr><tr><td/><td/><td colspan=\"3\">Within-language</td><td/><td/></tr><tr><td/><td>(1)</td><td>CC</td><td/><td colspan=\"2\">Base form</td><td/><td>Surface form</td></tr><tr><td/><td>OOPV</td><td>\ufa1d\u5f69</td><td/><td>jing 55 cai 21</td><td/><td/><td>j in 55 cai 21</td></tr><tr><td/><td/><td>\uf934\u5e2b</td><td/><td colspan=\"2\">lau 21 ,shii 55</td><td/><td>lau 21 sii 55</td></tr><tr><td/><td/><td>\u76f4\u7684</td><td/><td>dit 55 e 11</td><td/><td/><td>d i 55 e 11</td></tr><tr><td>(2)</td><td>MCC</td><td>MP</td><td>TCCL</td><td>TPL</td><td/><td>TCCC</td><td>TPC</td></tr><tr><td>CLS</td><td>\u4eca\u5929</td><td>Jin 55 -ten 55</td><td>\u4eca\u5929</td><td>gim 33 -ten 55</td><td/><td>\u4eca\u4ed4\u65e5</td><td>gin 33 -na 55 -lit 55</td></tr><tr><td/><td>\u660e\u5929</td><td>ming 35 -ten 55</td><td>\u660e\u5929</td><td colspan=\"2\">bhing 33 -ten 55</td><td>\u660e\u4ed4\u8f09</td><td>mi 33 -a 55 -zai 11</td></tr><tr><td>(3)</td><td>CC</td><td>MP, isolated</td><td colspan=\"2\">MP, connected</td><td colspan=\"2\">TP, isolated</td><td>TP, connected</td></tr><tr><td/><td/><td colspan=\"3\">Cross-language</td><td/><td/></tr><tr><td/><td>(4)</td><td>CC</td><td/><td>TP</td><td/><td/><td>MP</td></tr><tr><td/><td>OOMP</td><td>\u8d70</td><td/><td>zau 51</td><td/><td/><td>zou 21</td></tr><tr><td/><td/><td>\u96e8</td><td/><td>u 51 ,ho 33</td><td/><td/><td>yu 21</td></tr><tr><td/><td/><td>\ufa08</td><td/><td colspan=\"2\">giann 15 ,hing 15 ,hang 15</td><td/><td>sing 35 ,hang 35</td></tr><tr><td/><td/><td>(5)</td><td colspan=\"2\">Pronunciation</td><td/><td>TCC</td><td>MCC</td></tr><tr><td/><td colspan=\"2\">OPMO</td><td/><td>jia 55 -dan 51</td><td colspan=\"2\">[\u9019\uf9e8]\u7b49</td><td>\u52a0\u86cb</td></tr><tr><td/><td/><td/><td/><td>gau 55 -gai 51</td><td/><td>\u4e5d[\u6b21]</td><td>\u9ad8\u9223</td></tr></table>", |
|
"type_str": "table", |
|
"html": null |
|
}, |
|
"TABREF2": { |
|
"num": null, |
|
"text": "", |
|
"content": "<table><tr><td colspan=\"5\">. Statistics of the bi-lingual speech corpus used for training and testing sets.</td></tr><tr><td colspan=\"3\">M: Mandarin, T: Taiwanese.</td><td/><td/></tr><tr><td/><td>Langue</td><td>No. of</td><td>No. of</td><td>No. of</td></tr><tr><td/><td>ID.</td><td>Speakers</td><td>Words</td><td>Hours</td></tr><tr><td>Training</td><td>M</td><td>100</td><td>43078</td><td>11.3</td></tr><tr><td/><td>T</td><td>100</td><td>46086</td><td>11.2</td></tr><tr><td>Test_M</td><td>M</td><td>10</td><td>1000</td><td>0.28</td></tr><tr><td>Test_T</td><td>T</td><td>10</td><td>1000</td><td>0.28</td></tr></table>", |
|
"type_str": "table", |
|
"html": null |
|
} |
|
} |
|
} |
|
} |