|
{ |
|
"paper_id": "Y09-1005", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T13:43:22.783762Z" |
|
}, |
|
"title": "Developing Speech Recognition and Synthesis Technologies to Support Computer-Aided Pronunciation Training for Chinese Learners of English *", |
|
"authors": [ |
|
{ |
|
"first": "Helen", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "Human-Computer Communications Laboratory The", |
|
"institution": "Chinese University of Hong Kong", |
|
"location": {} |
|
}, |
|
"email": "hmmeng@se.cuhk.edu.hk" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "We describe ongoing research in the development of speech technologies that strives to raise the efficacy of computer-aided pronunciation training, especially for Chinese learners of English. Our approach is grounded on the theory of language transfer and involves a systematic phonological comparison between the primary language (L1 being Chinese) and secondary language (L2 being English) to predict possible segmental and suprasegmental realizations that constitute mispronunciations in L2 English. The predictions are validated based on a specially designed corpus that consists of several hundred hours of L2 English speech. The speech data supports the development of automatic speech recognition technologies that can detect and diagnose mispronunciations. The diagnosis aims to support the design of pedagogical and remedial instructions, which involves text-tospeech synthesis technologies in audiovisual forms.", |
|
"pdf_parse": { |
|
"paper_id": "Y09-1005", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "We describe ongoing research in the development of speech technologies that strives to raise the efficacy of computer-aided pronunciation training, especially for Chinese learners of English. Our approach is grounded on the theory of language transfer and involves a systematic phonological comparison between the primary language (L1 being Chinese) and secondary language (L2 being English) to predict possible segmental and suprasegmental realizations that constitute mispronunciations in L2 English. The predictions are validated based on a specially designed corpus that consists of several hundred hours of L2 English speech. The speech data supports the development of automatic speech recognition technologies that can detect and diagnose mispronunciations. The diagnosis aims to support the design of pedagogical and remedial instructions, which involves text-tospeech synthesis technologies in audiovisual forms.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "English is the lingua franca of our world. It is the second language (L2) most actively studied across Asia, as well as an official language or working language used in education, government, media, etc. in many regions. Hence, acquiring communicative competence in English is of prime importance. It has been estimated that by 2010 there will be 2 billion English learners worldwide, and the proportion in Asia alone will exceed the number of native speakers (Asia Economic News, 2006) . Second language learning, specifically pronunciation learning, involves correct perception and production of sounds in the target language. The learning process tends to be influenced by well established perceptions of sounds and articulatory motions in the primary language (L1). This cross-linguistic influence is often referred as language transfer. Negative transfer of L1 features causes inaccuracies and errors in L2 speech productions, which impede intelligibility. Consequently, the study of L2 speech productions (i.e. the \"interlanguage\" of learners who have not acquired native-like proficiency) is of great interest to phoneticians, linguists, language educators, as well as technologists engaged in the development of CAPT (computer-aided pronunciation training) systems. These applications can complement classroom teaching and provide unique benefits to the learner in terms of accessibility, reduced anxiety and individualized instructions.", |
|
"cite_spans": [ |
|
{ |
|
"start": 460, |
|
"end": 486, |
|
"text": "(Asia Economic News, 2006)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Effective pronunciation training tools need to provide learners with detailed mispronunciation detection, diagnosis and corrective feedback (Ehsani and Knodt, 1998) . Previous work has shown that automatic pronunciation scores at the word-level or sentencelevel correlate highly with human raters but fail to lead to measurable improvement in learner's overall pronunciation (Precoda et al., 2000) . However, locating mispronunciations at the phone-level to learners has been shown to lead to statistically significant improvement for the production of those targeted phones (Neri et al., 2006) . Moreover, diagnostic feedback to learners (e.g. \"you inserted a vowel at the end of the word\") has also been shown to lead to significant improvements in pronunciation training (Kim et al., 2004) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 140, |
|
"end": 164, |
|
"text": "(Ehsani and Knodt, 1998)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 375, |
|
"end": 397, |
|
"text": "(Precoda et al., 2000)", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 575, |
|
"end": 594, |
|
"text": "(Neri et al., 2006)", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 774, |
|
"end": 792, |
|
"text": "(Kim et al., 2004)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Speech recognition systems must be specially designed for computer-assisted pronunciation training (CAPT) in order to support detailed corrective feedback while still obtaining satisfactory performance (Ehsani and Knodt, 1998) . Although large vocabulary continuous speech recognition (LVCSR) systems are widely available in the commercial market, they are not necessarily appropriate for pronunciation training for non-native speakers. Generally, LVCSR systems are designed to accommodate a wide variety of accents and non-standard pronunciations. They are not intended to be used as a tool for discriminating phonetically similar pronunciations of a given word. Free phone recognition, in principle, could support a CAPT tool in providing detailed phone-level feedback to a learner. However, highly accurate free phone recognition is still not possible for native-speech, and is only expected to be even more difficult for the interlanguage of second language learners. Furthermore, non-native mispronunciations may be due to a diversity of factors, such as imperfect understanding of semantics, syntax, morphology, phonology, coarticulatory effects and letter-to-sound rules. As an initial step, we focus on the use of comparative phonological analysis between L1 (Cantonese) and L2 (English) to derive a set of 51 context-dependent phonological rules for mispronunciation prediction. The predicted deviations are validated against observations in the CU-CHLOE (Chinese University Chinese Learners of English) corpus. This consists of Chinese-accented English speech recordings from 100 speakers reading the AESOP fable, \"The North Wind and the Sun\", that we have collected to support our investigation. The phonological rules are incorporated in the automatic speech recognizer as finite-state transducers (FSTs) that form an extended recognition network (ERN) (Harrison et al., 2009) . The ERN enables us to model the target pronunciations and the related non-native mispronunciations in a unified framework. Evaluation based on a disjoint test set shows that the ERN achieves a phone-level recognition accuracy of 73.0% and models 58.4% of the mispronounced words in the test set. Automatic alignment between the recognized and target phone strings with a phonetically-sensitive alignment algorithm enables us to diagnose specific phonetic confusions and generate corrective feedback.", |
|
"cite_spans": [ |
|
{ |
|
"start": 202, |
|
"end": 226, |
|
"text": "(Ehsani and Knodt, 1998)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 1865, |
|
"end": 1888, |
|
"text": "(Harrison et al., 2009)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speech Technologies for CAPT", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Insufficient coverage of the mispronounced words in a disjoint test set is due to sparsity of training data and insensitive acoustic models. To raise the performance of mispronunciation detection, we have also devised a decision fusion methodology that augments the linguistically-motivated approach described above with pronunciation scoring and phone-dependent thresholding (Lo et al., 2008) . Such augmentation brought 30% relative improvement over the linguistically-motivated approach, in terms of phoneme decision errors (i.e. sum of the false acceptance and false rejection errors).", |
|
"cite_spans": [ |
|
{ |
|
"start": 376, |
|
"end": 393, |
|
"text": "(Lo et al., 2008)", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speech Technologies for CAPT", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "We have also begun to develop text-to-speech synthesis technologies that can automatically synthesize emphasis on localized, problematic speech segments, with the aim of enhancing correct feedback to the learner. The approach to synthesis is based on a perturbation model, where a set of acoustic features relating to energy, fundamental frequency and duration are modulated to present contrast between neutral and emphatic speech segments. This constitutes pedagogical and remedial instructions to facilitate perceptual training for the learner. This should, in turn, assist productive training that is supported by automatic speech recognition technologies. The invited talk will also showcase both the automatic speech recognition and text-to-speech synthesis technologies that are being developed under this research effort (Harrison et al., 2008; Meng et al., 2007; Wang et al., 2008a; Wang et al., 2008b) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 828, |
|
"end": 851, |
|
"text": "(Harrison et al., 2008;", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 852, |
|
"end": 870, |
|
"text": "Meng et al., 2007;", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 871, |
|
"end": 890, |
|
"text": "Wang et al., 2008a;", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 891, |
|
"end": 910, |
|
"text": "Wang et al., 2008b)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speech Technologies for CAPT", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "23rd Pacific Asia Conference on Language, Information and Computation, pages 40-42", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Speech technology in computer-aided language learning", |
|
"authors": [ |
|
{ |
|
"first": "F", |
|
"middle": [], |
|
"last": "Ehsani", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "E", |
|
"middle": [], |
|
"last": "Knodt", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Language Learning and Technology", |
|
"volume": "2", |
|
"issue": "1", |
|
"pages": "45--60", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "http://findarticles.com/p/articles/mi_m0WDP/is_2006_Feb_20/ai_n16086425/ Ehsani, F. and E. Knodt. 1998. Speech technology in computer-aided language learning. Language Learning and Technology, vol. 2, no. 1, pp. 45-60.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Implementation of an Extended Recognition Network for Mispronunciation Detection and Diagnosis in Computer-Aided Pronunciation Training", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Harrison", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [ |
|
"K" |
|
], |
|
"last": "Lo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "X", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Qian", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 2nd ISCA Workshop on Speech and Language Technology in Education", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Harrison, A., W.K. Lo, X.J. Qian and H. Meng. 2009. Implementation of an Extended Recognition Network for Mispronunciation Detection and Diagnosis in Computer-Aided Pronunciation Training. In Proceedings of the 2nd ISCA Workshop on Speech and Language Technology in Education, Warrickshire.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Improving Mispronunciation Detection and Diagnosis of Learners' Speech with Context-sensitive Phonological Rules based on Language Transfer", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Harrison", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lau", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of INTERSPEECH", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Harrison, A., W.Y. Lau, H. Meng and L. Wang. 2008. Improving Mispronunciation Detection and Diagnosis of Learners' Speech with Context-sensitive Phonological Rules based on Language Transfer. In Proceedings of INTERSPEECH.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "An interactive English pronunciation dictionary for Korean learners", |
|
"authors": [ |
|
{ |
|
"first": "J.-M", |
|
"middle": [], |
|
"last": "Kim", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Peabody", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Seneff", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of INTERSPEECH", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kim, J.-M., C. Wang, M. Peabody and S. Seneff. 2004. An interactive English pronunciation dictionary for Korean learners. In Proceedings of INTERSPEECH.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Decision Fusion for Improving Mispronunciation Detection using Language Transfer Knowledge and Phoneme-dependent Pronunciation Scoring", |
|
"authors": [ |
|
{ |
|
"first": "W", |
|
"middle": [ |
|
"K" |
|
], |
|
"last": "Lo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Harrison", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of the 6th International Symposium on Chinese Spoken Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lo, W.K., A. Harrison, H. Meng and L. Wang. 2008. Decision Fusion for Improving Mispronunciation Detection using Language Transfer Knowledge and Phoneme-dependent Pronunciation Scoring. In Proceedings of the 6th International Symposium on Chinese Spoken Language Processing, Kuming, China.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Deriving Salient Learners' Mispronunciations from Cross-Language Phonological Comparisons", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Lau", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of the IEEE Automatic Speech Recognition and Understanding Workshop", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Meng, H., Y.Y. Lo, L. Wang and W.Y. Lau. 2007. Deriving Salient Learners' Mispronunciations from Cross-Language Phonological Comparisons. In Proceedings of the IEEE Automatic Speech Recognition and Understanding Workshop.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "ASR-based corrective feedback on pronunciation: does it really work?", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Neri", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Cucchiarini", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Strik", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of INTERSPEECH", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1982--1985", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Neri, A., C. Cucchiarini and H. Strik. 2006. ASR-based corrective feedback on pronunciation: does it really work? In Proceedings of INTERSPEECH, Pittsburg, USA, pp. 1982-1985.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Effects of speech recognition-based pronunciation feedback on second-language pronunciation ability", |
|
"authors": [ |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Precoda", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"A" |
|
], |
|
"last": "Halverson", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Franco", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of InSTILL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Precoda, K., C.A. Halverson and H. Franco. 2000. Effects of speech recognition-based pronunciation feedback on second-language pronunciation ability. In Proceedings of InSTILL.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Automatic Generation and Pruning of Phonetic Mispronunciations to Support Computer-Aided Pronuinciation Training", |
|
"authors": [ |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Feng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of INTERSPEECH", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wang, L., X. Feng and H. Meng. 2008a. Automatic Generation and Pruning of Phonetic Mispronunciations to Support Computer-Aided Pronuinciation Training. In Proceedings of INTERSPEECH.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Mispronunciation Detection Based on Cross-Language Phonological Comparisons", |
|
"authors": [ |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Feng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Meng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of the IEEE IET International Conference on Audio, Language and Image Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wang, L., X. Feng and H. Meng. 2008b. Mispronunciation Detection Based on Cross-Language Phonological Comparisons. In Proceedings of the IEEE IET International Conference on Audio, Language and Image Processing.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": {} |
|
} |
|
} |