{ "paper_id": "P04-1022", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T08:44:51.942506Z" }, "title": "Collocation Translation Acquisition Using Monolingual Corpora", "authors": [ { "first": "Yajuan", "middle": [], "last": "L\u00fc", "suffix": "", "affiliation": {}, "email": "" }, { "first": "Ming", "middle": [], "last": "Zhou", "suffix": "", "affiliation": {}, "email": "mingzhou@microsoft.com" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "Collocation translation is important for machine translation and many other NLP tasks. Unlike previous methods using bilingual parallel corpora, this paper presents a new method for acquiring collocation translations by making use of monolingual corpora and linguistic knowledge. First, dependency triples are extracted from Chinese and English corpora with dependency parsers. Then, a dependency triple translation model is estimated using the EM algorithm based on a dependency correspondence assumption. The generated triple translation model is used to extract collocation translations from two monolingual corpora. Experiments show that our approach outperforms the existing monolingual corpus based methods in dependency triple translation and achieves promising results in collocation translation extraction.", "pdf_parse": { "paper_id": "P04-1022", "_pdf_hash": "", "abstract": [ { "text": "Collocation translation is important for machine translation and many other NLP tasks. Unlike previous methods using bilingual parallel corpora, this paper presents a new method for acquiring collocation translations by making use of monolingual corpora and linguistic knowledge. First, dependency triples are extracted from Chinese and English corpora with dependency parsers. Then, a dependency triple translation model is estimated using the EM algorithm based on a dependency correspondence assumption. The generated triple translation model is used to extract collocation translations from two monolingual corpora. Experiments show that our approach outperforms the existing monolingual corpus based methods in dependency triple translation and achieves promising results in collocation translation extraction.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "A collocation is an arbitrary and recurrent word combination (Benson, 1990) . Previous work in collocation acquisition varies in the kinds of collocations they detect. These range from twoword to multi-word, with or without syntactic structure (Smadja 1993; Lin, 1998; Pearce, 2001; Seretan et al. 2003) . In this paper, a collocation refers to a recurrent word pair linked with a certain syntactic relation. For instance, is a collocation with a syntactic relation verb-object.", "cite_spans": [ { "start": 61, "end": 75, "text": "(Benson, 1990)", "ref_id": "BIBREF0" }, { "start": 244, "end": 257, "text": "(Smadja 1993;", "ref_id": null }, { "start": 258, "end": 268, "text": "Lin, 1998;", "ref_id": null }, { "start": 269, "end": 282, "text": "Pearce, 2001;", "ref_id": null }, { "start": 283, "end": 303, "text": "Seretan et al. 2003)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Translation of collocations is difficult for nonnative speakers. Many collocation translations are idiosyncratic in the sense that they are unpredictable by syntactic or semantic features. Consider Chinese to English translation. The translations of \"\u89e3\u51b3\" can be \"solve\" or \"resolve\". The translations of \"\u95ee\u9898\" can be \"problem\" or \"issue\". However, translations of the collocation \"\u89e3\u51b3 ~ \u95ee\u9898\" as \"solve~problem\" or \"resolve~ issue\" is preferred over \"solve~issue\" or \"resolve ~problem\". Automatically acquiring these collocation translations will be very useful for machine translation, cross language information retrieval, second language learning and many other NLP applications. (Smadja et al., 1996; Gao et al., 2002; Wu and Zhou, 2003) .", "cite_spans": [ { "start": 679, "end": 700, "text": "(Smadja et al., 1996;", "ref_id": null }, { "start": 701, "end": 718, "text": "Gao et al., 2002;", "ref_id": "BIBREF7" }, { "start": 719, "end": 737, "text": "Wu and Zhou, 2003)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Some studies have been done for acquiring collocation translations using parallel corpora (Smadja et al, 1996; Kupiec, 1993; Echizen-ya et al., 2003) . These works implicitly assume that a bilingual corpus on a large scale can be obtained easily. However, despite efforts in compiling parallel corpora, sufficient amounts of such corpora are still unavailable. Instead of heavily relying on bilingual corpora, this paper aims to solve the bottleneck in a different way: to mine bilingual knowledge from structured monolingual corpora, which can be more easily obtained in a large volume.", "cite_spans": [ { "start": 90, "end": 110, "text": "(Smadja et al, 1996;", "ref_id": null }, { "start": 111, "end": 124, "text": "Kupiec, 1993;", "ref_id": null }, { "start": 125, "end": 149, "text": "Echizen-ya et al., 2003)", "ref_id": "BIBREF5" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Our method is based on the observation that despite the great differences between Chinese and English, the main dependency relations tend to have a strong direct correspondence (Zhou et al., 2001 ). Based on this assumption, a new translation model based on dependency triples is proposed. The translation probabilities are estimated from two monolingual corpora using the EM algorithm with the help of a bilingual translation dictionary. Experimental results show that the proposed triple translation model outperforms the other three models in comparison. The obtained triple translation model is also used for collocation translation extraction. Evaluation results demonstrate the effectiveness of our method.", "cite_spans": [ { "start": 177, "end": 195, "text": "(Zhou et al., 2001", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "The remainder of this paper is organized as follows. Section 2 provides a brief description on the related work. Section 3 describes our triple translation model and training algorithm. Section 4 extracts collocation translations from two independent monolingual corpora. Section 5 evaluates the proposed method, and the last section draws conclusions and presents the future work.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "There has been much previous work done on monolingual collocation extraction. They can in general be classified into two types: window-based and syntax-based methods. The former extracts collocations within a fixed window (Church and Hanks 1990; Smadja, 1993) . The latter extracts collocations which have a syntactic relationship (Lin, 1998; Seretan et al., 2003) . The syntax-based method becomes more favorable with recent significant increases in parsing efficiency and accuracy. Several metrics have been adopted to measure the association strength in collocation extraction. Thanopoulos et al. (2002) give comparative evaluations on these metrics.", "cite_spans": [ { "start": 222, "end": 245, "text": "(Church and Hanks 1990;", "ref_id": "BIBREF2" }, { "start": 246, "end": 259, "text": "Smadja, 1993)", "ref_id": null }, { "start": 331, "end": 342, "text": "(Lin, 1998;", "ref_id": null }, { "start": 343, "end": 364, "text": "Seretan et al., 2003)", "ref_id": null }, { "start": 581, "end": 606, "text": "Thanopoulos et al. (2002)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Related work", "sec_num": "2" }, { "text": "Most previous research in translation knowledge acquisition is based on parallel corpora (Brown et al., 1993) . As for collocation translation, Smadja et al. (1996) implement a system to extract collocation translations from a parallel English-French corpus. English collocations are first extracted using the Xtract system, then corresponding French translations are sought based on the Dice coefficient. Echizen-ya et al. (2003) propose a method to extract bilingual collocations using recursive chain-link-type learning. In addition to collocation translation, there is also some related work in acquiring phrase or term translations from parallel corpus (Kupiec, 1993; Yamamoto and Matsumoto 2000) .", "cite_spans": [ { "start": 89, "end": 109, "text": "(Brown et al., 1993)", "ref_id": null }, { "start": 144, "end": 164, "text": "Smadja et al. (1996)", "ref_id": null }, { "start": 406, "end": 430, "text": "Echizen-ya et al. (2003)", "ref_id": "BIBREF5" }, { "start": 658, "end": 672, "text": "(Kupiec, 1993;", "ref_id": null }, { "start": 673, "end": 701, "text": "Yamamoto and Matsumoto 2000)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Related work", "sec_num": "2" }, { "text": "Since large aligned bilingual corpora are hard to obtain, some research has been conducted to exploit translation knowledge from non-parallel corpora. Their work is mainly on word level. Koehn and Knight (2000) presents an approach to estimating word translation probabilities using unrelated monolingual corpora with the EM algorithm. The method exhibits promising results in selecting the right translation among several options provided by bilingual dictionary. Zhou et al.(2001) proposes a method to simulate translation probability with a cross language similarity score, which is estimated from monolingual corpora based on mutual information. The method achieves good results in word translation selection. In addition, (Dagan and Itai, 1994) and (Li, 2002) propose using two monolingual corpora for word sense disambiguation. (Fung, 1998) uses an IR approach to induce new word translations from comparable corpora. (Rapp, 1999) and (Koehn and Knight, 2002) extract new word translations from non-parallel corpus. (Cao and Li, 2002) acquire noun phrase translations by making use of web data. (Wu and Zhou, 2003) also make full use of large scale monolingual corpora and limited bilingual corpora for synonymous collocation extraction.", "cite_spans": [ { "start": 187, "end": 210, "text": "Koehn and Knight (2000)", "ref_id": null }, { "start": 465, "end": 482, "text": "Zhou et al.(2001)", "ref_id": null }, { "start": 727, "end": 749, "text": "(Dagan and Itai, 1994)", "ref_id": "BIBREF3" }, { "start": 754, "end": 764, "text": "(Li, 2002)", "ref_id": "BIBREF1" }, { "start": 834, "end": 846, "text": "(Fung, 1998)", "ref_id": "BIBREF6" }, { "start": 924, "end": 936, "text": "(Rapp, 1999)", "ref_id": null }, { "start": 941, "end": 965, "text": "(Koehn and Knight, 2002)", "ref_id": null }, { "start": 1022, "end": 1040, "text": "(Cao and Li, 2002)", "ref_id": "BIBREF1" }, { "start": 1101, "end": 1120, "text": "(Wu and Zhou, 2003)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Related work", "sec_num": "2" }, { "text": "In this section, we first describe the dependency correspondence assumption underlying our approach. Then a dependency triple translation model and the monolingual corpus based training algorithm are proposed. The obtained triple translation model will be used for collocation translation extraction in next section.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Training a triple translation model from monolingual corpora", "sec_num": "3" }, { "text": "A dependency triple consists of a head, a dependant, and a dependency relation. Using a dependency parser, a sentence can be analyzed into dependency triples. We represent a triple as (w 1 ,r,w 2 ), where w 1 and w 2 are words and r is the dependency relation. It means that w 2 has a dependency relation r with w 1 . For example, a triple (overcome, verb-object, difficulty) means that \"difficulty\" is the object of the verb \"overcome\".", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Dependency correspondence between Chinese and English", "sec_num": "3.1" }, { "text": "Among all the dependency relations, we only consider the following three key types that we think, are the most important in text analysis and machine translation: verb-object (VO), nounadj(AN), and verb-adv(AV).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Dependency correspondence between Chinese and English", "sec_num": "3.1" }, { "text": "It is our observation that there is a strong correspondence in major dependency relations in the translation between English and Chinese. For example, an object-verb relation in Chinese (e.g.(\u514b\u670d, VO, \u56f0\u96be)) is usually translated into the same verb-object relation in English(e.g. (overcome, VO, difficulty)).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Dependency correspondence between Chinese and English", "sec_num": "3.1" }, { "text": "This assumption has been experimentally justified based on a large and balanced bilingual corpus in our previous work (Zhou et al., 2001) . We come to the conclusion that more than 80% of the above dependency relations have a one-one mapping between Chinese and English. We can conclude that there is indeed a very strong correspondence between Chinese and English in the three considered dependency relations. This fact will be used to estimate triple translation model using two monolingual corpora. ", "cite_spans": [ { "start": 118, "end": 137, "text": "(Zhou et al., 2001)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Dependency correspondence between Chinese and English", "sec_num": "3.1" }, { "text": "The language model ) ( tri e p is calculated with English triples database. In order to tackle with the data sparseness problem, we smooth the language model with an interpolation method, as described below.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Language Model", "sec_num": null }, { "text": "When the given English triple occurs in the corpus, we can calculate it as in Equation 2. The wildcard symbol * means it can be any word or relation. With Equations (2) and (3), we get the interpolated language model as shown in (4). ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Language Model", "sec_num": null }, { "text": ") | ( ) | ( ) ( ) 1 ( ) ( ) ( 2 1 e e e tri", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Language Model", "sec_num": null }, { "text": ") ( 1 1 1 tri e freq + \u2212 = \u03bb (5)", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Language Model", "sec_num": null }, { "text": "We simplify the translation model according the following two assumptions. c and 2 c are conditionally independent. We have: Here, we present an approach to estimating these probabilities from two monolingual corpora based on the EM algorithm.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Translation Model", "sec_num": null }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": ") | ( ) , | ( ) , | ( ) | , ,", "eq_num": "( ) |" } ], "section": "Translation Model", "sec_num": null }, { "text": "Chinese and English corpora are first parsed using a dependency parser, and two dependency triple databases are generated. The candidate English translation set of Chinese triples is generated through a bilingual dictionary and the assumption of strong correspondence of dependency relations. There is a risk that unrelated triples in Chinese and English can be connected with this method. However, as the conditions that are used to make the connection are quite strong (i.e. possible word translations in the same triple structure), we believe that this risk, is not very severe. Then, the expectation maximization (EM) algorithm is introduced to iteratively strengthen the correct connections and weaken the incorrect connections.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Estimation of word translation probability using the EM algorithm", "sec_num": "3.3" }, { "text": "According to section 3. . The English language model can be estimated using Equation 4and the translation model can be calculated using Equation 7 ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "EM Algorithm", "sec_num": null }, { "text": "Where e \u0393 represents the translation set of the English word e. Then, the word translation probabilities are estimated iteratively using the EM algorithm. Figure 1 gives a formal description of the EM algorithm. that best explain the Chinese triple database as a translation from the English triple database. In each iteration, the normalized triple translation probabilities are used to update the word translation probabilities. Intuitively, after finding the most probable translation of the Chinese triple, we can collect counts for the word translation it contains. Since the English triple language model provides context information for the disambiguation of the Chinese words, only the appropriate occurrences are counted. Now, with the language model estimated using Equation 4and the translation probabilities estimated using EM algorithm, we can compute the best triple translation for a given Chinese triple using Equations (1) and 7.", "cite_spans": [], "ref_spans": [ { "start": 155, "end": 163, "text": "Figure 1", "ref_id": null } ], "eq_spans": [], "section": "EM Algorithm", "sec_num": null }, { "text": "This section describes how to extract collocation translation from independent monolingual corpora. First, collocations are extracted from a monolingual triples database. Then, collocation translations are acquired using the triple translation model obtained in section 3.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Collocation translation extraction from two monolingual corpora", "sec_num": "4" }, { "text": "As introduced in section 2, much work has been done to extract collocations. Among all the measure metrics, log likelihood ratio (LLR) has proved to give better results (Duning, 1993; Thanopoulos et al., 2002) . In this paper, we take LLR as the metric to extract collocations from a dependency triple database. Those triples whose LLR values are larger than a given threshold are taken as a collocation. This syntax-based collocation has the advantage that it can represent both adjacent and long distance word association. Here, we only extract the three main types of collocation that have been mentioned in section 3.1.", "cite_spans": [ { "start": 169, "end": 183, "text": "(Duning, 1993;", "ref_id": null }, { "start": 184, "end": 209, "text": "Thanopoulos et al., 2002)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Monolingual collocation extraction", "sec_num": "4.1" }, { "text": "For the acquired collocations, we try to extract their translations from the other monolingual corpus using the triple translation model trained with the method proposed in section 3.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Collocation translation extraction", "sec_num": "4.2" }, { "text": "Our objective is to acquire collocation translations as translation knowledge for a machine translation system, so only highly reliable collocation translations are extracted. Figure 2 describes the algorithm for Chinese-English collocation translation extraction. It can be seen that the best English triple candidate is extracted as the translation of the given Chinese collocation only if the Chinese collocation is also the best translation candidate of the English triple. But the English triple is not necessarily a collocation. English collocation translations can be extracted in a similar way. ", "cite_spans": [], "ref_spans": [ { "start": 176, "end": 184, "text": "Figure 2", "ref_id": "FIGREF3" } ], "eq_spans": [], "section": "Collocation translation extraction", "sec_num": "4.2" }, { "text": "Our English corpus is from Wall Street Journal (1987) (1988) (1989) (1990) (1991) (1992) and Associated Press (1988 -1990 , and the Chinese corpus is from People's Daily (1980) (1981) (1982) (1983) (1984) (1985) (1986) (1987) (1988) (1989) (1990) (1991) (1992) (1993) (1994) (1995) (1996) (1997) (1998) . The two corpora are parsed using the NLPWin parser 1 (Heidorn, 2000) . The statistics for three main types of dependency triples are shown in tables 1 and 2. Token refers to the total number of triple occurrences and Type refers to the number of unique triples in the corpus. Statistic for the extracted Chinese collocations and the collocation translations is shown in The translation dictionaries we used in training and translation are combined from two dictionaries: HITDic and NLPWinDic 2 . The final E-C dictionary contains 126,135 entries, and C-E dictionary contains 91,275 entries.", "cite_spans": [ { "start": 47, "end": 53, "text": "(1987)", "ref_id": null }, { "start": 54, "end": 60, "text": "(1988)", "ref_id": null }, { "start": 61, "end": 67, "text": "(1989)", "ref_id": null }, { "start": 68, "end": 74, "text": "(1990)", "ref_id": null }, { "start": 75, "end": 81, "text": "(1991)", "ref_id": null }, { "start": 82, "end": 88, "text": "(1992)", "ref_id": null }, { "start": 93, "end": 115, "text": "Associated Press (1988", "ref_id": null }, { "start": 116, "end": 121, "text": "-1990", "ref_id": "BIBREF0" }, { "start": 170, "end": 176, "text": "(1980)", "ref_id": null }, { "start": 177, "end": 183, "text": "(1981)", "ref_id": null }, { "start": 184, "end": 190, "text": "(1982)", "ref_id": null }, { "start": 191, "end": 197, "text": "(1983)", "ref_id": null }, { "start": 198, "end": 204, "text": "(1984)", "ref_id": null }, { "start": 205, "end": 211, "text": "(1985)", "ref_id": null }, { "start": 212, "end": 218, "text": "(1986)", "ref_id": null }, { "start": 219, "end": 225, "text": "(1987)", "ref_id": null }, { "start": 226, "end": 232, "text": "(1988)", "ref_id": null }, { "start": 233, "end": 239, "text": "(1989)", "ref_id": null }, { "start": 240, "end": 246, "text": "(1990)", "ref_id": null }, { "start": 247, "end": 253, "text": "(1991)", "ref_id": null }, { "start": 254, "end": 260, "text": "(1992)", "ref_id": null }, { "start": 261, "end": 267, "text": "(1993)", "ref_id": null }, { "start": 268, "end": 274, "text": "(1994)", "ref_id": null }, { "start": 275, "end": 281, "text": "(1995)", "ref_id": null }, { "start": 282, "end": 288, "text": "(1996)", "ref_id": null }, { "start": 289, "end": 295, "text": "(1997)", "ref_id": null }, { "start": 296, "end": 302, "text": "(1998)", "ref_id": null }, { "start": 358, "end": 373, "text": "(Heidorn, 2000)", "ref_id": "BIBREF8" } ], "ref_spans": [], "eq_spans": [], "section": "Implementation of our approach", "sec_num": "4.3" }, { "text": "To evaluate the effectiveness of our methods, two experiments have been conducted. The first one compares our method with three other monolingual corpus based methods in triple translation. The second one evaluates the accuracy of the acquired collocation translation.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experiments and evaluation", "sec_num": "5" }, { "text": "Triple translation experiments are conducted from Chinese to English. We randomly selected 2000 Chinese triples (whose frequency is larger than 2) from the dependency triple database. The standard translation answer sets were built manually by three linguistic experts. For each Chinese triple, its English translation set contain English triples provided by anyone of the three linguists. Among 2000 candidate triples, there are 101 triples that can't be translated into English triples with same relation. For example, the Chinese triple (\u8bb2, VO, \u4ef7\u94b1) should be translated into \"bargain\". The two words in triple cannot be translated separately. We call this kind of collocation translation no-compositional translations. Our current model cannot deal with this kind of translation. In addition, there are also 157 error dependency triples, which result from parsing mistakes. We filtered out these two kinds of triples and got a standard test set with 1,742 Chinese triples and 4,645 translations in total.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Dependency triple translation", "sec_num": "5.1" }, { "text": "We compare our triple translation model with three other models on the same standard test set with the same translation dictionary. As the baseline experiment, Model A selects the highestfrequency translation for each word in triple; Model B selects translation with the maximal target triple probability, as proposed in (Dagan 1994) ; Model C selects translation using both language model and translation model, but the translation probability is simulated by a similarity score which is estimated from monolingual corpus using mutual information measure (Zhou et al., 2001) . Table 4 , where coverage is the percentages of triples which can be translated. Some triples can't be translated by Model B, C and D because of the lack of dictionary translations or data sparseness in triples. In fact, the coverage of Model A is 100%. It was set to the same as others in order to compare accuracy using the same test set. The oracle score is the upper bound accuracy under the conditions of current translation dictionary and standard test set. Top N accuracy is defined as the percentage of triples whose selected top N translations include correct translations.", "cite_spans": [ { "start": 321, "end": 333, "text": "(Dagan 1994)", "ref_id": "BIBREF3" }, { "start": 556, "end": 575, "text": "(Zhou et al., 2001)", "ref_id": null } ], "ref_spans": [ { "start": 578, "end": 585, "text": "Table 4", "ref_id": "TABREF12" } ], "eq_spans": [], "section": "Dependency triple translation", "sec_num": "5.1" }, { "text": "We can see that both Model C and Model D achieve better results than Model B. This shows that the translation model trained from monolingual corpora really helps to improve the performance of translation. Our model also outperforms Model C, which demonstrates the probabilities trained by our EM algorithm achieve better performance than heuristic similarity scores.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Dependency triple translation", "sec_num": "5.1" }, { "text": "In fact, our evaluation method is very rigorous. To avoid bias in evaluation, we take human translation results as standard. The real translation accuracy is reasonably better than the evaluation results. But as we can see, compared to the oracle score, the current models still have much room for improvement. And coverage is also not high due to the limitations of the translation dictionary and the sparse data problem.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Dependency triple translation", "sec_num": "5.1" }, { "text": "47,632 Chinese collocation translations are extracted with the method proposed in section 4. We randomly selected 1000 translations for evaluation. Three linguistic experts tag the acceptability of the translation. Those translations that are tagged as acceptable by at least two experts are evaluated as correct. The evaluation results are shown in Table 5 We can see that the extracted collocation translations achieve a much better result than triple translation. The average accuracy is 63.20% and the collocations with relation AN achieve the highest accuracy of 68.15%. If we only consider those Chinese collocations whose translations are also English collocations, we obtain an even better accuracy of 72.16% as shown in the last row of Table 5 . The results justify our idea that we can acquire reliable translation for collocation by making use of triple translation model in two directions.", "cite_spans": [], "ref_spans": [ { "start": 350, "end": 357, "text": "Table 5", "ref_id": null }, { "start": 745, "end": 752, "text": "Table 5", "ref_id": null } ], "eq_spans": [], "section": "Collocation translation extraction", "sec_num": "5.2" }, { "text": "These acquired collocation translations are very valuable for translation knowledge building. Manually crafting collocation translations can be time-consuming and cannot ensure high quality in a consistent way. Our work will certainly improve the quality and efficiency of collocation translation acquisition.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Collocation translation extraction", "sec_num": "5.2" }, { "text": "Although our approach achieves promising results, it still has some limitations to be remedied in future work.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "5.3" }, { "text": "(1) Translation dictionary extension Due to the limited coverage of the dictionary, a correct translation may not be stored in the dictionary. This naturally limits the coverage of triple translations. Some research has been done to expand translation dictionary using a non-parallel corpus (Rapp, 1999; Keohn and Knight, 2002) . It can be used to improve our work.", "cite_spans": [ { "start": 291, "end": 303, "text": "(Rapp, 1999;", "ref_id": null }, { "start": 304, "end": 327, "text": "Keohn and Knight, 2002)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "5.3" }, { "text": "(2) Noise filtering of parsers Since we use parsers to generate dependency triple databases, this inevitably introduces some parsing mistakes. From our triple translation test data, we can see that 7.85% (157/2000) types of triples are error triples. These errors will certainly influence the translation probability estimation in the training process. We need to find an effective way to filter out mistakes and perform necessary automatic correction.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "5.3" }, { "text": "(3) Non-compositional collocation translation.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "5.3" }, { "text": "Our model is based on the dependency correspondence assumption, which assumes that a triple's translation is also a triple. But there are still some collocations that can't be translated word by word. For example, the Chinese triple (\u5bcc\u6709, VO, \u6210\u6548) usually be translated into \"be effective\"; the English triple (take, VO, place) usually be translated into \"\u53d1\u751f\". The two words in triple cannot be translated separately. Our current model cannot deal with this kind of non-compositional collocation translation. Melamed (1997) and Lin (1999) have done some research on noncompositional phrases discovery. We will consider taking their work as a complement to our model.", "cite_spans": [ { "start": 507, "end": 521, "text": "Melamed (1997)", "ref_id": null }, { "start": 526, "end": 536, "text": "Lin (1999)", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "5.3" }, { "text": "This paper proposes a novel method to train a triple translation model and extract collocation translations from two independent monolingual corpora. Evaluation results show that it outperforms the existing monolingual corpus based methods in triple translation, mainly due to the employment of EM algorithm in cross language translation probability estimation. By making use of the acquired triple translation model in two directions, promising results are achieved in collocation translation extraction.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion and future work", "sec_num": "6" }, { "text": "Our work also demonstrates the possibility of making full use of monolingual resources, such as corpora and parsers for bilingual tasks. This can help overcome the bottleneck of the lack of a large-scale bilingual corpus. This approach is also applicable to comparable corpora, which are also easier to access than bilingual corpora.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion and future work", "sec_num": "6" }, { "text": "In future work, we are interested in extending our method to solving the problem of noncompositional collocation translation. We are also interested in incorporating our triple translation model for sentence level translation.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion and future work", "sec_num": "6" }, { "text": "& Chinese Language Processing. 6(1): 1-26", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "The authors would like to thank John Chen, Jianfeng Gao and Yunbo Cao for their valuable suggestions and comments on a preliminary draft of this paper.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Acknowledgements", "sec_num": "7" } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "Collocations and generalpurpose dictionaries", "authors": [ { "first": "Morton", "middle": [], "last": "Benson", "suffix": "" } ], "year": 1990, "venue": "International Journal of Lexicography", "volume": "3", "issue": "1", "pages": "23--35", "other_ids": {}, "num": null, "urls": [], "raw_text": "Morton Benson. 1990. Collocations and general- purpose dictionaries. International Journal of Lexicography. 3(1):23-35", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "Base noun phrase translation using Web data and the EM algorithm", "authors": [ { "first": "Yunbo", "middle": [], "last": "Cao", "suffix": "" }, { "first": "Hang", "middle": [], "last": "Li", "suffix": "" } ], "year": 2002, "venue": "The 19th International Conference on Computational Linguistics", "volume": "", "issue": "", "pages": "127--133", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yunbo Cao, Hang Li. 2002. Base noun phrase translation using Web data and the EM algorithm. The 19th International Conference on Computational Linguistics. pp.127-133", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "Word association norms, mutural information, and lexicography", "authors": [ { "first": "W", "middle": [], "last": "Kenneth", "suffix": "" }, { "first": "Patrick", "middle": [], "last": "Church", "suffix": "" }, { "first": "", "middle": [], "last": "Hanks", "suffix": "" } ], "year": 1990, "venue": "Computational Linguistics", "volume": "16", "issue": "1", "pages": "22--29", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kenneth W. Church and Patrick Hanks. 1990. Word association norms, mutural information, and lexicography. Computational Linguistics, 16(1):22-29", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "Word sense disambiguation using a second language monolingual corpus", "authors": [ { "first": "Ido", "middle": [], "last": "Dagan", "suffix": "" }, { "first": "Alon", "middle": [], "last": "Itai", "suffix": "" } ], "year": 1994, "venue": "Computational Linguistics", "volume": "20", "issue": "4", "pages": "563--596", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ido Dagan and Alon Itai. 1994. Word sense disambiguation using a second language monolingual corpus. Computational Linguistics, 20(4):563-596", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "Accurate methods for the statistics of surprise and coincidence", "authors": [ { "first": "Ted", "middle": [], "last": "Dunning", "suffix": "" } ], "year": 1993, "venue": "Computational Linguistics", "volume": "19", "issue": "1", "pages": "61--74", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ted Dunning. 1993. Accurate methods for the statistics of surprise and coincidence. Computational Linguistics. 19(1):61-74", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Effectiveness of automatic extraction of bilingual collocations using recursive chain-link-type learning. The 9th Machine Translation Summit", "authors": [ { "first": "Kenji", "middle": [], "last": "Hiroshi Echizen-Ya", "suffix": "" }, { "first": "Yoshi", "middle": [], "last": "Araki", "suffix": "" }, { "first": "Koji", "middle": [], "last": "Momouchi", "suffix": "" }, { "first": "", "middle": [], "last": "Tochinai", "suffix": "" } ], "year": 2003, "venue": "", "volume": "", "issue": "", "pages": "102--109", "other_ids": {}, "num": null, "urls": [], "raw_text": "Hiroshi Echizen-ya, Kenji Araki, Yoshi Momouchi, Koji Tochinai. 2003. Effectiveness of automatic extraction of bilingual collocations using recursive chain-link-type learning. The 9th Machine Translation Summit. pp.102-109", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "An IR approach for translating new words from nonparallel, comparable Texts. The 36th annual conference of the Association for Computational Linguistics", "authors": [ { "first": "Pascale", "middle": [], "last": "Fung", "suffix": "" }, { "first": "Yee Lo", "middle": [], "last": "Yuen", "suffix": "" } ], "year": 1998, "venue": "", "volume": "", "issue": "", "pages": "414--420", "other_ids": {}, "num": null, "urls": [], "raw_text": "Pascale Fung, and Yee Lo Yuen. 1998. An IR approach for translating new words from nonparallel, comparable Texts. The 36th annual conference of the Association for Computational Linguistics. pp. 414-420", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "Resolving query translation ambiguity using a decaying cooccurrence model and syntactic dependence relations", "authors": [ { "first": "Jianfeng", "middle": [], "last": "Gao", "suffix": "" }, { "first": "Jianyun", "middle": [], "last": "Nie", "suffix": "" }, { "first": "Hongzhao", "middle": [], "last": "He", "suffix": "" }, { "first": "Weijun", "middle": [], "last": "Chen", "suffix": "" }, { "first": "Ming", "middle": [], "last": "Zhou", "suffix": "" } ], "year": 2002, "venue": "The 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval", "volume": "", "issue": "", "pages": "183--190", "other_ids": {}, "num": null, "urls": [], "raw_text": "Jianfeng Gao, Jianyun Nie, Hongzhao He, Weijun Chen, Ming Zhou. 2002. Resolving query translation ambiguity using a decaying co- occurrence model and syntactic dependence relations. The 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. pp.183 - 190", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "Intelligent writing assistant", "authors": [ { "first": "G", "middle": [], "last": "Heidorn", "suffix": "" } ], "year": 2000, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "G. Heidorn. 2000. Intelligent writing assistant. In R. Dale, H. Moisl, and H. Somers, editors, A", "links": null } }, "ref_entries": { "FIGREF0": { "type_str": "figure", "text": "of triple tri e . N represents the total counts of all the English triples in the training corpus.", "uris": null, "num": null }, "FIGREF1": { "type_str": "figure", "text": "Figure 1: EM algorithm", "uris": null, "num": null }, "FIGREF2": { "type_str": "figure", "text": "is the total counts of all Chinese triples.", "uris": null, "num": null }, "FIGREF3": { "type_str": "figure", "text": "Collocation translation extraction", "uris": null, "num": null }, "FIGREF4": { "type_str": "figure", "text": "These two dictionaries are built by Harbin Institute of Technology and Microsoft Research respectively.", "uris": null, "num": null }, "FIGREF5": { "type_str": "figure", "text": "", "uris": null, "num": null }, "TABREF8": { "content": "
Class#Type#Token
VO1,579,78319,168,229
AN311,5605,383,200
AV546,0549,467,103
", "type_str": "table", "num": null, "html": null, "text": "" }, "TABREF9": { "content": "
Class#Type#Token
VO1,526,7478,943,903
AN1,163,4406,386,097
AV215,1101,034,410
", "type_str": "table", "num": null, "html": null, "text": "Chinese dependency triples1 The NLPWin parser is a rule-based parser developed at Microsoft research, which parses several languages including Chinese and English. Its output can be a phrase structure parse tree or a logical form which is represented with dependency triples." }, "TABREF10": { "content": "
Class#Type#Translated
VO99,60928,841
AN35,95112,615
AV46,5156,176
Table 3: Extracted Chinese collocations
and E-C translation pairs
", "type_str": "table", "num": null, "html": null, "text": "English dependency triples" }, "TABREF12": { "content": "", "type_str": "table", "num": null, "html": null, "text": "Translation results comparisonThe evaluation results on the standard test set are shown in" } } } }