|
{ |
|
"paper_id": "Y08-1025", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T13:37:45.478512Z" |
|
}, |
|
"title": "An Improved Corpus Comparison Approach to Domain Specific Term Recognition *", |
|
"authors": [ |
|
{ |
|
"first": "Xiaoyue", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "University of Hong Kong Tat Chee Ave", |
|
"location": { |
|
"settlement": "Kowloon, Hong Kong" |
|
} |
|
}, |
|
"email": "xyliu0@cityu.edu.hk" |
|
}, |
|
{ |
|
"first": "Chunyu", |
|
"middle": [], |
|
"last": "Kit", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "University of Hong Kong Tat Chee Ave", |
|
"location": { |
|
"settlement": "Kowloon, Hong Kong" |
|
} |
|
}, |
|
"email": "ctckit@cityu.edu.hk" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "Domain specific terms are words carrying special conceptual meanings in a subject field. Automatic term recognition plays an important role in many natural language processing and knowledge engineering applications such as information retrieval and knowledge mining. This paper explores a novel approach to automatic term extraction based on the basic ideas of corpus comparison and emerging pattern with significant elaboration. It measures the termhood of a term candidate in terms of its peculiarity to a given domain via comparison to several background domains. Our experiments confirm its outperformance against other approaches, achieving an average precision of 83% on the top 10% candidates in terms of their termhood.", |
|
"pdf_parse": { |
|
"paper_id": "Y08-1025", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "Domain specific terms are words carrying special conceptual meanings in a subject field. Automatic term recognition plays an important role in many natural language processing and knowledge engineering applications such as information retrieval and knowledge mining. This paper explores a novel approach to automatic term extraction based on the basic ideas of corpus comparison and emerging pattern with significant elaboration. It measures the termhood of a term candidate in terms of its peculiarity to a given domain via comparison to several background domains. Our experiments confirm its outperformance against other approaches, achieving an average precision of 83% on the top 10% candidates in terms of their termhood.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Automatic term recognition (ATR) plays an important role in many natural language processing applications, e.g., information retrieval (IR) (Chowdhury, 1999; Zhou and Nie, 2005) , information extraction (Yangarber et al., 2000) , domain specific lexicon construction (Hull, 2001) , and topic extraction (Lin, 2004) . Its technological advancements can facilitate all these applications for performance enhancement.", |
|
"cite_spans": [ |
|
{ |
|
"start": 140, |
|
"end": 157, |
|
"text": "(Chowdhury, 1999;", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 158, |
|
"end": 177, |
|
"text": "Zhou and Nie, 2005)", |
|
"ref_id": "BIBREF33" |
|
}, |
|
{ |
|
"start": 203, |
|
"end": 227, |
|
"text": "(Yangarber et al., 2000)", |
|
"ref_id": "BIBREF32" |
|
}, |
|
{ |
|
"start": 267, |
|
"end": 279, |
|
"text": "(Hull, 2001)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 303, |
|
"end": 314, |
|
"text": "(Lin, 2004)", |
|
"ref_id": "BIBREF22" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "Despite the large volume of literature on ATR, further significant success in the field still relies heavily on a sound resolution of two basic issues, namely, the unithood and termhood of a term candidate, as identified in Kageura and Umino (1996) . The former quantifies the unity of a candidate (especially, a multi-word candidate), indicating how likely a candidate is to be an atomic linguistic unit. It works more like a filter to exclude non-atomic (thus unqualified) term candidates but has little authority to determine which atomic language unit is a true term. The latter measures how likely a qualified candidate is to be a true term in a subject field. It plays a decisive role in licensing a term.", |
|
"cite_spans": [ |
|
{ |
|
"start": 224, |
|
"end": 248, |
|
"text": "Kageura and Umino (1996)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "Regardless of the previous progress in term extraction, termhood measurement remains the most critical problem to be solved. Novel technologies and methodologies are needed in order to bring up new insights into our understanding of this problem. This paper is intended to pre-sent a novel statistical approach to domain specific term extraction from a collection of thematic documents following the basic ideas of corpus comparison and emerging pattern. It measures the termhood of a term candidate in a subject domain in terms of its peculiarity to this subject domain via comparison to several background domains.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "In order to avoid unexpected interference from unithood issues, our study focuses on investigating the termhood measurement for mono-word terms. Nevertheless, this is by no means to imply that mono-word ATR can be any easier than multi-word ATR in any sense. It is pointed out in Daille (1994) that the automatic identification of mono-word terms is possibly more complex than that of multi-word ones. One of the reasons is that all structural information that can be utilized for multi-word ATR is not available for mono-word ATR. In this sense, the latter needs to tackle a fundamental issue, that is, how to differentiate between terms and non-terms without resorting to structure information.", |
|
"cite_spans": [ |
|
{ |
|
"start": 280, |
|
"end": 293, |
|
"text": "Daille (1994)", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "The rest of the paper is organized as follows. Section 2 presents a brief review of previous work on ATR, to give a background for our research. Section 3 formulates our approach and the working procedure involved. A series of experiments are then reported in Section 4 for the purpose of evaluation. Section 5 concludes the paper with a highlight on the advantages of our approach.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1." |
|
}, |
|
{ |
|
"text": "Various approaches to ATR were developed in the past. From a methodological point of view, the existing approaches can be classified into the following categories.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Previous Work", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "Linguistic A linguistic approach played a dominant role in the early research on ATR. As early as twenty years ago, Ananiadou (1988) studied the effectiveness of theoretically motivated linguistic knowledge (e.g. morphology) in term recognition. A common procedure involved in this kind of approach is to carry out part-of-speech tagging first and then some pre-defined syntactic patterns, e.g., noun-noun compounds (Dagan and Church, 1994; Wu and Hsu, 2002) and base noun phrases (Justeson and Katz, 1995) , can be applied to identify term candidates. All word combinations that match none of the predefined patterns are filtered out. This approach was reported to achieve good results on small scale corpora. However, its disadvantages include inadequate coverage of pre-defined syntactic patterns, low transplantability to other domains or languages, and incapability of excluding non-term candidates consistent with the pre-defined patterns.", |
|
"cite_spans": [ |
|
{ |
|
"start": 116, |
|
"end": 132, |
|
"text": "Ananiadou (1988)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 416, |
|
"end": 440, |
|
"text": "(Dagan and Church, 1994;", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 441, |
|
"end": 458, |
|
"text": "Wu and Hsu, 2002)", |
|
"ref_id": "BIBREF29" |
|
}, |
|
{ |
|
"start": 481, |
|
"end": 506, |
|
"text": "(Justeson and Katz, 1995)", |
|
"ref_id": "BIBREF15" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Previous Work", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "Statistical Various kinds of statistical information can be utilized to support term extraction, e.g., frequency (Damerau, 1990) , mutual information (Damerau, 1993) , C-value (Frantzi and Ananiadou, 1996) , NC-value (Frantzi et al., 1998) , imp function (Nakagawa, 2001) , KFIDF measure (Xu et al., 2002) , standard deviation (Lin, 2004) and entropy (Chang, 2005) , to name but a few. A multi-word term is assumed to carry a key concept and is thus expected to behave like an atomic text unit. Many of these statistical measures are applied to explore such unity or structural stability of a multi-word candidate, namely, its unithood. Besides, a bootstrapping approach is reported in Chen et al. (2003) to learn domain specific terms from unannotated texts on a subject. Wermter and Hahn (2005) identify multi-word terms among n-grams of words in a large biomedical corpus, measuring their termhood in terms of their paradigmatic modifiability. Although statistical approaches share an advantage, i.e., their language independency, they are far from reliable while working on small corpora.", |
|
"cite_spans": [ |
|
{ |
|
"start": 113, |
|
"end": 128, |
|
"text": "(Damerau, 1990)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 150, |
|
"end": 165, |
|
"text": "(Damerau, 1993)", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 176, |
|
"end": 205, |
|
"text": "(Frantzi and Ananiadou, 1996)", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 217, |
|
"end": 239, |
|
"text": "(Frantzi et al., 1998)", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 255, |
|
"end": 271, |
|
"text": "(Nakagawa, 2001)", |
|
"ref_id": "BIBREF23" |
|
}, |
|
{ |
|
"start": 288, |
|
"end": 305, |
|
"text": "(Xu et al., 2002)", |
|
"ref_id": "BIBREF30" |
|
}, |
|
{ |
|
"start": 327, |
|
"end": 338, |
|
"text": "(Lin, 2004)", |
|
"ref_id": "BIBREF22" |
|
}, |
|
{ |
|
"start": 351, |
|
"end": 364, |
|
"text": "(Chang, 2005)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 686, |
|
"end": 704, |
|
"text": "Chen et al. (2003)", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 773, |
|
"end": 796, |
|
"text": "Wermter and Hahn (2005)", |
|
"ref_id": "BIBREF28" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Previous Work", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "Hybrid Linguistic knowledge is used in conjunction with statistical information in most hybrid approaches to ATR. For example, some syntactic patterns are first applied to identify term candidates, by filtering out those unqualified ones, and then a statistical measure is applied to validate the true terms among them. Daille (1994) presents an integrated approach to ATR that works this way. In addition, a hybrid approach can also be applied to combine several independent term recognizers for a better performance than any of them alone, as reported in Vivaldi et al. (2001) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 320, |
|
"end": 333, |
|
"text": "Daille (1994)", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 557, |
|
"end": 578, |
|
"text": "Vivaldi et al. (2001)", |
|
"ref_id": "BIBREF27" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Previous Work", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "Corpus Comparison This approach is a popular direction in recent ATR research. Its basic idea is to utilize the distinct distributions of terms and non-terms in different corpora to facilitate term extraction. That is, true terms are more prominent in their own subject field than in others. The original idea of this approach can be traced back to Yang (1986) that attempts to identify scientific terms by their statistical distributional difference between science and general texts. The statistics in use for this purpose include document frequency, average frequency, relative standard-deviation, etc. Ahmad et al. (1994) quantify similar contrasting distributions of terms in different corpora by means of the ratio of relative frequencies of a word in a domain corpus and a background corpus. The words with a score larger than 1.0 are then identified as the most potential terms. Chung (2003) applies a similar scheme called normalized frequency ratio to extract single-word terms in anatomy, reporting a performance of about 86% overlap with the results from a manual rating approach. In Uchimoto et al. (2001) , more statistical characteristics (e.g., term frequency, document frequency and field frequency) of a term candidate are explored, achieving an F-score of 58.49%. Kit and Liu (2007) propose to measure mono-word termhood in terms of a candidate's rank difference in a domain and a background corpus.", |
|
"cite_spans": [ |
|
{ |
|
"start": 349, |
|
"end": 360, |
|
"text": "Yang (1986)", |
|
"ref_id": "BIBREF31" |
|
}, |
|
{ |
|
"start": 606, |
|
"end": 625, |
|
"text": "Ahmad et al. (1994)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 887, |
|
"end": 899, |
|
"text": "Chung (2003)", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 1096, |
|
"end": 1118, |
|
"text": "Uchimoto et al. (2001)", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 1283, |
|
"end": 1301, |
|
"text": "Kit and Liu (2007)", |
|
"ref_id": "BIBREF18" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Previous Work", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "The approach of corpus comparison can also be accomplished by statistical tests based on the null hypothesis that there is no difference between the observed frequencies for the same word in different corpora. Words with large testing values indicate a statistically significant difference between corpora and hence are more likely to be terms. The statistical tests include loglikelihood ratio (Rayson and Garside 2000) , \u03c7 2 -test, Mann-Whitney ranks test and t-test (Kilgarriff 2001). Drouin (2003) bases a term extraction process on a statistical test, which uses a normal distribution as an approximation to words' binomial distribution, obtaining an overall precision of 81% on term recognition. Based on Drouin's work, Lemay et al. (2005) examine various corpus comparison approaches in different terminological settings. Among them, one is to use a general corpus as background and another is to break down the specialized corpus into six topical sub-corpora for comparison to the entire specialized corpus.", |
|
"cite_spans": [ |
|
{ |
|
"start": 395, |
|
"end": 420, |
|
"text": "(Rayson and Garside 2000)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 488, |
|
"end": 501, |
|
"text": "Drouin (2003)", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 726, |
|
"end": 745, |
|
"text": "Lemay et al. (2005)", |
|
"ref_id": "BIBREF20" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Previous Work", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "Methodologically, a corpus comparison approach takes advantage of the intrinsic statistical characteristics of true terms in different corpora and thus has a preferable theoretical grounding over others that utilize only a special domain corpus. Our research reported in this paper falls into the category of corpus comparison approach with necessary elaboration for further enhancement.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Previous Work", |
|
"sec_num": "2." |
|
}, |
|
{ |
|
"text": "Terms are linguistic representations of domain specific concepts to encode our special knowledge about a subject field. Emerging pattern (EP) (Dong and Li, 1999 ) presents a similar idea to corpus comparison in the field of database for knowledge discovery. EPs are defined as itemsets whose growth rates, i.e., the ratios of their supports 1 in one dataset over those in another, are larger than a predefined threshold. When applied to datasets with classes (e.g., cancerous vs. normal tissues, poisonous vs. edible mushrooms), EPs can capture significant differences or useful contrasts between the classes in terms of their growth rates. In principle, the larger the growth rates, the more significant the patterns. This approach has been successfully deployed in several applications of data mining, e.g., Li and Wong (2002) on identification of good diagnostic gene groups from gene expression profiles.", |
|
"cite_spans": [ |
|
{ |
|
"start": 142, |
|
"end": 160, |
|
"text": "(Dong and Li, 1999", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 810, |
|
"end": 828, |
|
"text": "Li and Wong (2002)", |
|
"ref_id": "BIBREF21" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "While the EP approach works on well-structured databases, corpus comparison deals with unstructured texts. Following the essential principle shared by the two, we consider domain specific term recognition from a thematic corpus of documents an issue of identifying words and expressions as EPs in the form of string highly peculiar to their own subject fields than to any others. In this sense, the higher peculiarity of a term candidate to a particular domain but lower to the others, the more likely it is to be a true term in that domain.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "Accordingly, we opt to quantify the peculiarity of a term candidate to a subject field in terms of its emerging difference, which is to be scored according to statistical information such as fre-quency difference in its own subject field and another field as background. This illustrates the basic idea of corpus comparison. To our knowledge, however, very few existing approaches of corpus comparison to ATR use more than one background corpus. When comparing a thematic corpus with multiple background corpora, we need to find an appropriate way to sum up the comparison results into the final termhood scores for the term candidates in question. This would pay off if such comparison and summing up can make the termhood scoring more reliable. We will follow this idea to derive the termhood for a term candidate in a target subject field.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "Given a collection of n documents (or corpora) each representing a subject field, henceforth referred to as thematic documents, we follow the following working procedure to extract monoword terms from each corpus. 1. Extraction: Extract all mono-words as term candidates, including those appearing only once, via stop word filtering, and then assign to each of them a weight in terms of the statistical measure in use (e.g., frequency, or tf-idf score). 2. Normalization: Normalize a weight in each subject field according to the sum of all weights in that field. 3. Computing emerging difference: For a candidate w in a target subject field i, calculate its emerging difference d ij via comparison with another subject field j as", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": ") ( ) ( ) ( w s w s w d j i ij = \u2212", |
|
"eq_num": "(1)" |
|
} |
|
], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "where s i (w) and s j (w) are w's normalized weights in fields i and j (0< i, j \u2264n), respectively. Consequently each candidate in each field will have n-1 emerging difference scores corresponding to the n-1 background fields involved in the comparison. 4. Ranking: Rank all candidates in each field i in terms of their emerging differences d ij (\u2022), resulting in n-1 ranking lists for each i accordingly to n-1 background corpora in use. In each ranking list, candidate w has a rank r ij (w) corresponding to its score d ij (w). 5. Sorting: Sort all candidates in each field i in terms of their termhood defined as", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "\u2211 = = n j ij i w r w 1 ) ( ) ( \u03c4 (2)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "where j\u2260i. 6. Evaluation: Examine the sorted list for each subject field to check how true terms are pushed to the top of the list by their termhood. In our approach, a subject corpus is compared with more than one background corpus and all comparison results for each candidate are summed up together to represent its termhood (or peculiarity) in a subject field in question. However, a term candidate may get a negative score from (1) above, which will, unreasonably, weaken the total sum of scores from the comparison to all other background corpora. This certainly would not help to achieve the right ranking outcomes in an ATR output list sorted by the termhood scores so resulted. To alleviate this problem, we opt to rank term candidates first according to their emerging differences before the summing-up is conducted. The ranking in this way keeps the relative position (or relationship) of term candidates in the candidate list while avoiding the unexpected problem. Finally, the termhood of a term candidate is measured by the sum of its ranks that are derived from the comparison to a number of other domain corpora.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Term Extraction", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "A number of experiments following the above working procedure are carried out on the BLIS corpus (Kit et al., 2005) to extract legal terms in HK laws. The corpus consists of all ordinances and subsidiary legislation to prescribe laws and regulations involving almost every aspect of livings in HK. Excluding those repealed, ceased, expired, or not adopted ordinances, there are a total of 503 chapters in the current version of the corpus. Each ordinance can be regarded as a sub-corpus for a \"field\", e.g., \"Public Finance Ordinance\", \"Forest and Countryside Ordinance\", \"Hospital Authority Ordinance\", and \"Marriage Ordinance\". We use the English texts from the corpus as data for our experiments.", |
|
"cite_spans": [ |
|
{ |
|
"start": 97, |
|
"end": 115, |
|
"text": "(Kit et al., 2005)", |
|
"ref_id": "BIBREF19" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "4.1.Data", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In order to get reliable statistical information about individual words during the extraction procedure, a series of preprocessing steps are carried out, including word tokenization, lemmatization, and filtering of stop words (including words with only digits or without letters, monocharacter words and function words). The evaluation of our approach focuses on the precision of ATR output, for our main concern is how to capture as many true terms as possible at the high end of a candidate list sorted by termhood. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "4.1.Data", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Applying the ATR procedure given in Section 3 above to the BLIS corpus with tf-idf (Salton, 1992) scoring for the initial weighting in the first step, in total we get 503 output lists sorted by termhood. Among them 30 lists are randomly picked for evaluation. For each selected list, the top 10% candidates are manually checked. Chang (2005) presents a modified tf-idf model based on inter-domain entropy calculation, giving an average precision of 65.4% on the top 10% of ATR output. It is assumed that around 10% words in each sub-domain are domain specific terms. The exact percentage of true terms among words, however, may vary significantly in different domains. Table 1 presents a distribution of precision scores over the 30 BLIS chapters manually evaluated, each of which corresponds to an ATR output list. The precision score var-ies from above 70% to 100%, giving an average of 82.98%. As a whole, twenty-two chapters, i.e., 73% of the evaluated lists, have a precision greater than 80%. A few output samples are presented in Table 2 , illustrating the top 10 term candidates in the ATR output for four BLIS chapters, with non-terms highlighted in italic font. An illustration of the performance of the tf-idf scoring vs. corpus size is presented in Table 3 , resulted from the experiments on the same 30 chapters as above. The number of recognized terms (i.e., 10% of the total number of word types in a corpus) in those chapters varies from 15 to 148. The best performance, a precision of 100%, is achieved on a small chapter of only 15 candidates, whereas a precision of 93% is on a chapter of 103 candidates. This seems to suggest that our approach is able to achieve an excellent performance even on a very small corpus. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 83, |
|
"end": 97, |
|
"text": "(Salton, 1992)", |
|
"ref_id": "BIBREF25" |
|
}, |
|
{ |
|
"start": 329, |
|
"end": 341, |
|
"text": "Chang (2005)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 669, |
|
"end": 676, |
|
"text": "Table 1", |
|
"ref_id": "TABREF0" |
|
}, |
|
{ |
|
"start": 1037, |
|
"end": 1044, |
|
"text": "Table 2", |
|
"ref_id": "TABREF1" |
|
}, |
|
{ |
|
"start": 1261, |
|
"end": 1268, |
|
"text": "Table 3", |
|
"ref_id": "TABREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "4.2.Weighting with tf-idf", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "On the same data set, several other scoring schemes have also been tested for a comparison with the tf-idf scoring above. The comparison is carried out on the same 30 chapters. 1. Frequency for initial scoring: Use frequency as the initial scoring measure, and then follow the above working procedures in Section 3 to derive a term list for each sub-corpus. 2. Corpus comparison using frequency ratio: For each mono-word term candidate, compute the ratio of its frequencies in its own subject sub-corpus and a general background corpus. 2 Candidates with a ratio greater than 1.0 are recognized as true terms. 3. Using tf-idf as termhood: Sort the term candidates in a sub-corpus (i.e., an ordinance as a document in our case) in terms of their tf-idf scores as given by (3) below, where freq(w) is the frequency of a candidate w in the sub-corpus, N the total number of documents in the entire corpus in use and d(w) the number of documents containing w.", |
|
"cite_spans": [ |
|
{ |
|
"start": 537, |
|
"end": 538, |
|
"text": "2", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "4.3.Comparison with Other Approaches", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": ") ( log ) ( ) ( w d N w freq w idf tf \u22c5 = \u2212", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "4.3.Comparison with Other Approaches", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(3) Figure 1 presents the performance of the four approaches given above in terms of their precision scores on the 30 chapters, and the average precision achieved by each approach is presented in Table 4 . From this figure and table we can see that corpus comparison using frequency ratio directly is outperformed by our working procedure with frequency for initial scoring. Similarly, the working procedure with tf-idf for initial scoring outperforms that using tf-idf directly as termhood, and performs the best among all the four ATR procedures formulated above. All these verify the significance and effectiveness of the improved corpus comparison approach that we have implemented for enhancing the current ATR technology.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 4, |
|
"end": 12, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
}, |
|
{ |
|
"start": 196, |
|
"end": 203, |
|
"text": "Table 4", |
|
"ref_id": "TABREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "4.3.Comparison with Other Approaches", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We have presented in the above sections an improved approach of corpus comparison to automatic extraction of domain specific terms from thematic corpora, which extends the basic principle of corpus comparison and emerging pattern effectively for performance enhancement. Different from the previous approaches, our approach compares a subject corpus with more than one background corpora and sums up the respective comparison results for each term candidate as its termhood score in the subject field in question.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "Accordingly, we have implemented a novel working procedure for term extraction to examine the effectiveness of this approach. The experiments we have carried out on the BLIS corpus of HK laws show that the proposed approach outperforms other approaches of direct corpus comparison and achieves an average precision of 82.98% on the top 10% of candidates according to their termhood in each domain corpus. Also, a nice advantage of this approach is that its performance seems sustainable on a small corpus.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "22nd Pacific Asia Conference on Language, Information and Computation, pages 253-261", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In database studies, the term support refers to the frequency of an itemset in a dataset.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "British National Corpus is used as the background corpus here for comparison. See its official site at http://www.natcorp.ox.ac.uk/ for more information.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "What is a Term? The Semi-Automatic Extraction of Terms from Text", |
|
"authors": [ |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Ahmad", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Davies", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Fulford", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Rogers", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Translation Studies: An Interdiscipline", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "267--278", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ahmad, K., A. Davies, H. Fulford and M. Rogers. 1994. What is a Term? The Semi-Automatic Extraction of Terms from Text. In M.S. Hornby, F. Pochhacker and K. Kaindl (eds), Translation Studies: An Interdiscipline, pp. 267-278. Amsterdam: John Benjamins Pub- lishing Company.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "A Methodology for Automatic Term Recognition", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Ananiadou", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1988, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ananiadou, S. 1988. A Methodology for Automatic Term Recognition. Ph.D. thesis, University of Manchester Institute of Science and Technology.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Domain Specific Word Extraction from Hierarchical Web Documents: A First Step toward Building Lexicon Trees from Web Corpora", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Chang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the Fourth SIGHAN Workshop on Chinese Language Learning", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "64--71", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chang, J. S. 2005. Domain Specific Word Extraction from Hierarchical Web Documents: A First Step toward Building Lexicon Trees from Web Corpora. Proceedings of the Fourth SIGHAN Workshop on Chinese Language Learning, pp. 64-71. Korea.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Automatic Learning Field Words by Bootstrapping", |
|
"authors": [ |
|
{ |
|
"first": "W", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"B" |
|
], |
|
"last": "Zhu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "T", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Yao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [ |
|
"X" |
|
], |
|
"last": "Zhang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the Joint Seminar on Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "67--72", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chen, W. L., J. B. Zhu, T. S. Yao and Y. X. Zhang. 2003. Automatic Learning Field Words by Bootstrapping. Proceedings of the Joint Seminar on Computational Linguistics 2003, pp. 67-72. Beijing: Tsinghua University Press.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Introduction to Modern Information Retrieval", |
|
"authors": [ |
|
{ |
|
"first": "G", |
|
"middle": [ |
|
"G" |
|
], |
|
"last": "Chowdhury", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "Library Association", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chowdhury, G. G. 1999. Introduction to Modern Information Retrieval. London: Library Asso- ciation.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "A Corpus Comparison Approach for Terminology Extraction", |
|
"authors": [ |
|
{ |
|
"first": "T", |
|
"middle": [], |
|
"last": "Chung", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Terminology", |
|
"volume": "9", |
|
"issue": "2", |
|
"pages": "221--246", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chung, T. 2003. A Corpus Comparison Approach for Terminology Extraction. Terminology, 9(2), 221-246.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Approche Mixte pour l'extraction Automatique de Terminologie: Statistique Lexicale et Filtres Linguistiques", |
|
"authors": [ |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Daille", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Daille, B. 1994. Approche Mixte pour l'extraction Automatique de Terminologie: Statistique Lexicale et Filtres Linguistiques. Ph.D. thesis, University Paris 7, France.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Termight: Identifying and Translating Technical Terminology", |
|
"authors": [ |
|
{ |
|
"first": "I", |
|
"middle": [], |
|
"last": "Dagan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Church", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Proceedings of the 4th Conference on Applied Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "34--40", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dagan, I. and K. Church. 1994. Termight: Identifying and Translating Technical Terminology. Proceedings of the 4th Conference on Applied Natural Language Processing, pp. 34-40. Stuttgart, Germany.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Generating and Evaluating Domain-Oriented Multi-Word Terms from Texts", |
|
"authors": [ |
|
{ |
|
"first": "F", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Damerau", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1993, |
|
"venue": "Information Processing & Management", |
|
"volume": "29", |
|
"issue": "4", |
|
"pages": "433--447", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Damerau, F. J. 1993. Generating and Evaluating Domain-Oriented Multi-Word Terms from Texts. Information Processing & Management, 29(4), 433-447.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Evaluating Computer-Generated Domain-Oriented Vocabularies. Information Processing & Management", |
|
"authors": [ |
|
{ |
|
"first": "F", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Damerau", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1990, |
|
"venue": "", |
|
"volume": "26", |
|
"issue": "", |
|
"pages": "791--801", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Damerau, F. J. 1990. Evaluating Computer-Generated Domain-Oriented Vocabularies. Informa- tion Processing & Management, 26(6), 791-801.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Efficient Mining of Emerging Patterns: Discovering Trends and Differences", |
|
"authors": [ |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Dong", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "43--52", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dong, G. and J. Li. 1999. Efficient Mining of Emerging Patterns: Discovering Trends and Dif- ferences. Proceedings of the Fifth ACM SIGKDD International Conference on Knowl- edge Discovery and Data Mining, pp. 43-52. San Diego, CA: ACM Press.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Term Extraction Using Non-technical Corpora as a Point of Leverage", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Drouin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Terminology", |
|
"volume": "9", |
|
"issue": "1", |
|
"pages": "99--115", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Drouin, P. 2003. Term Extraction Using Non-technical Corpora as a Point of Leverage. Termi- nology, 9(1), 99-115.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "The C-value/NC-value Method of Automatic Recognition for Multi-Word Terms", |
|
"authors": [ |
|
{ |
|
"first": "K", |
|
"middle": [ |
|
"T" |
|
], |
|
"last": "Frantzi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Ananiadou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Tsujii", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the 2nd European Conference on Research and Advanced Technology for Digital Libraries", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "585--604", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Frantzi, K. T., S. Ananiadou and J. Tsujii. 1998. The C-value/NC-value Method of Automatic Recognition for Multi-Word Terms. In C. Nikolaou and C. Stephanidis (eds.), Proceed- ings of the 2nd European Conference on Research and Advanced Technology for Digital Libraries, pp. 585-604. Heraklion, Crete, Greece.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Extracting Nested Collocations", |
|
"authors": [ |
|
{ |
|
"first": "K", |
|
"middle": [ |
|
"T" |
|
], |
|
"last": "Frantzi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Ananiadou", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1996, |
|
"venue": "Proceedings of the 16th International Conference on Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "41--46", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Frantzi, K. T. and S. Ananiadou. 1996. Extracting Nested Collocations. Proceedings of the 16th International Conference on Computational Linguistics, pp. 41-46. Copenhagen, Den- mark.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Software Tools to Support the Construction of Bilingual Terminology Lexicons", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [ |
|
"A" |
|
], |
|
"last": "Hull", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Recent Advances in Computational Terminology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "225--244", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hull, D. A. 2001. Software Tools to Support the Construction of Bilingual Terminology Lexi- cons. In D. Bourigault, C. Jacquemin and M.C. L'Homme (eds), Recent Advances in Computational Terminology, pp. 225-244. Amsterdam: John Benjamins Publishing Com- pany.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Technical Terminology: Some Linguistic Properties and an Algorithm for Identification in Text", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Justeson", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Katz", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1995, |
|
"venue": "Natural Language Engineering", |
|
"volume": "1", |
|
"issue": "1", |
|
"pages": "9--27", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Justeson, J. S. and S. M. Katz. 1995. Technical Terminology: Some Linguistic Properties and an Algorithm for Identification in Text. Natural Language Engineering, 1(1), 9-27.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Comparing Corpora", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Kilgarriff", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "International Journal of Corpus Linguistics", |
|
"volume": "6", |
|
"issue": "1", |
|
"pages": "97--133", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kilgarriff, A. 2001. Comparing Corpora. International Journal of Corpus Linguistics, 6(1), 97- 133.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "Mono-Word Termhood as Rank Difference in Domain and Background Corpora", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Kit", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of International Conference: Keyness in Text", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "41--45", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kit, C. and X. Liu. 2007. Mono-Word Termhood as Rank Difference in Domain and Back- ground Corpora. Proceedings of International Conference: Keyness in Text, pp. 41-45. Pontignano, Siena, Italy.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Harvesting the Bitexts of the Laws of Hong Kong from the Web", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Kit", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [ |
|
"K" |
|
], |
|
"last": "Sin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Webster", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the 5th Workshop on Asian Language Resources", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "71--78", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kit, C., X. Liu, K. K. Sin and J. J. Webster. 2005. Harvesting the Bitexts of the Laws of Hong Kong from the Web. Proceedings of the 5th Workshop on Asian Language Resources, pp. 71-78. Jeju Island, Korea.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Two Methods for Extracting \"Specific\" Single-Word Terms from Specialized Corpora: Experimentation and Evaluation", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Lemay", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "M.-C. L'homme", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Drouin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "International Journal of Corpus Linguistics", |
|
"volume": "10", |
|
"issue": "2", |
|
"pages": "227--255", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lemay, C., M.-C. L'Homme and P. Drouin. 2005. Two Methods for Extracting \"Specific\" Sin- gle-Word Terms from Specialized Corpora: Experimentation and Evaluation. Interna- tional Journal of Corpus Linguistics, 10(2), 227-255.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Identifying Good Diagnostic Gene Groups from Gene Expression Profiles Using the Concept of Emerging Patterns", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Wong", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Bioinformatics", |
|
"volume": "18", |
|
"issue": "5", |
|
"pages": "725--734", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Li, J. and L. Wong. 2002. Identifying Good Diagnostic Gene Groups from Gene Expression Profiles Using the Concept of Emerging Patterns. Bioinformatics, 18(5), 725-734.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Topic Extraction Based on Techniques of Term Extraction and Term Clustering", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Lin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Computational Linguistics and Chinese Language Processing", |
|
"volume": "9", |
|
"issue": "", |
|
"pages": "97--112", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lin, S. C. 2004. Topic Extraction Based on Techniques of Term Extraction and Term Clustering. Computational Linguistics and Chinese Language Processing, 9(2), 97-112.", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "Automatic Term Recognition Based on Statistics of Compound Nouns", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Nakagawa", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Terminology", |
|
"volume": "6", |
|
"issue": "2", |
|
"pages": "195--210", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Nakagawa, H. 2001. Automatic Term Recognition Based on Statistics of Compound Nouns. Terminology, 6(2), 195-210.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Comparing Corpora Using Frequency Profiling", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Rayson", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Garside", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of the Workshop on Comparing Corpora, the 38th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--6", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rayson, P. and R. Garside. 2000. Comparing Corpora Using Frequency Profiling. Proceedings of the Workshop on Comparing Corpora, the 38th Annual Meeting of the Association for Computational Linguistics, pp. 1-6.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "The State of Retrieval System Evaluation", |
|
"authors": [ |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Salton", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1992, |
|
"venue": "", |
|
"volume": "28", |
|
"issue": "", |
|
"pages": "441--449", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Salton, G. 1992. The State of Retrieval System Evaluation. Information Processing & Manage- ment, 28(4), 441-449.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Term Recognition Using Corpora from Different Fields", |
|
"authors": [ |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Uchimoto", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Sekine", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Murata", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Ozaku", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Isahara", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Japanese Term Extraction: Special issue of Terminology", |
|
"volume": "6", |
|
"issue": "2", |
|
"pages": "233--256", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Uchimoto, K., S. Sekine, M. Murata, H. Ozaku and H. Isahara. 2001. Term Recognition Using Corpora from Different Fields. Japanese Term Extraction: Special issue of Terminology, 6(2), 233-256.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Improving Term Extraction by System Combination Using Boosting", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Vivaldi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "M\u00e0rquez", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Rodr\u00edguez", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Proceedings of the 12th European Conference on Machine Learning", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "515--526", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Vivaldi, J., L. M\u00e0rquez and H. Rodr\u00edguez. 2001. Improving Term Extraction by System Combi- nation Using Boosting. Proceedings of the 12th European Conference on Machine Learn- ing, pp. 515-526. Freiburg, Germany.", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "Finding New Terminology in Very Large Corpora", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Wermter", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "U", |
|
"middle": [], |
|
"last": "Hahn", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the Third International Conference on Knowledge Capture", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "137--144", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wermter, J. and U. Hahn. 2005. Finding New Terminology in Very Large Corpora. Proceedings of the Third International Conference on Knowledge Capture, pp. 137-144. Banff, Al- berta, Canada.", |
|
"links": null |
|
}, |
|
"BIBREF29": { |
|
"ref_id": "b29", |
|
"title": "SOAT: A Semi-Automatic Domain Ontology Acquisition Tool from Chinese Corpus", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [ |
|
"H" |
|
], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Hsu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of the 19th international conference on Computational linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--5", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wu, S. H. and W. L. Hsu. 2002. SOAT: A Semi-Automatic Domain Ontology Acquisition Tool from Chinese Corpus. Proceedings of the 19th international conference on Computa- tional linguistics, pp. 1-5. Taipei.", |
|
"links": null |
|
}, |
|
"BIBREF30": { |
|
"ref_id": "b30", |
|
"title": "A Domain Adaptive Approach to Automatic Acquisition of Domain Relevant Terms and Their Relations with Bootstrapping", |
|
"authors": [ |
|
{ |
|
"first": "F", |
|
"middle": [ |
|
"Y" |
|
], |
|
"last": "Xu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Kurz", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Piskorski", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Schmeier", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of the Third International Conference on Language Resources and Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Xu, F. Y., D. Kurz, J. Piskorski and S. Schmeier. 2002. A Domain Adaptive Approach to Automatic Acquisition of Domain Relevant Terms and Their Relations with Bootstrap- ping. Proceedings of the Third International Conference on Language Resources and Evaluation, Spain.", |
|
"links": null |
|
}, |
|
"BIBREF31": { |
|
"ref_id": "b31", |
|
"title": "A New Technique for Identifying Scientific/Technical Terms and Describing Science Texts", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [ |
|
"Z" |
|
], |
|
"last": "Yang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1986, |
|
"venue": "Literary and Linguistic Computing", |
|
"volume": "1", |
|
"issue": "2", |
|
"pages": "93--103", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yang, H. Z. 1986. A New Technique for Identifying Scientific/Technical Terms and Describing Science Texts. Literary and Linguistic Computing, 1(2), 93-103.", |
|
"links": null |
|
}, |
|
"BIBREF32": { |
|
"ref_id": "b32", |
|
"title": "Automatic Acquisition of Domain Knowledge for Information Extraction", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Yangarber", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Grishman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Tapanainen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Huttunen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of the 18th International Conference on Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yangarber, R., R. Grishman, P. Tapanainen and S. Huttunen. 2000. Automatic Acquisition of Domain Knowledge for Information Extraction. Proceedings of the 18th International Conference on Computational Linguistics. Germany.", |
|
"links": null |
|
}, |
|
"BIBREF33": { |
|
"ref_id": "b33", |
|
"title": "Improving Retrieval Effectiveness by Using Key Terms in Top Retrieved Documents", |
|
"authors": [ |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Y", |
|
"middle": [], |
|
"last": "Nie", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Advances in Information Retrieval", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhou, G. and Y. Nie. 2005. Improving Retrieval Effectiveness by Using Key Terms in Top Re- trieved Documents. Advances in Information Retrieval, 3408.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"num": null, |
|
"type_str": "figure", |
|
"text": "Performance comparison", |
|
"uris": null |
|
}, |
|
"TABREF0": { |
|
"num": null, |
|
"html": null, |
|
"type_str": "table", |
|
"content": "<table><tr><td>Range of precision</td><td>1 \u2265 p \u2265 0.9</td><td colspan=\"2\">0.9 > p \u2265 0.8 0.8 > p \u2265 0.7</td></tr><tr><td>Number of chapters</td><td>6</td><td>16</td><td>8</td></tr></table>", |
|
"text": "Distribution of precision score" |
|
}, |
|
"TABREF1": { |
|
"num": null, |
|
"html": null, |
|
"type_str": "table", |
|
"content": "<table><tr><td>Chapter</td><td>Terms in the output</td></tr><tr><td>Probate and Administration</td><td>executor, probate, administrator, caveat, testator,</td></tr><tr><td>Ordinance</td><td>box, decease, affidavit, estate, renter</td></tr><tr><td>Import and Export Ordi-</td><td>export, import, manifest, cargo, article,</td></tr><tr><td>nance</td><td>transhipment, trader, smuggle, validate, customs</td></tr><tr><td>Theft Ordinance</td><td>steal, deception, burglary, theft, deceit, inducement, indictment, menace, robbery, cheat</td></tr><tr><td>Adoption Ordinance</td><td>adoption, infant, litem, accredit, adopt, guardian, parent, adoptive, wedlock, applicant</td></tr></table>", |
|
"text": "Samples of ATR output" |
|
}, |
|
"TABREF2": { |
|
"num": null, |
|
"html": null, |
|
"type_str": "table", |
|
"content": "<table><tr><td>Precision</td><td>Size</td><td>Precision</td><td>Size</td><td>Precision</td><td>Size</td></tr><tr><td>1.0000</td><td>15</td><td>0.8684</td><td>38</td><td>0.8052</td><td>77</td></tr><tr><td>0.9655</td><td>29</td><td>0.8649</td><td>148</td><td>0.8000</td><td>45</td></tr><tr><td>0.9474</td><td>19</td><td>0.8594</td><td>64</td><td>0.7980</td><td>99</td></tr><tr><td>0.9320</td><td>103</td><td>0.8571</td><td>42</td><td>0.7826</td><td>92</td></tr><tr><td>0.9048</td><td>21</td><td>0.8293</td><td>82</td><td>0.7805</td><td>41</td></tr><tr><td>0.9000</td><td>30</td><td>0.8250</td><td>40</td><td>0.7798</td><td>109</td></tr><tr><td>0.8889</td><td>27</td><td>0.8228</td><td>79</td><td>0.7714</td><td>35</td></tr><tr><td>0.8727</td><td>55</td><td>0.8167</td><td>60</td><td>0.7419</td><td>93</td></tr><tr><td>0.8718</td><td>117</td><td>0.8078</td><td>26</td><td>0.7391</td><td>46</td></tr><tr><td>0.8710</td><td>31</td><td>0.8065</td><td>31</td><td>0.7246</td><td>69</td></tr></table>", |
|
"text": "Performance vs. corpus size" |
|
}, |
|
"TABREF3": { |
|
"num": null, |
|
"html": null, |
|
"type_str": "table", |
|
"content": "<table><tr><td>Approach</td><td>tf-idf as weight</td><td>freq as weight</td><td colspan=\"2\">freq Ratio tf-idf as termhood</td></tr><tr><td>Average precision</td><td>82.98%</td><td>75.34%</td><td>66.93%</td><td>81.46%</td></tr></table>", |
|
"text": "The average precision" |
|
} |
|
} |
|
} |
|
} |