|
{ |
|
"paper_id": "D10-1042", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:51:42.821934Z" |
|
}, |
|
"title": "Mining Name Translations from Entity Graph Mapping *", |
|
"authors": [ |
|
{ |
|
"first": "Gae-Won", |
|
"middle": [], |
|
"last": "You", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Pohang University of Science and Technology", |
|
"location": { |
|
"settlement": "Pohang", |
|
"country": "Republic of Korea" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Seung-Won", |
|
"middle": [], |
|
"last": "Hwang", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Pohang University of Science and Technology", |
|
"location": { |
|
"settlement": "Pohang", |
|
"country": "Republic of Korea" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Young-In", |
|
"middle": [], |
|
"last": "Song", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Long", |
|
"middle": [], |
|
"last": "Jiang", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Zaiqing", |
|
"middle": [], |
|
"last": "Nie", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "This paper studies the problem of mining entity translation, specifically, mining English and Chinese name pairs. Existing efforts can be categorized into (a) a transliterationbased approach leveraging phonetic similarity and (b) a corpus-based approach exploiting bilingual co-occurrences, each of which suffers from inaccuracy and scarcity respectively. In clear contrast, we use unleveraged resources of monolingual entity co-occurrences, crawled from entity search engines, represented as two entity-relationship graphs extracted from two language corpora respectively. Our problem is then abstracted as finding correct mappings across two graphs. To achieve this goal, we propose a holistic approach, of exploiting both transliteration similarity and monolingual co-occurrences. This approach, building upon monolingual corpora, complements existing corpus-based work, requiring scarce resources of parallel or comparable corpus, while significantly boosting the accuracy of transliteration-based work. We validate our proposed system using real-life datasets.", |
|
"pdf_parse": { |
|
"paper_id": "D10-1042", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "This paper studies the problem of mining entity translation, specifically, mining English and Chinese name pairs. Existing efforts can be categorized into (a) a transliterationbased approach leveraging phonetic similarity and (b) a corpus-based approach exploiting bilingual co-occurrences, each of which suffers from inaccuracy and scarcity respectively. In clear contrast, we use unleveraged resources of monolingual entity co-occurrences, crawled from entity search engines, represented as two entity-relationship graphs extracted from two language corpora respectively. Our problem is then abstracted as finding correct mappings across two graphs. To achieve this goal, we propose a holistic approach, of exploiting both transliteration similarity and monolingual co-occurrences. This approach, building upon monolingual corpora, complements existing corpus-based work, requiring scarce resources of parallel or comparable corpus, while significantly boosting the accuracy of transliteration-based work. We validate our proposed system using real-life datasets.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Entity translation aims at mapping the entity names (e.g., people, locations, and organizations) in source language into their corresponding names in target language. While high quality entity translation is essential in cross-lingual information access and trans-lation, it is non-trivial to achieve, due to the challenge that entity translation, though typically bearing pronunciation similarity, can also be arbitrary, e.g., Jackie Chan and (pronounced Cheng Long). Existing efforts to address these challenges can be categorized into transliteration-and corpusbased approaches. Transliteration-based approaches (Wan and Verspoor, 1998; Knight and Graehl, 1998) identify translations based on pronunciation similarity, while corpus-based approaches mine bilingual co-occurrences of translation pairs obtained from parallel (Kupiec, 1993; Feng et al., 2004) or comparable (Fung and Yee, 1998) corpora, or alternatively mined from bilingual sentences (Lin et al., 2008; Jiang et al., 2009) . These two approaches have complementary strength-transliteration-based similarity can be computed for any name pair but cannot mine translations of little (or none) phonetic similarity. Corpus-based similarity can support arbitrary translations, but require highly scarce resources of bilingual co-occurrences, obtained from parallel or comparable bilingual corpora.", |
|
"cite_spans": [ |
|
{ |
|
"start": 615, |
|
"end": 639, |
|
"text": "(Wan and Verspoor, 1998;", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 640, |
|
"end": 664, |
|
"text": "Knight and Graehl, 1998)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 826, |
|
"end": 840, |
|
"text": "(Kupiec, 1993;", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 841, |
|
"end": 859, |
|
"text": "Feng et al., 2004)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 874, |
|
"end": 894, |
|
"text": "(Fung and Yee, 1998)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 952, |
|
"end": 970, |
|
"text": "(Lin et al., 2008;", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 971, |
|
"end": 990, |
|
"text": "Jiang et al., 2009)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this paper, we propose a holistic approach, leveraging both transliteration-and corpus-based similarity. Our key contribution is to replace the use of scarce resources of bilingual co-occurrences with the use of untapped and significantly larger resources of monolingual co-occurrences for translation. In particular, we extract monolingual cooccurrences of entities from English and Chinese Web corpora, which are readily available from entity search engines such as PeopleEntityCube 1 , deployed by Microsoft Research Asia. Such engine automatically extracts people names from text and their co-occurrences to retrieve related entities based on co-occurrences. To illustrate, Figure 1 (a) demonstrates the query result for \"Bill Gates,\" retrieving and visualizing the \"entity-relationship graph\" of related people names that frequently co-occur with Bill in English corpus. Similarly, entity-relationship graphs can be built over other language corpora, as Figure 1 (b) demonstrates the corresponding results for the same query, from Renlifang 2 on Chinese Web corpus. From this point on, for the sake of simplicity, we refer to English and Chinese graphs, simply as G e and G c respectively. Though we illustrate with English-Chinese pairs in the paper, our method can be easily adapted to other language pairs.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 681, |
|
"end": 689, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF2" |
|
}, |
|
{ |
|
"start": 962, |
|
"end": 970, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In particular, we propose a novel approach of abstracting entity translation as a graph matching problem of two graphs G e and G c in Figures 1(a) and (b). Specifically, the similarity between two nodes v e and v c in G e and G c is initialized as their transliteration similarity, which is iteratively refined based on relational similarity obtained from monolingual cooccurrences. To illustrate this, an English news article mentioning \"Bill Gates\" and \"Melinda Gates\" evidences a relationship between the two entities, which can be quantified from their co-occurrences in the entire English Web corpus. Similarly, we can mine Chinese news articles to obtain the relationships between \"", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 134, |
|
"end": 146, |
|
"text": "Figures 1(a)", |
|
"ref_id": "FIGREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "\u2022 \" and \" \u2022 \". Once these two bilingual graphs of people and their relationships are harvested, entity translation can leverage these parallel relationships to further evidence the mapping between translation pairs, as Figure 1 (c) illustrates.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 219, |
|
"end": 227, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "To highlight the advantage of our proposed approach, we compare our results with commercial machine translators (1) Engkoo 3 developed in Microsoft Research Asia and (2) Google Translator 4 . In particular, Figure 2 reports the precision for two groups-\"heads\" that belong to top-100 popular people (determined by the number of hits), among randomly sampled 304 people names 5 from six graph pairs of size 1,000 each, and the remaining \"tails\". Commercial translators such as Google, leveraging bilingual co-occurrences that are scarce for tails, show significantly lower precision for tails. Meanwhile, our work, depending solely on monolingual co-occurrences, shows high precision, for both heads and tails.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 207, |
|
"end": 215, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Our focus is to boost translation accuracy for long tails with non-trivial Web occurrences in each monolingual corpus, but not with much bilingual cooccurrences, e.g., researchers publishing actively in two languages but not famous enough to be featured in multi-lingual Wikipedia entries or news articles. As existing translators are already highly accurate for popular heads, this focus well addresses the remaining challenges for entity translation.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "To summarize, we believe that this paper has the following contributions:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "\u2022 We abstract entity translation problem as a graph mapping between entity-relationship graphs in two languages.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "\u2022 We develop an effective matching algorithm leveraging both pronunciation and cooccurrence similarity. This holistic approach complements existing approaches and enhances the translation coverage and accuracy.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "\u2022 We validate the effectiveness of our approach using various real-life datasets.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The rest of this paper is organized as follows. Section 2 reviews existing work. Section 3 then develops our framework. Section 4 reports experimental results and Section 5 concludes our work. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this section, we first survey related efforts, categorized into transliteration-based and corpus-based approaches. Our approach leveraging both is complementary to these efforts.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Many name translations are loosely based on phonetic similarity, which naturally inspires transliteration-based translation of finding the translation with the closest pronunciation similarity, using either rule-based (Wan and Verspoor, 1998) or statistical (Knight and Graehl, 1998; Li et al., 2004) approaches. However, people are free to designate arbitrary bilingual names of little (or none) phonetic similarity, for which the transliteration-based approach is not effective.", |
|
"cite_spans": [ |
|
{ |
|
"start": 218, |
|
"end": 242, |
|
"text": "(Wan and Verspoor, 1998)", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 258, |
|
"end": 283, |
|
"text": "(Knight and Graehl, 1998;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 284, |
|
"end": 300, |
|
"text": "Li et al., 2004)", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Transliteration-based Approaches", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Corpus-based approach can mine arbitrary translation pairs, by mining bilingual co-occurrences from parallel and comparable bilingual corpora. Using parallel corpora (Kupiec, 1993; Feng et al., 2004) , e.g., bilingual Wikipedia entries on the same person, renders high accuracy but suffers from high scarcity. To alleviate such scarcity, (Fung and Yee, 1998; Shao and Ng, 2004 ) explore a more vast resource of comparable corpora, which share no parallel document-or sentence-alignments as in parallel corpora but describe similar contents in two languages, e.g., news articles on the same event. Alternatively, (Lin et al., 2008) extracts bilingual cooccurrences from bilingual sentences, such as annotating terms with their corresponding translations in English inside parentheses. Similarly, (Jiang et al., 2009) identifies potential translation pairs from bilingual sentences using lexical pattern analysis.", |
|
"cite_spans": [ |
|
{ |
|
"start": 166, |
|
"end": 180, |
|
"text": "(Kupiec, 1993;", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 181, |
|
"end": 199, |
|
"text": "Feng et al., 2004)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 338, |
|
"end": 358, |
|
"text": "(Fung and Yee, 1998;", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 359, |
|
"end": 376, |
|
"text": "Shao and Ng, 2004", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 612, |
|
"end": 630, |
|
"text": "(Lin et al., 2008)", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 795, |
|
"end": 815, |
|
"text": "(Jiang et al., 2009)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Corpus-based Approaches", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "The complementary strength of the above two approaches naturally calls for a holistic approach, such as recent work combining transliterationand corpus-based similarity mining bilingual cooccurrences using general search engines. Specifically, (Al-Onaizan and Knight, 2002) uses transliteration to generate candidates and then web corpora to identify translations. Later, (Jiang et al., 2007) enhances to use transliteration to guide web mining.", |
|
"cite_spans": [ |
|
{ |
|
"start": 260, |
|
"end": 273, |
|
"text": "Knight, 2002)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 372, |
|
"end": 392, |
|
"text": "(Jiang et al., 2007)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Holistic Approaches", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "Our work is also a holistic approach, but leveraging significantly larger corpora, specifically by exploiting monolingual co-occurrences. Such expansion enables to translate \"long-tail\" people entities with non-trivial Web occurrences in each monolingual corpus, but not much bilingual co-occurrences. Specifically, we initialize name pair similarity using transliteration-based approach, and iteratively reinforces base similarity using relational similarity.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Holistic Approaches", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "Given two graphs G e = (V e , E e ) and G c = (V c , E c ) harvested from English and Chinese corpora respectively, our goal is to find translation pairs, or a set S of matching node pairs such that S \u2286 V e \u00d7 V c . Let R be a |V e |-by-|V c | matrix where each R ij denotes the similarity between two nodes i \u2208 V e and j \u2208 V c .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Our Framework", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Overall, with the matrix R, our approach consists of the following three steps, as we will discuss in the following three sections respectively: 1. Initialization: computing base translation similarities R ij between two entity nodes using transliteration similarity 2. Reinforcement model: reinforcing the trans-lation similarities R ij by exploiting the monolingual co-occurrences 3. Matching extraction: extracting the matching pairs from the final translation similarities R ij", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Our Framework", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We initialize the translation similarity R ij as the transliteration similarity. This section explains how to get the transliteration similarity between English and Chinese names using an unsupervised approach. Formally, let an English name N e = (e 1 , e 2 , \u2022 \u2022 \u2022 , e n ) and a Chinese name", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "N c = (c 1 , c 2 , \u2022 \u2022 \u2022 , c m ) be given,", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "where e i is an English word and N e is a sequence of the words, and c i is a Chinese character and N c is a sequence of the characters. Our goal is to compute a score indicating the similarity between the pronunciations of the two names.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We first convert N c into its Pinyin representation", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "P Y c = (s 1 , s 2 , \u2022 \u2022 \u2022 , s m ), where s i is the Pinyin rep- resentation of c i .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Pinyin is the romanization representation of pronunciation of Chinese character. For example, the Pinyin representation of N e = (\"Barack\", \"Obama\") is P Y c =(\"ba\", \"la\", \"ke\", \"ao\", \"ba\", \"ma\"). The Pinyin representations of Chinese characters can be easily obtained from Chinese character pronunciation dictionary. In our experiments, we use an in-house dictionary, which contains pronunciations of 20, 774 Chinese characters. For the Chinese characters having multiple pronunciations, we only use the most popular one.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Calculation of transliteration similarity between N e and N c is now transformed to calculation of pronunciation similarity between N e and P Y c . Because letters in Chinese Pinyins and English strings are pronounced similarly, we can further approximate pronunciation similarity between N e and P Y c using their spelling similarity. In this paper, we use Edit Distance (ED) to measure the spelling similarity. Moreover, since words in N e are transliterated into characters in P Y c independently, it is more accurate to compute the ED between N e and P Y c , i.e., ED name (N e , P Y c ), as the sum of the EDs of all component transliteration pairs, i.e., every e i in N e and its corresponding transliteration (s i ) in P Y c . In other words, we need to first align all s j 's in P Y c with corresponding e i in N e based on whether they are translations of each other. Then based on the alignment, we can calculate ED name (N e , P Y c ) using the following formula.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "ED name (N e , P Y c ) = \u2211 i ED(e i , es i ) (1)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "where es i is a string generated by concatenating all s i 's that are aligned to e i and ED(e i , es i ) is the Edit Distance between e i and es i , i.e., the minimum number of edit operations (including insertion, deletion and substitution) needed to transform e i into es i . Because an English word usually consists of multiple syllables but every Chinese character consists of only one syllable, when aligning e i 's with s j 's, we add the constraint that each e i is allowed to be aligned with 0 to 4 s i 's but each s i can only be aligned with 0 to 1 e i . To get the alignment between P Y c and N e which has the minimal ED name (N e , P Y c ), we use a Dynamic Programming based algorithm as defined in the following formula:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "EDname(N 1,i e , P Y 1,j c ) = min( EDname(N 1,i\u22121 e , P Y 1,j c ) + Len(ei), EDname(N 1,i e , P Y 1,j\u22121 c ) + Len(sj), EDname(N 1,i\u22121 e , P Y 1,j\u22121 c ) + ED(ei, sj), EDname(N 1,i\u22121 e , P Y 1,j\u22122 c ) + ED(ei, P Y j\u22121,j c )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": ",", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "EDname(N 1,i\u22121 e , P Y 1,j\u22123 c ) + ED(ei, P Y j\u22122,j c )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": ",", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "EDname(N 1,i\u22121 e , P Y 1,j\u22124 c ) + ED(ei, P Y j\u22123,j c ))", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "where, given a sequence", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "X = (x 1 , x 2 , \u2022 \u2022 \u2022) such that x i is a word, X i,j is the subsequence (x i , x i+1 , \u2022 \u2022 \u2022 , x j ) of X and Len(X)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "is the number of letters except spaces in the sequence X. For instance, the minimal Edit Distance between the English name \"Barack Obama\" and the Chinese Pinyin representation \"ba la ke ao ba ma\" is 4, as the best alignment is: \"Barack\" \u2194 \"ba la ke\" (ED: 3), \"Obama\" \u2194 \"ao ba ma\" (ED: 1). Finally the transliteration similarity between N c and N e is calculated using the following formula.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "Sim tl (N c , N e ) = 1 \u2212 ED name (N e , P Y c ) Len(P Y c ) + Len(N e )", |
|
"eq_num": "(2)" |
|
} |
|
], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "For example, Sim tl (\"Barack Obama\", \"", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "\u2022 \") is 1 \u2212 4 11+12 = 0.826.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Initialization with Transliteration", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "From the initial similarity, we model our problem as an iterative approach that iteratively reinforces the similarity R ij of the nodes i and j from the matching similarities of their neighbor nodes u and v.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "The basic intuition is built on exploiting the similarity between monolingual co-occurrences of two different languages. In particular, we assume two entities with strong relationship co-occur frequently in both corpora. In order to express this intuition, we formally define an iterative reinforcement model as follows. Let R t ij denote the similarity of nodes i and j at t-th iteration:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "R t+1 ij = \u03bb \u2211 (u,v) k \u2208B t (i,j,\u03b8) R t uv 2 k + (1 \u2212 \u03bb)R 0 ij (3)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "The model is expressed as a linear combination of (a) the relational similarity \u2211 R t uv /2 k and (b) transliteration similarity R 0 ij . (\u03bb is the coefficient for interpolating two similarities.)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "In the relational similarity, B t (i, j, \u03b8) is an ordered set of the best matching pairs between neighbor nodes of i and ones of j such that \u2200 ) k is the matching pair with k-th highest similarity score. We consider (u, v) with similarity over some threshold \u03b8, or R t uv \u2265 \u03b8, as a matching pair. In this neighbor matching process, if many-to-many matches exist, we select only one with the greatest matching score. Figure 3 describes such matching process more formally. N (i) and N (j) are the sets of neighbor nodes of i and j, respectively, and H is a priority queue sorting pairs in the decreasing order of similarity scores.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 143, |
|
"end": 144, |
|
"text": ")", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 416, |
|
"end": 424, |
|
"text": "Figure 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "(u, v) k \u2208 B t (i, j, \u03b8), R t uv \u2265 \u03b8, where (u, v", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Meanwhile, note that, in order to express that the confidence for matching (i, j) progressively converges as the number of matched neighbors increases, we empirically use decaying coefficient 1/2 k for R t uv , because", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "\u2211 \u221e k=1 1/2 k = 1.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reinforcement Model", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "After the convergence of the above model, we get the |V e |-by-|V c | similarity matrix R \u221e . From this matrix, we extract one-to-one matches maximizing the overall similarity. More formally, this problem can be stated as the maximum weighted bipartite matching (West, ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "(i, j, \u03b8) \u2190 B t (i, j, \u03b8) \u222a {(u, v)} end if end while return B t (i, j, \u03b8)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "Figure 3: How to get the ordered set B t (i, j, \u03b8) 2000)-Given two groups of entities V e and V c from the two graphs G e and G c , we can build a weighted bipartite graph is", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 41, |
|
"end": 50, |
|
"text": "(i, j, \u03b8)", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "G = (V, E), where V = V e \u222a V c", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "and E is a set of edges (u, v) with weight R \u221e uv . To filter out null alignment, we construct only the edges with weight R \u221e uv \u2265 \u03b4. From this bipartite graph, the maximum weighted bipartite matching problem finds a set of pairwise non-adjacent edges S \u2286 E such that", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "\u2211 (u,v)\u2208S R \u221e", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "uv is the maximum. Wellknown algorithms include Hungarian algorithm with time complexity of O(|V | 2 log |V | + |V ||E|) (West, 2000) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 121, |
|
"end": 133, |
|
"text": "(West, 2000)", |
|
"ref_id": "BIBREF12" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "In this paper, to speed up processing, we consider a greedy alternative with the following steps-(1) choose the pair with the highest similarity score, (2) remove the corresponding row and column from the matrix, and (3) repeat (1) and (2) until their matching scores are over a specific threshold \u03b4.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Matching Extraction", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "This section reports our experimental results to evaluate our proposed approach. First, we report our experimental setting in Section 4.1. Second, we validate the effectiveness and the scalability of our approach over a real-life dataset in Section 4.2.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiments", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "This section describes (1) how we collect the English and Chinese EntityCube datasets, (2) how to build ground-truth test datasets for evaluating our framework, and (3) how to set up three parameters \u03bb, \u03b8, and \u03b4.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Settings", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "First, we crawled G e = (V e , E e ) and G c = (V c , E c ) from English and Chinese EntityCubes. Specifically, we built a graph pairs (G e , G c ) expanding from a \"seed pair\" of nodes s e \u2208 V e and s c \u2208 V c until the number of nodes for each graph becomes 1,000 6 . More specifically, when we build a graph G e by expanding from s e , we use a queue Q. We first initialize Q by pushing the seed node s e . We then iteratively pop a node v e from Q, save v e into V e , and push its neighbor nodes in decreasing order of co-occurrence scores with v e . Similarly, we can get G c from a counterpart seed node v c . By using this procedure, we built six graph pairs from six different seed pairs. In particular, the six seed nodes are English names and its corresponding Chinese names representing a wide range of occupation domains (e.g., 'Barack Obama,' 'Bill Gates,' 'Britney Spears,' 'Bruno Senna,' 'Chris Paul,' and 'Eminem') as Table 1 depicts. Meanwhile, though we demonstrate the effectiveness of the proposed method for mining name translations in Chinese and English languages, the method can be easily adapted to other language pairs. Second, we manually searched for about 50 \"ground-truth\" matched translations for each graph pair to build test datasets T i , by randomly selecting nodes within two hops 7 from the seed pair (s e , s c ), since nodes outside two hops may include nodes whose neighbors are not fully crawled. More specifically, due to our crawling process expanding to add neighbors from the seed, the nodes close to the seed have all the neighbors they would have in the full graph, while those far from the node may not. In order to pick the nodes that well represent the actual neighbors, we built test datasets among those within two hops. However, this crawling is used for the evaluation sake only, and thus does not suggest the bias in our proposed framework. Table 1 describes the size of such test dataset for each graph pair. Lastly, we set up the three parameters \u03bb, \u03b8, and \u03b4 using 6-fold cross validation with 6 test datasets T i 's of the graphs. More specifically, for each dataset T i , we decide \u03bb i and \u03b8 i such that average MRR for the other 5 test datasets is maximized. (About MRR, see more details of Equation (4) in Section 4.2.) We then decide \u03b4 i such that average F1-score is maximized. Figure 4 shows the average MRR for \u03bb i and \u03b8 i with default values \u03b8 = 0.66 and \u03bb = 0.2. Based on these results, we set \u03bb i with values {0.2, 0.15, 0.2, 0.15, 0.2, 0.15} that optimize MRR in datasets T 1 , . . . T 6 , and similarly \u03b8 i with {0.67, 0.65, 0.67, 0.67, 0.65, 0.67}. We also set \u03b4 i with values {0.63, 0.63, 0.61, 0.61, 0.61, 0.61} optimizing F1-score with the same default values \u03bb = 0.2 and \u03b8 = 0.66. We can observe the variances of optimal parameter setting values are low, which suggests the robustness of our framework.", |
|
"cite_spans": [ |
|
{ |
|
"start": 840, |
|
"end": 930, |
|
"text": "'Barack Obama,' 'Bill Gates,' 'Britney Spears,' 'Bruno Senna,' 'Chris Paul,' and 'Eminem')", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 934, |
|
"end": 941, |
|
"text": "Table 1", |
|
"ref_id": "TABREF1" |
|
}, |
|
{ |
|
"start": 1896, |
|
"end": 1965, |
|
"text": "Table 1 describes the size of such test dataset for each graph pair.", |
|
"ref_id": "TABREF1" |
|
}, |
|
{ |
|
"start": 2342, |
|
"end": 2350, |
|
"text": "Figure 4", |
|
"ref_id": "FIGREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experimental Settings", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "This section reports our experimental results using the evaluation datasets explained in previous section. For each graph pair, we evaluated the effectiveness of (1) reinforcement model using MRR measure in Section 4.2.1 and (2) overall framework using precision, recall, and F1 measures in Section 4.2.2. We also validated (3) scalability of our framework over larger scale of graphs (with up to five thousand nodes) in Section 4.2.3. (In all experimental results, Bold numbers indicate the best performance for each metric.)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Results", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "We evaluated the reinforcement model over MRR (Voorhees, 2001) , the average of the reciprocal ranks of the query results as the following formula:", |
|
"cite_spans": [ |
|
{ |
|
"start": 46, |
|
"end": 62, |
|
"text": "(Voorhees, 2001)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Effectiveness of reinforcement model", |
|
"sec_num": "4.2.1" |
|
}, |
|
{ |
|
"text": "MRR = 1 |Q| \u2211 q\u2208Q 1 rank q (4)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Effectiveness of reinforcement model", |
|
"sec_num": "4.2.1" |
|
}, |
|
{ |
|
"text": "Each q is a ground-truth matched pair (u, v) such that u \u2208 V e and v \u2208 V c , and rank q is the rank of the similarity score of R uv among all R uk 's such that k \u2208 V c . Q is a set of such queries. By comparing MRRs for two matrices R 0 and R \u221e , we can validate the effectiveness of the reinforcement model.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Effectiveness of reinforcement model", |
|
"sec_num": "4.2.1" |
|
}, |
|
{ |
|
"text": "\u2022 Baseline matrix (R 0 ): using only the transliteration similarity score, i.e., without reinforcement", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Effectiveness of reinforcement model", |
|
"sec_num": "4.2.1" |
|
}, |
|
{ |
|
"text": "\u2022 Reinforced matrix (R \u221e ): using the reinforced similarity score obtained from Equation 3 We empirically observed that the iterative model converges within 5 iterations. In all experiments, we used 5 iterations for the reinforcement. Table 2 summarizes our experimental results. As these figures show, MRR scores significantly increase after applying our reinforcement model except for the set T 4 (on average from 69% to 81%), which indirectly shows the effectiveness of our reinforcement model.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 235, |
|
"end": 242, |
|
"text": "Table 2", |
|
"ref_id": "TABREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Effectiveness of reinforcement model", |
|
"sec_num": "4.2.1" |
|
}, |
|
{ |
|
"text": "Based on the reinforced matrix, we evaluated the effectiveness of our overall matching framework using the following three measures-(1) precision: how accurately the method returns matching pairs, (2) recall: how many the method returns correct matching pairs, and (3) F1-score: the harmonic mean of precision and recall. We compared our approach with a baseline, mapping two graphs with only transliteration similarity.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Effectiveness of overall framework", |
|
"sec_num": "4.2.2" |
|
}, |
|
{ |
|
"text": "\u2022 Baseline: in matching extraction, using R 0 as the similarity matrix by bypassing the reinforcement step", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Effectiveness of overall framework", |
|
"sec_num": "4.2.2" |
|
}, |
|
{ |
|
"text": "\u2022 Ours: using R \u221e , the similarity matrix converged by Equation (3) In addition, we compared ours with the machine translators of Engkoo and Google. Table 3 summarizes our experimental results.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 149, |
|
"end": 156, |
|
"text": "Table 3", |
|
"ref_id": "TABREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Effectiveness of overall framework", |
|
"sec_num": "4.2.2" |
|
}, |
|
{ |
|
"text": "As this table shows, our approach results in the highest precision (about 80% on average) without compromising the best recall of Google, i.e., 61% of Google vs. 63% of ours. Overall, our approach outperforms others in all three measures.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Effectiveness of overall framework", |
|
"sec_num": "4.2.2" |
|
}, |
|
{ |
|
"text": "Meanwhile, in order to validate the translation accuracy over popular head and long-tail, as discussed in Section 1, we separated the test data into two groups and analyzed the effectiveness separately. Figure 5 plots the number of hits returned for the names from Google search engine. According to the distribution, we separate the test data into top-100 popular people with the highest hits and the remaining, denoted head and tail, respectively. Table 4 shows the effectiveness with both datasets, respectively. As difference of the effectiveness between tail and head (denoted diff ) with respect to three measures shows, our approach shows stably high precision, for both heads and tails.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 203, |
|
"end": 211, |
|
"text": "Figure 5", |
|
"ref_id": "FIGREF6" |
|
}, |
|
{ |
|
"start": 450, |
|
"end": 457, |
|
"text": "Table 4", |
|
"ref_id": "TABREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Effectiveness of overall framework", |
|
"sec_num": "4.2.2" |
|
}, |
|
{ |
|
"text": "To validate the scalability of our approach, we evaluated the effectiveness of our approach over the number of nodes in two graphs. We built larger six graph pairs (G e , G c ) by expanding them from the seed pairs further until the number of nodes reaches 5,000. Figure 6 shows the number of matched translations according to such increase. Overall, the number of matched pairs linearly increases as the number of nodes increases, which suggests scalability. The ratio of node overlap in two graphs is about between 7% and 9% of total node size. ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 264, |
|
"end": 272, |
|
"text": "Figure 6", |
|
"ref_id": "FIGREF8" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Scalability", |
|
"sec_num": "4.2.3" |
|
}, |
|
{ |
|
"text": "This paper abstracted name translation problem as a matching problem of two entity-relationship graphs. This novel approach complements existing name translation work, by not requiring rare resources of parallel or comparable corpus yet outperforming the state-of-the-art. More specifically, we combine bilingual phonetic similarity and monolingual Web co-occurrence similarity, to compute a holistic notion of entity similarity. To achieve this goal, we de- veloped a graph alignment algorithm that iteratively reinforces the matching similarity exploiting relational similarity and then extracts correct matches. Our evaluation results empirically validated the accuracy of our algorithm over real-life datasets, and showed the effectiveness on our proposed perspective.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "http://people.entitycube.com", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "http://renlifang.msra.cn 3 http://www.engkoo.com 4 http://translate.google.com 5 See Section 4 for the sampling process.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Note, this is just a default setting, which we later increase for scalability evaluation inFigure 6.7 Note that the numbers of nodes within two hops in Ge and Gc are 327 and 399 on average respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "This work was supported by Microsoft Research Asia NLP theme funding and MKE (Ministry of Knowledge Economy), Korea, under the ITRC (Information Technology Research Center) support program supervised by the IITA (Institute for Information Technology Advancement) (IITA-2009-C1090-0902-0045).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgments", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Translating Named Entities Using Monolingual and Bilingual Resources", |
|
"authors": [ |
|
{ |
|
"first": "Yaser", |
|
"middle": [], |
|
"last": "Al-Onaizan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kevin", |
|
"middle": [], |
|
"last": "Knight", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of the 40th Annual Meeting on Association for Computational Linguistics (ACL'02)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "400--408", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yaser Al-Onaizan and Kevin Knight. 2002. Trans- lating Named Entities Using Monolingual and Bilin- gual Resources. In Proceedings of the 40th Annual Meeting on Association for Computational Linguistics (ACL'02), pages 400-408. Association for Computa- tional Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "A New Approach for English-Chinese Named Entity Alignment", |
|
"authors": [ |
|
{ |
|
"first": "Donghui", |
|
"middle": [], |
|
"last": "Feng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yajuan", |
|
"middle": [], |
|
"last": "L\u00fc", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP'04)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "372--379", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Donghui Feng, Yajuan L\u00fc, and Ming Zhou. 2004. A New Approach for English-Chinese Named En- tity Alignment. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP'04), pages 372-379. Association for Com- putational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "An IR Approach for Translating New Words from Nonparallel,Comparable Texts", |
|
"authors": [ |
|
{ |
|
"first": "Pascale", |
|
"middle": [], |
|
"last": "Fung", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yee", |
|
"middle": [], |
|
"last": "Lo Yuen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the 17th International Conference on Computational Linguistics (COLING'98)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "414--420", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Pascale Fung and Lo Yuen Yee. 1998. An IR Ap- proach for Translating New Words from Nonparal- lel,Comparable Texts. In Proceedings of the 17th In- ternational Conference on Computational Linguistics (COLING'98), pages 414-420. Association for Com- putational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Named Entity Translation with Web Mining and Transliteration", |
|
"authors": [ |
|
{ |
|
"first": "Long", |
|
"middle": [], |
|
"last": "Jiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Cheng", |
|
"middle": [], |
|
"last": "Lee Feng Chien", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Niu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of the 20th International Joint Conference on Artificial Intelligence (IJ-CAI'07)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1629--1634", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Long Jiang, Ming Zhou, Lee feng Chien, and Cheng Niu. 2007. Named Entity Translation with Web Mining and Transliteration. In Proceedings of the 20th Interna- tional Joint Conference on Artificial Intelligence (IJ- CAI'07), pages 1629-1634. Morgan Kaufmann Pub- lishers Inc.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Mining Bilingual Data from the Web with Adaptively Learnt Patterns", |
|
"authors": [ |
|
{ |
|
"first": "Long", |
|
"middle": [], |
|
"last": "Jiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Shiquan", |
|
"middle": [], |
|
"last": "Yang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaohua", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Qingsheng", |
|
"middle": [], |
|
"last": "Zhu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 47th Annual Meeting of the Association for Computational Linguistics (ACL'09)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "870--878", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Long Jiang, Shiquan Yang, Ming Zhou, Xiaohua Liu, and Qingsheng Zhu. 2009. Mining Bilingual Data from the Web with Adaptively Learnt Patterns. In Proceed- ings of the 47th Annual Meeting of the Association for Computational Linguistics (ACL'09), pages 870-878. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "An Algorithm for finding Noun Phrase Correspondences in Bilingual Corpora", |
|
"authors": [ |
|
{ |
|
"first": "Julian", |
|
"middle": [], |
|
"last": "Kupiec", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1993, |
|
"venue": "Proceedings of the 31th Annual Meeting of the Association for Computational Linguistics (ACL'93)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "17--22", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Julian Kupiec. 1993. An Algorithm for finding Noun Phrase Correspondences in Bilingual Corpora. In Pro- ceedings of the 31th Annual Meeting of the Association for Computational Linguistics (ACL'93), pages 17-22. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "A Joint Source-Channel Model for Machine Transliteration", |
|
"authors": [ |
|
{ |
|
"first": "Haizhou", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Zhang", |
|
"middle": [], |
|
"last": "Min", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Su", |
|
"middle": [], |
|
"last": "Jian", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics (ACL'04)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "159--166", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Haizhou Li, Zhang Min, and Su Jian. 2004. A Joint Source-Channel Model for Machine Transliteration. In Proceedings of the 42nd Annual Meeting on Associ- ation for Computational Linguistics (ACL'04), pages 159-166. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Mining Parenthetical Translations from the Web by Word Alignment", |
|
"authors": [ |
|
{ |
|
"first": "Dekang", |
|
"middle": [], |
|
"last": "Lin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Shaojun", |
|
"middle": [], |
|
"last": "Zhao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Benjamin", |
|
"middle": [], |
|
"last": "Van Durme", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Marius", |
|
"middle": [], |
|
"last": "Pasca", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of the 46th Annual Meeting of the Association for Computational Linguistics (ACL'08)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "994--1002", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dekang Lin, Shaojun Zhao, Benjamin Van Durme, and Marius Pasca. 2008. Mining Parenthetical Transla- tions from the Web by Word Alignment. In Proceed- ings of the 46th Annual Meeting of the Association for Computational Linguistics (ACL'08), pages 994- 1002. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Mining New Word Translations from Comparable Corpora", |
|
"authors": [ |
|
{ |
|
"first": "Li", |
|
"middle": [], |
|
"last": "Shao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hwee Tou", |
|
"middle": [], |
|
"last": "Ng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the 20th International Conference on Computational Linguistics (COLING'04)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "618--624", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Li Shao and Hwee Tou Ng. 2004. Mining New Word Translations from Comparable Corpora. In Proceed- ings of the 20th International Conference on Computa- tional Linguistics (COLING'04), pages 618-624. As- sociation for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "The trec question answering track", |
|
"authors": [ |
|
{ |
|
"first": "Ellen", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Voorhees", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Natural Language Engineering", |
|
"volume": "7", |
|
"issue": "4", |
|
"pages": "361--378", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ellen M. Voorhees. 2001. The trec question answering track. Natural Language Engineering, 7(4):361-378.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Automatic English-Chinese Name Transliteration for Development of Multilingual Resources", |
|
"authors": [ |
|
{ |
|
"first": "Stephen", |
|
"middle": [], |
|
"last": "Wan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Cornelia", |
|
"middle": [ |
|
"Maria" |
|
], |
|
"last": "Verspoor", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "Proceedings of the 17th International Conference on Computational Linguistics (COLING'98)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1352--1356", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Stephen Wan and Cornelia Maria Verspoor. 1998. Auto- matic English-Chinese Name Transliteration for De- velopment of Multilingual Resources. In Proceed- ings of the 17th International Conference on Compu- tational Linguistics (COLING'98), pages 1352-1356. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Introduction to Graph Theory", |
|
"authors": [ |
|
{ |
|
"first": "Douglas", |
|
"middle": [], |
|
"last": "Brent", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "West", |
|
"middle": [], |
|
"last": "", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Douglas Brent West. 2000. Introduction to Graph The- ory. Prentice Hall, second edition.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Comparison for Head and Tail datasets", |
|
"uris": null |
|
}, |
|
"FIGREF1": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "(a) English PeopleEntityCube G e (b) Chinese Renlifang G c (c) Abstracting translation as graph mapping", |
|
"uris": null |
|
}, |
|
"FIGREF2": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Illustration of entity-relationship graphs", |
|
"uris": null |
|
}, |
|
"FIGREF4": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Parameter setup for \u03bb, \u03b8, and \u03b4", |
|
"uris": null |
|
}, |
|
"FIGREF6": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Distribution over number of hits", |
|
"uris": null |
|
}, |
|
"FIGREF8": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Matched translations over |V e | and |V c |", |
|
"uris": null |
|
}, |
|
"TABREF1": { |
|
"content": "<table><tr><td colspan=\"4\">i |Ve|, |Vc| |Ti| English Name</td><td>Chinese Name</td></tr><tr><td>1</td><td>1,000</td><td>51</td><td>Barack Obama</td><td>\u2022</td></tr><tr><td>2</td><td>1,000</td><td>52</td><td>Bill Gates</td><td>\u2022</td></tr><tr><td>3</td><td>1,000</td><td>40</td><td>Britney Spears</td><td>\u2022</td></tr><tr><td>4</td><td>1,000</td><td>53</td><td>Bruno Senna</td><td>\u2022</td></tr><tr><td>5</td><td>1,000</td><td>51</td><td>Chris Paul</td><td>\u2022</td></tr><tr><td>6</td><td>1,000</td><td>57</td><td>Eminem</td><td/></tr></table>", |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"text": "Summary for graphs and test datasets obtained from each seed pair" |
|
}, |
|
"TABREF2": { |
|
"content": "<table><tr><td colspan=\"3\">: MRR of baseline and reinforced matrices</td></tr><tr><td>Set</td><td colspan=\"2\">MRR Baseline R 0 Reinforced R \u221e</td></tr><tr><td>T 1</td><td>0.6964</td><td>0.8377</td></tr><tr><td>T 2</td><td>0.6213</td><td>0.7581</td></tr><tr><td>T 3</td><td>0.7095</td><td>0.7989</td></tr><tr><td>T 4</td><td>0.8159</td><td>0.8378</td></tr><tr><td>T 5</td><td>0.6984</td><td>0.8158</td></tr><tr><td>T 6</td><td>0.5982</td><td>0.8011</td></tr><tr><td>Average</td><td>0.6900</td><td>0.8082</td></tr></table>", |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"text": "" |
|
}, |
|
"TABREF3": { |
|
"content": "<table><tr><td>Set</td><td colspan=\"3\">Precision Engkoo Google Baseline</td><td>Ours</td><td>Engkoo</td><td colspan=\"2\">Recall Google Baseline</td><td>Ours</td><td colspan=\"3\">F1-score Engkoo Google Baseline</td><td>Ours</td></tr><tr><td>T 1</td><td>0.5263</td><td>0.4510</td><td>0.5263</td><td>0.8974</td><td>0.3922</td><td>0.4510</td><td>0.1961</td><td>0.6863</td><td>0.4494</td><td>0.4510</td><td>0.2857</td><td>0.7778</td></tr><tr><td>T 2</td><td>0.7551</td><td>0.75</td><td>0.7143</td><td>0.8056</td><td>0.7115</td><td>0.75</td><td>0.2885</td><td>0.5577</td><td>0.7327</td><td>0.75</td><td>0.4110</td><td>0.6591</td></tr><tr><td>T 3</td><td>0.5833</td><td>0.7925</td><td>0.5556</td><td>0.7949</td><td>0.5283</td><td>0.7925</td><td>0.1887</td><td>0.5849</td><td>0.5545</td><td>0.7925</td><td>0.2817</td><td>0.6739</td></tr><tr><td>T 4</td><td>0.5</td><td>0.45</td><td>0.7368</td><td>0.7353</td><td>0.425</td><td>0.45</td><td>0.35</td><td>0.625</td><td>0.4595</td><td>0.45</td><td>0.4746</td><td>0.6757</td></tr><tr><td>T 5</td><td>0.6111</td><td>0.3137</td><td>0.5</td><td>0.7234</td><td>0.4314</td><td>0.3137</td><td>0.1765</td><td>0.6667</td><td>0.5057</td><td>0.3137</td><td>0.2609</td><td>0.6939</td></tr><tr><td>T 6</td><td>0.5636</td><td>0.8947</td><td>0.6</td><td>0.8605</td><td>0.5438</td><td>0.8947</td><td>0.1053</td><td>0.6491</td><td>0.5536</td><td>0.8947</td><td>0.1791</td><td>0.74</td></tr><tr><td>AVG</td><td>0.5899</td><td>0.6086</td><td>0.6055</td><td>0.8028</td><td>0.5054</td><td>0.6086</td><td>0.2175</td><td>0.6283</td><td>0.5426</td><td>0.6086</td><td>0.3155</td><td>0.7034</td></tr></table>", |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"text": "Precision, Recall, and F1-score of Baseline, Engkoo, Google, and Ours over test sets T i" |
|
}, |
|
"TABREF4": { |
|
"content": "<table><tr><td>Method</td><td>head</td><td>Precision tail</td><td>diff</td><td>head</td><td>Recall tail</td><td>diff</td><td>head</td><td>F1-score tail</td><td>diff</td></tr><tr><td colspan=\"10\">Engkoo 0.6082 0.5854 0.0229 0.59 0.4706 0.1194 0.5990 0.5217 0.0772</td></tr><tr><td>Google</td><td>0.75</td><td colspan=\"5\">0.5588 0.1912 0.75 0.5588 0.1912</td><td>0.75</td><td colspan=\"2\">0.5588 0.1912</td></tr><tr><td>Ours</td><td colspan=\"9\">0.8462 0.7812 0.0649 0.66 0.6127 0.0473 0.7416 0.6868 0.0548</td></tr></table>", |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"text": "Precision, Recall, and F1-score of Engkoo, Google, and Ours with head and tail datasets" |
|
} |
|
} |
|
} |
|
} |