|
{ |
|
"paper_id": "U15-1019", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T03:10:04.935803Z" |
|
}, |
|
"title": "Word Transformation Heuristics Against Lexicons for Cognate Detection", |
|
"authors": [ |
|
{ |
|
"first": "Alexandra", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Uitdenbogerd", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "RMIT University GPO", |
|
"location": { |
|
"postBox": "Box 2476", |
|
"postCode": "3001", |
|
"settlement": "Melbourne", |
|
"region": "VIC", |
|
"country": "Australia" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "One of the most common lexical transformations between cognates in French and English is the presence or absence of a terminal \"e\". However, many other transformations exist, such as a vowel with a circumflex corresponding to the vowel and the letter s. Our algorithms tested the effectiveness of taking the entire English and French lexicons from Treetagger, deaccenting the French lexicon, and taking the intersection of the two. Words shorter than 6 letters were excluded from the list, and a set of lexical transformations were also used prior to intersecting, to increase the potential pool of cognates. The result was 15% above the baseline cognate list in the initial test set, but only 1% above it in the final test set. However, its accuracy was consistant at about 37% for both test sets.", |
|
"pdf_parse": { |
|
"paper_id": "U15-1019", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "One of the most common lexical transformations between cognates in French and English is the presence or absence of a terminal \"e\". However, many other transformations exist, such as a vowel with a circumflex corresponding to the vowel and the letter s. Our algorithms tested the effectiveness of taking the entire English and French lexicons from Treetagger, deaccenting the French lexicon, and taking the intersection of the two. Words shorter than 6 letters were excluded from the list, and a set of lexical transformations were also used prior to intersecting, to increase the potential pool of cognates. The result was 15% above the baseline cognate list in the initial test set, but only 1% above it in the final test set. However, its accuracy was consistant at about 37% for both test sets.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "When assessing readability of English for French native speakers, or French for English native speakers, the cognates -words with similar appearance and meaning -tend to be relatively difficult words, making traditional readability measures less effective than a simple average words per sentence ( Uitdenbogerd, 2005) . While most words that look similar in French and English are cognates, some commonly occurring words that look similar, such as \"a\", \"as\", and \"an\", tend to be false friends. Other words are partial cognates, having a similar meaning only in some situations (Wang and Sitbon, 2014) . Our approach ignored all context and focussed on simple heuristics. All approaches were based on taking the intersection of the French and English lexicons of the Treetagger Part of Speech tagger (Schmid, 1994) , after applying a set of lexical transformations. The submission was a \"quick and dirty hack\" implemented on little sleep during conference attendance, and merely demonstrates that a simple heuristic-based algorithm can beat a manually curated list of cognates, albeit not by much. However, the approach should perform better than demonstrated in the ALTW challenge if applied more comprehensively.", |
|
"cite_spans": [ |
|
{ |
|
"start": 297, |
|
"end": 318, |
|
"text": "( Uitdenbogerd, 2005)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 579, |
|
"end": 602, |
|
"text": "(Wang and Sitbon, 2014)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 801, |
|
"end": 815, |
|
"text": "(Schmid, 1994)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Credits 2 Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The first \"quick and dirty\" baseline algorithm (Algorithm 1) only looks for exact matches once case and accents are preprocessed:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "1. Replace the accented letters in the French lexicon with unaccented letters. For example, replace \"\u00ea\" and \"\u00e9\" with \"e\".", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "2. Casefold and remove punctuation from the words in the source and cognate file.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "3. Take the intersection of the sanitised source and cognate file words.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "4. All words in the text that are in the intersection are counted as cognates.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Algorithm 2 discards any Algorithm 1 words of 5 letters or less as false friends. Common false friend English words that become discarded as a result include: \"a\", \"as\", \"an\". However, the following cognates are also discarded: \"ah\", \"oh\".", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Algorithm 3 uses lexical transformations to the French lexicon list before intersecting with the English lexicon. It is done with and without plurals. The list is based on observation on reading a French text. While there are many transformation rules, they are not comprehensive. In particular, words in which multiple transformations are required to create an exact match are likely to be missed. Figure 1 shows the regular expressionbased substitutions applied to the file. Algorithm 4 combines Algorithm 3's lexical transformations with discarding words that are 5 letters or fewer in length.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 399, |
|
"end": 407, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We also tried stemming but the result was half the accuracy of the original baseline. The final submission used the transformations as well as discarding words of 5 letters or less. Table 1 shows the precision, recall and F1 measure for the training data set. Table 2 shows the overall results for the ALTW challenge. As can be seen, our entry (toe in) had the most consistent performance across the two test sets. Of our submissions, Algorithm 3 performed the best on the public test dataset. The best private data submission was a version of Algorithm 3 that didn't discard short words.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 182, |
|
"end": 189, |
|
"text": "Table 1", |
|
"ref_id": "TABREF0" |
|
}, |
|
{ |
|
"start": 260, |
|
"end": 267, |
|
"text": "Table 2", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Algorithms", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In a post-analysis using the training data set we looked at the effect of varying the minimum word length for cognates, holding the base word list constant. Table 3 shows the effect on precision, recall and F measure. Precision increases as the minimum word length is increased, and recall decreases. The sweet spot in the training data set is to discard words that are 4 letters long or less. ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 157, |
|
"end": 164, |
|
"text": "Table 3", |
|
"ref_id": "TABREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "The experimental results demonstrated that a lexically transformed French lexicon intersected with an English lexicon with the shortest words discarded can be a substitute for a manually curated list of cognates, achieving 1 to 15% higher accuracy on the given test sets. A more comprehensive set of lexical transformations is likely to give a slightly higher accuracy again. However, as the ALTW challenge results demonstrate, this context-free, heuristic approach has only about half the accuracy of the best technique.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "grep \"e$\" $1 | sed \"s/e$//\" grep \"ait$\" $1 | sed \"s/ait$/ed/\" grep \"aient$\" $1 | sed \"s/aient$/ed/\" grep \"gue$\" $1 | sed \"s/gue/g/\" grep \"\u00e9$\" $1 | sed \"s/\u00e9$/y/\" grep \"euse$\" $1 | sed \"s/euse/ous/\" grep \"eux$\" $1 | sed \"s/eux/ous/\" grep \"ique$\" $1 | sed \"s/ique/ic/\" grep \"\u02c6d\u00e9$\" $1 | sed \"s/\u02c6d\u00e9/dis/\" grep \"ont$\" $1 | sed \"s/ont$/ount/\" grep \"ond$\" $1 | sed \"s/ond$/ound/\" grep \"ant$\" $1 | sed \"s/ant$/ing/\" grep \"ain$\" $1 | sed \"s/ain$/an/\" grep \"aine$\" $1 | sed \"s/aine/an/\" grep \"re$\" $1 | sed \"s/re$/er/\" grep \"ment$\" $1 | sed \"s/ment$/ly/\" grep \"\u00e9$\" $1 | sed \"s/\u00e9$/ated/\" grep \"\u00e9$\" $1 | sed \"s/\u00e9$/ed/\" grep \"\u00e9e$\" $1 | sed \"s/\u00e9e$/ated/\" grep \"\u00e9e$\" $1 | sed \"s/\u00e9e$/ed/\" grep \"i$\" $1 | sed \"s/i$/ished/\" grep \"ir$\" $1 | sed \"s/ir$/ish/\" grep \"er$\" $1 | sed \"s/er$/e/\" grep \"\u00f4\" $1 | sed \"s/\u00f4/os/\" grep \"\u00ea\" $1 | sed \"s/\u00ea/es/\" grep \"\u00ee\" $1 | sed \"s/\u00ee/is/\" grep \"ement$\" $1 | sed \"s/ement$/ly/\" grep \"eusement$ $1| sed \"s/eusement$/ously/\" grep \"isme$\" $1 | sed \"s/isme$/ism/\" grep \"if$\" $1| sed \"s/if$/ive/\" grep \"asse$\" $1 | sed \"s/asse$/ace/\" grep \"eur$\" $1 | sed \"s/eur$/or/\" grep \"eur$\" $1 | sed \"s/eur$/er/\" grep \"eur$\" $1 | sed \"s/eur$/our/\" grep \"\u02c6\u00e9\" $1 | sed \"s/\u02c6\u00e9/es/\" grep \"\u02c6\u00e9\" $1 | sed \"s/\u02c6\u00e9/s/\" grep \"o\u00fbt\" $1 | sed \"s/o\u00fbt/ost/\" grep \"\u02c6av\" $1 | sed \"s/\u02c6av/adv\" grep \"\u02c6aj\" $1 | sed \"s/\u02c6aj/adj\" grep \"elle$ $1 | sed \"s/elle$/al/\" grep \"ette$\" $1 | sed \"s/ette$/et/\" grep \"onne$\" $1 | sed \"s/onne$/on/\" grep \"quer$\" $1 | sed \"s/quer$/cate/\" grep \"ai\" $1 | sed \"s/ai/ea/\" grep \"\u02c6en\" $1 | sed \"s/\u02c6en/in/\" grep \"ier$\" $1 | sed \"s/ier$/er/\" Figure 1 : The set of lexical transformations applied to the French lexicon prior to intersection with the English lexicon. \"$1\" is the file containing the French lexicon.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 1542, |
|
"end": 1550, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "5" |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "Some text processing code was written by Aidan Martin.Support for this project has been provided by the Australian Government Office for Learning and Teaching. The views in this project do not necessarily reflect the views of the Australian Government Office for Learning and Teaching.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgments", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Readability of French as a foreign language and its uses", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Uitdenbogerd", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Australasian Document Computing Symposium", |
|
"volume": "10", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "A. L. Uitdenbogerd. 2005. Readability of French as a foreign language and its uses. In A. Turpin and R. Wilkinson, editors, Australasian Document Com- puting Symposium, volume 10, December.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Probabilitic part-of-speech tagging using decision trees", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Schmid", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "International Conference on New Methods in Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "H. Schmid. 1994. Probabilitic part-of-speech tagging using decision trees. In International Conference on New Methods in Language Processing.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Multilingual lexical resources to detect cognates in non-aligned texts", |
|
"authors": [ |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "L", |
|
"middle": [], |
|
"last": "Sitbon", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the Australasian Language Technology Association Workshop", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "14--22", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "H. Wang and L. Sitbon. 2014. Multilingual lexical resources to detect cognates in non-aligned texts. In G. Ferraro and S. Wan, editors, Proceedings of the Australasian Language Technology Association Workshop, pages 14-22, Melbourne, Victoria, Aus- tralia, November. ALTA.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF0": { |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"content": "<table><tr><td/><td colspan=\"2\">Actual Algorithm 3</td></tr><tr><td>Cognates</td><td>1670</td><td>703</td></tr><tr><td>Non-Cognates</td><td>9425</td><td>10392</td></tr><tr><td>Total Words</td><td>11095</td><td/></tr><tr><td>Proportion of Cognates</td><td>0.150</td><td>0.063</td></tr><tr><td>Precision</td><td/><td>0.661</td></tr><tr><td>Recall</td><td/><td>0.278</td></tr><tr><td>F1</td><td/><td>0.390</td></tr></table>", |
|
"text": "Training Set Statistics" |
|
}, |
|
"TABREF1": { |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"content": "<table><tr><td colspan=\"3\">: ALTW Challenge 2015 Results</td></tr><tr><td>Team</td><td>Public</td><td>Private</td></tr><tr><td>LookForward</td><td colspan=\"2\">0.70478 0.77001</td></tr><tr><td>Little Monkey</td><td colspan=\"2\">0.67118 0.71415</td></tr><tr><td>MAC</td><td colspan=\"2\">0.59927 0.66857</td></tr><tr><td colspan=\"3\">toe in (Alg. 4 lex.tr.-shrt) 0.37019 0.36697</td></tr><tr><td>Alg. 3 (lex. trans.)</td><td colspan=\"2\">0.31394 0.37272</td></tr><tr><td colspan=\"3\">Alg. 2 (Exact -shrtwrds) 0.23583 0.23908</td></tr><tr><td>Baseline</td><td colspan=\"2\">0.22951 0.34158</td></tr><tr><td>Alg. 1 (Exact Match)</td><td colspan=\"2\">0.22107 0.27406</td></tr><tr><td>Stemmed lexicon match</td><td colspan=\"2\">0.11347 0.14116</td></tr></table>", |
|
"text": "" |
|
}, |
|
"TABREF2": { |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"content": "<table><tr><td colspan=\"4\">Min Length Precision Recall F measure</td></tr><tr><td>3</td><td>.457</td><td>.346</td><td>.393</td></tr><tr><td>4</td><td>.457</td><td>.346</td><td>.393</td></tr><tr><td>5</td><td>.579</td><td>.323</td><td>.414</td></tr><tr><td>6</td><td>.658</td><td>.266</td><td>.378</td></tr><tr><td>7</td><td>.711</td><td>.198</td><td>.309</td></tr></table>", |
|
"text": "The effect of minimum word length on cognate detection reliability" |
|
} |
|
} |
|
} |
|
} |