|
{ |
|
"paper_id": "2022", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T12:11:28.690164Z" |
|
}, |
|
"title": "Using Cross-Lingual Part of Speech Tagging for Partially Reconstructing the Classic Language Family Tree Model", |
|
"authors": [ |
|
{ |
|
"first": "Anat", |
|
"middle": [], |
|
"last": "Samohi", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Daniel", |
|
"middle": [ |
|
"Weisberg" |
|
], |
|
"last": "Mitelman", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Kfir", |
|
"middle": [], |
|
"last": "Bar", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "The tree model is well known for expressing the historic evolution of languages. This model has been considered as a method of describing genetic relationships between languages. Nevertheless, some researchers question the model's ability to predict the proximity between two languages, since it represents genetic relatedness rather than linguistic resemblance. Defining other language proximity models has been an active research area for many years. In this paper we explore a part-of-speech model for defining proximity between languages using a multilingual language model that was fine-tuned on the task of cross-lingual part-of-speech tagging. We train the model on one language and evaluate it on another; the measured performance is then used to define the proximity between the two languages. By further developing the model, we show that it can reconstruct some parts of the tree model.", |
|
"pdf_parse": { |
|
"paper_id": "2022", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "The tree model is well known for expressing the historic evolution of languages. This model has been considered as a method of describing genetic relationships between languages. Nevertheless, some researchers question the model's ability to predict the proximity between two languages, since it represents genetic relatedness rather than linguistic resemblance. Defining other language proximity models has been an active research area for many years. In this paper we explore a part-of-speech model for defining proximity between languages using a multilingual language model that was fine-tuned on the task of cross-lingual part-of-speech tagging. We train the model on one language and evaluate it on another; the measured performance is then used to define the proximity between the two languages. By further developing the model, we show that it can reconstruct some parts of the tree model.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Language families are defined by the evolution of languages over the history, providing indications regarding the proximity between them. The tree model, which was first introduced by Augustus Schleicher (Schleicher, 1853) is considered as the consensual language-family model. For example, Figure 1 shows the Indo-European branch of the tree model; a full version of the model is nicely presented on Ethnologue 1 . In this paper, we refer to this source as a reference for the classic family tree model. Concomitantly, there have been theories that question the tree model as being an indicator for language proximity, since it represents genetic relatedness rather than lexical resemblance. Loanwords, as well as other lexical influences are usually not expressed in the classic tree model. Representing historical relatedness, the tree is agnostic to various linguistic influences. Consequently, some claim that language families should be defined by alternative models (Geisler and List, 2013) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 973, |
|
"end": 997, |
|
"text": "(Geisler and List, 2013)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 291, |
|
"end": 299, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The Universal Grammar, introduced by Noam Chomsky, is usually defined as the \"system of categories, mechanisms and constraints shared by all human languages and considered to be innate\" (Dobrovolsky et al., 2016) . In other words, a human language is derived from a set of structural rules, typically referred to as generative grammar, which we are usually totally unaware of. We can intuitively distinguish between nouns and verbs; children can phrase a sentence they have not heard before by ordering parts of speech they are familiar with in a valid grammatical order. A child can identify a noun without knowing what a noun is, or without even understanding the meaning of that specific noun. It may be assumed that rather than this aspect of universal grammar being specific to language, it is more generally a part of human cognition, and there might be a common structure for different languages. Still, the ability to classify words into parts of speech requires some knowledge of the structure of the specific language.", |
|
"cite_spans": [ |
|
{ |
|
"start": 186, |
|
"end": 212, |
|
"text": "(Dobrovolsky et al., 2016)", |
|
"ref_id": "BIBREF8" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The hypothesis we examine in this paper relies on the assumption that historically close languages, like French and Spanish, share some information that may help the classification of words into partof-speech (POS) tags. While identifying this type of information is out of scope for this paper, we will show that this information can be used by a neural network for predicting POS tags of one language only using examples from another language.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Our goal is to redefine the proximity between languages to achieve a comparable model to the classic tree model, by considering only POS tags. The Indo-European language branch; the graph was created by the igraph package for Python (Csardi and Nepusz, 2006) . The languages are represented using their equivalent two-letter ISO 639-1 code.", |
|
"cite_spans": [ |
|
{ |
|
"start": 233, |
|
"end": 258, |
|
"text": "(Csardi and Nepusz, 2006)", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "To enable transferability between languages, we suggest using a multilingual pre-trained language model (MPLM), fine-tuned for the POS tagging task in a multilingual environment. Specifically, we take a multilingual zero-shot training approach by fine-tuning an MPLM to predict POS tags for texts written in one language, the source language, and evaluating it on texts written in another language, the target language. The performance metrics are then used to estimate the similarity between the source and target languages. As a final step, we generate a language-similarity graph, which we describe as an approximation for the classic tree model. We make two main contributions: (1) Reconstructing part of the classic tree model using POS-based similarity scores; and, (2) Providing some insights into the cross-lingual generalization of MPLMs.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "We proceed as follows: In Section 2 we cite some related studies, following by a detailed description of our method, provided in Section 3. We end Section 3 with reporting on some results. We discuss the results in Section 4 and make some conclusions.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "There have been some prior studies on measuring distance between languages. In their paper, Chiswick and Miller (2005) presented some empirical observations of how rapidly speakers of a given language gained proficiency in another tongue. Specifically, they measured the speed of English acquisition by immigrants of various linguistic backgrounds in the United States and Canada. Their first languages were ranked for the distance from English, on a scale from 1.0 (very different than English) to 3.0 (closest to English). It has been found empirically that the greater the distance between an immigrant's origin language and English, the lower is the level of the immigrant's English language proficiency.", |
|
"cite_spans": [ |
|
{ |
|
"start": 92, |
|
"end": 118, |
|
"text": "Chiswick and Miller (2005)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "There have been many attempts to use computational tools to infer the relations between languages; the dominant approach is known as phylogenetic linguistics. Phylogenetic linguistics is about establishing historical relationships among languages, by considering the evolutionary nature of human languages. In computational phylogenetics, words and/or phonemes of what counts as the same language over time, are analyzed and compared among languages.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Specifically, Swadesh (1950) was first to introduce a computational phylogenetic technique called lexicostatistics for comparing between two languages. In lexicostatistics, the similarity between two languages is calculated by a function of the percentage of cognates found in a predefined list of words of the two languages. Swadesh's work has been followed by a number of studies that use lexicostatistics or a minor variation of it (Nakhleh et al., 2005; Bakker et al., 2009; Petroni and Serva, 2010; Barban\u00e7on et al., 2013) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 14, |
|
"end": 28, |
|
"text": "Swadesh (1950)", |
|
"ref_id": "BIBREF23" |
|
}, |
|
{ |
|
"start": 435, |
|
"end": 457, |
|
"text": "(Nakhleh et al., 2005;", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 458, |
|
"end": 478, |
|
"text": "Bakker et al., 2009;", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 479, |
|
"end": 503, |
|
"text": "Petroni and Serva, 2010;", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 504, |
|
"end": 527, |
|
"text": "Barban\u00e7on et al., 2013)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Instead of measuring the percentage of cognates, Petroni and Serva proposed to calculate a normalized Levenstein distance among words with the same meaning and then to take the average over the words contained in a cross-lingual list. M\u00fcller et al. (2010) conducted a lexical comparison using the Levenstein distance approach, between 4,350 languages of the ASJP database , and created a full diagram of lexical proximity. They showed that lexical resemblance is related to genetic affiliation. However, some of the languages that have been found as lexically similar, according to their technique, are not closely genetically associated.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Another computational approach for measuring language similarity is based on corpus analysis. Gamallo et al. (2017) used the known per-plexity score of a probabilistic n-gram language model to measure the distance between European languages. Asgari and Mofrad (2016) compared 50 languages from different families by training a monolingual language model on each language individually, using a parallel corpus of the Bible (Christodouloupoulos and Steedman, 2015) , and apply them to calculate perplexity on all the other languages. In some of the works that are mentioned above, the proximity between languages is not perfectly aligned with the classic tree model.", |
|
"cite_spans": [ |
|
{ |
|
"start": 94, |
|
"end": 115, |
|
"text": "Gamallo et al. (2017)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 422, |
|
"end": 462, |
|
"text": "(Christodouloupoulos and Steedman, 2015)", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "While the main focus has always been on lexical similarity, some attempts were made to compare languages on the syntactic level. Longobardi and Guardiano (2009) characterized 28 languages, mostly Indo-European ones, using a set of 63 predefined morpho-syntactic parameters. They calculated a normalized Hamming distance over those parametric representations, with which they were able to generate a language tree. They showed that this tree is equivalent to a tree that was generated based on a traditional lexicostatistics approach, suggesting that syntactic characteristics are sufficiently robust to reconstruct a plausible historical language tree. The same method was re-used in (Longobardi et al., 2013) , which was concluded in a similar way. In a recent work, Shu et al. (2021) applied a different comparison technique on the same syntactic characteristics, using Markov models. In all of those works, the selection of the syntactic characteristics to be used for comparison, plays an important role in the creation of a language proximity model.", |
|
"cite_spans": [ |
|
{ |
|
"start": 129, |
|
"end": 160, |
|
"text": "Longobardi and Guardiano (2009)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 684, |
|
"end": 709, |
|
"text": "(Longobardi et al., 2013)", |
|
"ref_id": "BIBREF15" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "To the best of our knowledge, there have not been attempts to compare languages using syntactic information in a non-parametric way. In this work, we take a corpus-based approach to automatically extract part-of-speech tags from a given text in order to generate a language-proximity model. In that sense, we consider our approach as a nonparametric estimation method, since we do not need to manually define specific syntactic parameters to consider for calculating similarity between languages.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "To transfer information across languages, we use mBERT, a multilingual version of BERT (Devlin et al., 2019) , that was pre-trained on texts written in over 100 languages based on a shared vocabulary. 2 During pre-training, the training documents are given to mBERT without any indication on the language that they have been written with. Like every other pre-trained language model (PLM), the pre-trained mBERT model is typically fine-tuned on a training set of a specific downstream task, which could be either monolingual or multilingual. This unique multilingual design allows mBERT to handle multilingual tasks in a transfer-learning way. In another study, Wu and Dredze (2019) reported an impressive performance using mBERT in a zero-shot cross-lingual transfer learning setting on several NLP tasks, including POS tagging. They claimed that mBERT may learn a cross-lingual representation by generalizing and abstracting some language-specific information. A similar observation was made by Gonen et al. (2020) who claimed that mBERT learns information by two components, one that encodes the language and another that encodes some abstract information that can be used in a cross-lingual way.", |
|
"cite_spans": [ |
|
{ |
|
"start": 87, |
|
"end": 108, |
|
"text": "(Devlin et al., 2019)", |
|
"ref_id": "BIBREF7" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "For every pair of languages, source language and target language, we measure their similarity as the performance of an mBERT-based POS tagger finetuned on the source language, and evaluated on the target language. For training and evaluation, we use treebanks from Universal Dependencies (UD). 3 In particular, we use the Universal POS labels 4 assigned for every syntactic word in the text. The Universal POS tagset contains the following core part-of-speech categories that can be used for any UD language: adjective, adposition, adverb, auxiliary, coordinating conjunction, determiner, interjection, noun, numeral, particle, pronoun, proper noun, punctuation, subordinating conjunction, symbol, verb and other. Each treebank is divided to train and test sets. Therefore, we fine-tune mBERT on the UPOS (universal POS) tagging task using the source language's training set, and evaluate it on the target language's testing set.", |
|
"cite_spans": [ |
|
{ |
|
"start": 294, |
|
"end": 295, |
|
"text": "3", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Our selected evaluation metric is the micro average F1 score. Clearly, for every pair of languages we calculate two F1 scores, one for each direction. The two scores are not necessarily equivalent. In all our experiments, we use the commonly used pre-trained language model bert-base-multilingual-cased, provided by the Hugging Face transformers library (Wolf et al., 2019) . For every language we fine-tune the model for the standard token classification downstream task for three epochs, using a learning rate value of 5e \u2212 5.", |
|
"cite_spans": [ |
|
{ |
|
"start": 354, |
|
"end": 373, |
|
"text": "(Wolf et al., 2019)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We include 36 languages in our study, taken from a diversity of language families and subfamilies. The full list of languages is provided in Figure 2 . For each language we indicate its two-letter ISO 639-1 code, which we use throughout the paper. All the 36 languages we process are covered by mBERT.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 141, |
|
"end": 149, |
|
"text": "Figure 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Overall, we calculate the F1 score for every pair of languages, resulting in 36 2 = 1296 scores. A partial list of the scores is provided in As mentioned before, the two F1 scores that were calculated for each pair of different languages, are not necessarily equal. In fact, they are very unlikely to be equal, since the performance of the tagger is affected not only by the difference between the languages, but also by the size and the quality of the training sets, as well as the volume and quality of the texts in each relevant language, which were used for training mBERT.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "The average of the absolute difference between all language pairs is 0.0874 and the standard deviation is 0.074. While some pairs have relatively similar scores in both directions, some other have significantly different ones. However, as we show later, we do not use the F1 scores directly as some sort of a distance function between the languages. Instead, we represent each language l by a vector of F1 scores calculated by all other models during evaluation on l's testing set, and use a clustering Figure 2 : The 36 languages we include in our study. We chose languages from different families and subfamilies. The two-letter ISO 639-1 code is provided in parentheses next to each language name. algorithm to organize these vectors into language families.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 503, |
|
"end": 511, |
|
"text": "Figure 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Before we show how we do that, first, we argue that our cross-lingual F1 score is an important piece of information for reconstructing the classic tree model. Our argument is based on the correlation between our cross-lingual F1 score and the proximity of language pairs in the classic tree model. In order to measure the proximity between two languages in the classic tree model, we use the Wu-Palmer similarity (Wu and Palmer, 1994) metric, which was originally invented for measuring relatedness of two synsets in a WordNet taxonomy. For the context of using Wu-Palmer, the tree model has the same characteristics as WordNet; language family names are represented by intermediate nodes, while language names are represented by the leaves. Therefore, the Wu-Palmer score for two languages L 1 , L 2 is calculated as follows:", |
|
"cite_spans": [ |
|
{ |
|
"start": 413, |
|
"end": 434, |
|
"text": "(Wu and Palmer, 1994)", |
|
"ref_id": "BIBREF27" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "2 \u2022 depth(lcs(L 1 , L 2 )) depth(L 1 ) + depth(L 2 )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "with lcs representing the least common subsumer, that is, the first common ancestor of the two languages in the language-family tree. The score ranges between 0 and 1, but it can never go to zero since the depth of lcs(L 1 , L 2 ) is never zero (the model tree has a single root).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We denote the Wu-Palmer score as WP. As opposed to our cross-lingual F1 score, WP is symmetric.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We calculate WP for every language pair, and compare with our F1; the results are shown in Figure 3. Every data point in this chart represents a single language pair out of the 36 2 pairs. Overall, we learn that the F1 score increases along with WP, except maybe on relatively small WP values, representing pairs of languages taken from significantly different branches of the language-family tree.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 91, |
|
"end": 97, |
|
"text": "Figure", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Furthermore, we measure the correlation between the two metrics using Pearson (for linear correlation) and Spearman (for monotonic correlation) and realize that both are strongly correlated with Pearson= 0.64 (at p < 0.001), and Spearman= 0.59, (at p < 0.001). Figure 4 visualizes the F1 scores of all language pairs as a heatmap, with target languages provided as rows and source languages as columns. For each target language, all the 36 source languages are sorted according to the F1 scores (from the highest to the lowest). The color represents the proximity, as calculated by the WP score; a lighter color is equivalent to a higher proximity. For example in the fifth row, the best performance on the Spanish test set is observed by the Spanish model, followed by other Romance languages, Catalan, Italian, Portuguese and so on. The worst performance was recorded by the Welsh model. Evidently, higher proximity values (light boxes on the left side of the heatmap) derive higher performance on the crosslingual POS task, indicating that the closer two languages are, the encoded information in their corresponding models tends to be more helpful for POS tagging.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 261, |
|
"end": 269, |
|
"text": "Figure 4", |
|
"ref_id": "FIGREF2" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Language Similarity Score", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "In this section we show how we use the resulting F1 scores, calculated for every language pair, to reconstruct the language-family tree.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reconstructing Language Families", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "We represent every language l by a 36dimensional vector consisted of the F1 scores of the models that have been trained on all other languages, evaluated on l. We generate exactly 36 vectors, one for each language. Conceptually, the vector of language l represents the similarity of l to all the other languages, by considering only cross-lingual UPOS information, as captured by mBERT.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reconstructing Language Families", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "To identify families and subfamilies of languages, we use k-means (Lloyd, 1982) to cluster the 36 vectors. In addition to the collection of vectors, k-means receives as input a parameter k that denotes the number of clusters.", |
|
"cite_spans": [ |
|
{ |
|
"start": 66, |
|
"end": 79, |
|
"text": "(Lloyd, 1982)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Reconstructing Language Families", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "According to Figure 2 , the tree model organizes the 36 languages into 9 families; therefore, we run k-means with value of k = 9. In Figure 5 we visualize the resulting clusters. The color of the circle next to the language name marks the cluster. Note that while the k-means algorithm works with 36-dimensional vectors given as an input, we visualize the vectors on a 2-dimensional axis, which we calculate using the principal component analysis (PCA) algorithm for reducing dimensions. The clusters are summarized in Table 2 . We discuss the results in the following section.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 13, |
|
"end": 21, |
|
"text": "Figure 2", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 133, |
|
"end": 141, |
|
"text": "Figure 5", |
|
"ref_id": "FIGREF3" |
|
}, |
|
{ |
|
"start": 519, |
|
"end": 526, |
|
"text": "Table 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Reconstructing Language Families", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "The alternative partitioning for language families that we get, partially align with the classic tree model.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "Cluster 1 contains only Romance languages. All languages in cluster 2, except Romanian, are considered as Germanic in the classic tree model. Cluster 3 contains all Slavic languages excluding the Baltic languages. Cluster 4 includes the Baltic languages (Lithuanian and Latvian) as well as two Uralic languages (Finnish and Estonian). Those four languages are spoken in the geographically close countries Lithuania, Latvia, Finland and Estonia, respectively, suggesting that there might be a geographical dimension in our POS-based language proximity method. We plan to further investigate this discovery as one of our future directions. Cluster 6 contains only Hindustani languages. The two Semitic languages (Hebrew and Arabic) are grouped together in cluster 7, which also includes Welsh Celtic Table 2 : The clusters obtained by running k-means with k = 9. We provide some information about the language families of each cluster in the third column.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 798, |
|
"end": 805, |
|
"text": "Table 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "Persian probably due to historical influences. Hungarian is the only language in cluster 5. Clusters 8 and 9 represents two languages of the Celtic family. They should have probably been clustered together.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "Overall although there are a few misplacements, our clustering method was able to reconstruct parts of the tree model. 31 out of 36 languages were classified correctly according to the classic model. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "In this work we used a cross-lingual model trained on UPOS for measuring the proximity between languages. We showed that our new languageproximity model can reconstruct families of genetically related languages, suggesting that POS information plays a major role in modelling similarity between languages.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusions", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "We believe that we have demonstrated the potential of a fine-tuned mBERT model to capture some cross-lingual information that is needed for assigning UPOS tags to a text written in an unseen language. On average, models of genetically related languages perform better on each other in this task, even if they are not written in the same script. For example, in Table 1 we show that a Spanish (ES) model performs similarly on English (EN) and Russian (RU), although both Spanish and English are written in the Latin script while Russian is written in the Cyrillic script.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 361, |
|
"end": 368, |
|
"text": "Table 1", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusions", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "There are a few caveats to this research to note.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusions", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "mBERT was pre-trained on the full collections of Wikipedia articles in the relevant languages. Therefore, the size of those collections varies proportionally to the number of active speakers. To handle that bias, the authors of mBERT had decided to upsample the Wikipedia collections of the less dominant languages, in the main training loop. Wu and Dredze (2020) have recently addressed that problem and showed that mBERT performs better on cross-lingual zero-shot tasks on languages that have large Wikipedia collections. In our work, we handle that bias by designing each individual language vector to have F1 scores from all other languages, including both high-resource and low-resource languages. Therefore, every language is represented by F1 scores achieved by models trained on exactly the same language set.", |
|
"cite_spans": [ |
|
{ |
|
"start": 343, |
|
"end": 363, |
|
"text": "Wu and Dredze (2020)", |
|
"ref_id": "BIBREF26" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusions", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Another caveat is the size and quality of the treebanks we use for training and testing our models. As noted before, we believe that our approach to represent a language using scores from models trained on all the 36 language included in this research, mitigates this risk.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusions", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "We make a final practical observation. The results of our study suggest that for UPOS tagging, mBERT may benefit from training on texts written in languages that are genetically similar to the target language, based on the classic tree model. These results are aligned with what have been reported by Wu and Dredze (2020) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 301, |
|
"end": 321, |
|
"text": "Wu and Dredze (2020)", |
|
"ref_id": "BIBREF26" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusions", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Similar to BERT, mBERT's tokens are subwords.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "https://universaldependencies.org 4 https://universaldependencies.org/u/ pos", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "A Appendix: F1 Scores ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "annex", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Comparing fifty natural languages and twelve genetic languages using word embedding language divergence (WELD) as a quantitative measure of language distance", |
|
"authors": [ |
|
{ |
|
"first": "Ehsaneddin", |
|
"middle": [], |
|
"last": "Asgari", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [ |
|
"K" |
|
], |
|
"last": "Mohammad", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Mofrad", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the Workshop on Multilingual and Cross-lingual Methods in NLP", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "65--74", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/W16-1208" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ehsaneddin Asgari and Mohammad R.K. Mofrad. 2016. Comparing fifty natural languages and twelve genetic languages using word embedding language divergence (WELD) as a quantitative measure of lan- guage distance. In Proceedings of the Workshop on Multilingual and Cross-lingual Methods in NLP, pages 65-74, San Diego, California. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Adding typology to lexicostatistics: A", |
|
"authors": [ |
|
{ |
|
"first": "Dik", |
|
"middle": [], |
|
"last": "Bakker", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andr\u00e9", |
|
"middle": [], |
|
"last": "M\u00fcller", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Viveka", |
|
"middle": [], |
|
"last": "Velupillai", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S\u00f8ren", |
|
"middle": [], |
|
"last": "Wichmann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Cecil", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Pamela", |
|
"middle": [], |
|
"last": "Brown", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dmitry", |
|
"middle": [], |
|
"last": "Brown", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Robert", |
|
"middle": [], |
|
"last": "Egorov", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Anthony", |
|
"middle": [], |
|
"last": "Mailhammer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Eric", |
|
"middle": [ |
|
"W" |
|
], |
|
"last": "Grant", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Holman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dik Bakker, Andr\u00e9 M\u00fcller, Viveka Velupillai, S\u00f8ren Wichmann, Cecil H Brown, Pamela Brown, Dmitry Egorov, Robert Mailhammer, Anthony Grant, and Eric W Holman. 2009. Adding typology to lexico- statistics: A combined approach to language classifi- cation.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "An experimental study comparing linguistic phylogenetic reconstruction methods", |
|
"authors": [ |
|
{ |
|
"first": "Fran\u00e7ois", |
|
"middle": [], |
|
"last": "Barban\u00e7on", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Steven", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luay", |
|
"middle": [], |
|
"last": "Evans", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Don", |
|
"middle": [], |
|
"last": "Nakhleh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tandy", |
|
"middle": [], |
|
"last": "Ringe", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Warnow", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "Diachronica", |
|
"volume": "30", |
|
"issue": "2", |
|
"pages": "143--170", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Fran\u00e7ois Barban\u00e7on, Steven N Evans, Luay Nakhleh, Don Ringe, and Tandy Warnow. 2013. An exper- imental study comparing linguistic phylogenetic re- construction methods. Diachronica, 30(2):143-170.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Automated classification of the world s languages: a description of the method and preliminary results", |
|
"authors": [ |
|
{ |
|
"first": "Eric", |
|
"middle": [ |
|
"W" |
|
], |
|
"last": "Cecil H Brown", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S\u00f8ren", |
|
"middle": [], |
|
"last": "Holman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Viveka", |
|
"middle": [], |
|
"last": "Wichmann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Velupillai", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Language Typology and Universals", |
|
"volume": "61", |
|
"issue": "4", |
|
"pages": "285--308", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Cecil H Brown, Eric W Holman, S\u00f8ren Wichmann, and Viveka Velupillai. 2008. Automated classification of the world s languages: a description of the method and preliminary results. Language Typology and Universals, 61(4):285-308.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Linguistic distance: A quantitative measure of the distance between English and other languages", |
|
"authors": [ |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Barry", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paul", |
|
"middle": [ |
|
"W" |
|
], |
|
"last": "Chiswick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Miller", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Journal of Multilingual and Multicultural Development", |
|
"volume": "26", |
|
"issue": "1", |
|
"pages": "1--11", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Barry R Chiswick and Paul W Miller. 2005. Linguistic distance: A quantitative measure of the distance be- tween English and other languages. Journal of Mul- tilingual and Multicultural Development, 26(1):1- 11.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "A massively parallel corpus: The Bible in 100 languages. Language resources and evaluation", |
|
"authors": [ |
|
{ |
|
"first": "Christos", |
|
"middle": [], |
|
"last": "Christodouloupoulos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Steedman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "", |
|
"volume": "49", |
|
"issue": "", |
|
"pages": "375--395", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Christos Christodouloupoulos and Mark Steedman. 2015. A massively parallel corpus: The Bible in 100 languages. Language resources and evaluation, 49(2):375-395.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "The igraph software package for complex network research", |
|
"authors": [ |
|
{ |
|
"first": "Gabor", |
|
"middle": [], |
|
"last": "Csardi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tamas", |
|
"middle": [], |
|
"last": "Nepusz", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Gabor Csardi and Tamas Nepusz. 2006. The igraph software package for complex network research. In- terJournal, Complex Systems:1695.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
|
"authors": [ |
|
{ |
|
"first": "Jacob", |
|
"middle": [], |
|
"last": "Devlin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming-Wei", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kenton", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kristina", |
|
"middle": [], |
|
"last": "Toutanova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "4171--4186", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/N19-1423" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171-4186, Minneapolis, Minnesota. Associ- ation for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Contemporary Linguistics. Longman", |
|
"authors": [ |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Dobrovolsky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "William", |
|
"middle": [], |
|
"last": "Delaney", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "O'", |
|
"middle": [], |
|
"last": "Grady", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Francis", |
|
"middle": [], |
|
"last": "Katamba", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Michael Dobrovolsky, William Delaney O'Grady, and Francis Katamba. 2016. Contemporary Linguistics. Longman.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "From language identification to language distance", |
|
"authors": [ |
|
{ |
|
"first": "Pablo", |
|
"middle": [], |
|
"last": "Gamallo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "I\u00f1aki", |
|
"middle": [], |
|
"last": "Jos\u00e9 Ramom Pichel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Alegria", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Physica A: Statistical Mechanics and its Applications", |
|
"volume": "484", |
|
"issue": "", |
|
"pages": "152--162", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Pablo Gamallo, Jos\u00e9 Ramom Pichel, and I\u00f1aki Alegria. 2017. From language identification to language dis- tance. Physica A: Statistical Mechanics and its Ap- plications, 484:152-162.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Do languages grow on trees? The tree metaphor in the history of linguistics. Classification and evolution in biology, linguistics and the history of science. Concepts-methods-visualization", |
|
"authors": [ |
|
{ |
|
"first": "Hans", |
|
"middle": [], |
|
"last": "Geisler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Johann-Mattis", |
|
"middle": [], |
|
"last": "List", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "111--124", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hans Geisler and Johann-Mattis List. 2013. Do lan- guages grow on trees? The tree metaphor in the history of linguistics. Classification and evolution in biology, linguistics and the history of science. Concepts-methods-visualization, pages 111-124.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "It's not Greek to mBERT: Inducing word-level translations from multilingual BERT", |
|
"authors": [ |
|
{ |
|
"first": "Shauli", |
|
"middle": [], |
|
"last": "Hila Gonen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yanai", |
|
"middle": [], |
|
"last": "Ravfogel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yoav", |
|
"middle": [], |
|
"last": "Elazar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Goldberg", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "45--56", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.blackboxnlp-1.5" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hila Gonen, Shauli Ravfogel, Yanai Elazar, and Yoav Goldberg. 2020. It's not Greek to mBERT: Induc- ing word-level translations from multilingual BERT. In Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, pages 45-56, Online. Association for Compu- tational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Explorations in automated lexicostatistics", |
|
"authors": [ |
|
{ |
|
"first": "S\u00f8ren", |
|
"middle": [], |
|
"last": "Eric W Holman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Wichmann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "Cecil", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Viveka", |
|
"middle": [], |
|
"last": "Brown", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andr\u00e9", |
|
"middle": [], |
|
"last": "Velupillai", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dik", |
|
"middle": [], |
|
"last": "M\u00fcller", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Bakker", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Folia Linguistica", |
|
"volume": "42", |
|
"issue": "2", |
|
"pages": "331--354", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Eric W Holman, S\u00f8ren Wichmann, Cecil H Brown, Viveka Velupillai, Andr\u00e9 M\u00fcller, and Dik Bakker. 2008. Explorations in automated lexicostatistics. Folia Linguistica, 42(2):331-354.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Least squares quantization in PCM", |
|
"authors": [ |
|
{ |
|
"first": "Stuart", |
|
"middle": [], |
|
"last": "Lloyd", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1982, |
|
"venue": "IEEE transactions on information theory", |
|
"volume": "28", |
|
"issue": "2", |
|
"pages": "129--137", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Stuart Lloyd. 1982. Least squares quantization in PCM. IEEE transactions on information theory, 28(2):129-137.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Evidence for syntax as a signal of historical relatedness", |
|
"authors": [ |
|
{ |
|
"first": "Giuseppe", |
|
"middle": [], |
|
"last": "Longobardi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Cristina", |
|
"middle": [], |
|
"last": "Guardiano", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Lingua", |
|
"volume": "119", |
|
"issue": "11", |
|
"pages": "1679--1706", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Giuseppe Longobardi and Cristina Guardiano. 2009. Evidence for syntax as a signal of historical relat- edness. Lingua, 119(11):1679-1706.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Toward a syntactic phylogeny of modern Indo-European languages", |
|
"authors": [ |
|
{ |
|
"first": "Giuseppe", |
|
"middle": [], |
|
"last": "Longobardi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Cristina", |
|
"middle": [], |
|
"last": "Guardiano", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Giuseppina", |
|
"middle": [], |
|
"last": "Silvestri", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Alessio", |
|
"middle": [], |
|
"last": "Boattini", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andrea", |
|
"middle": [], |
|
"last": "Ceolin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "Journal of Historical Linguistics", |
|
"volume": "3", |
|
"issue": "1", |
|
"pages": "122--152", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Giuseppe Longobardi, Cristina Guardiano, Giusep- pina Silvestri, Alessio Boattini, and Andrea Ceolin. 2013. Toward a syntactic phylogeny of modern Indo-European languages. Journal of Historical Lin- guistics, 3(1):122-152.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Perfect phylogenetic networks: A new methodology for reconstructing the evolutionary history of natural languages. Language", |
|
"authors": [ |
|
{ |
|
"first": "Luay", |
|
"middle": [], |
|
"last": "Nakhleh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Don", |
|
"middle": [], |
|
"last": "Ringe", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tandy", |
|
"middle": [], |
|
"last": "Warnow", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "382--420", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Luay Nakhleh, Don Ringe, and Tandy Warnow. 2005. Perfect phylogenetic networks: A new methodology for reconstructing the evolutionary history of natural languages. Language, pages 382-420.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "Language distance and tree reconstruction", |
|
"authors": [ |
|
{ |
|
"first": "Filippo", |
|
"middle": [], |
|
"last": "Petroni", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Maurizio", |
|
"middle": [], |
|
"last": "Serva", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Journal of Statistical Mechanics: Theory and Experiment", |
|
"volume": "", |
|
"issue": "08", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Filippo Petroni and Maurizio Serva. 2008. Lan- guage distance and tree reconstruction. Journal of Statistical Mechanics: Theory and Experiment, 2008(08):P08012.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Measures of lexical distance between languages", |
|
"authors": [ |
|
{ |
|
"first": "Filippo", |
|
"middle": [], |
|
"last": "Petroni", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Maurizio", |
|
"middle": [], |
|
"last": "Serva", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Physica A: Statistical Mechanics and its Applications", |
|
"volume": "389", |
|
"issue": "11", |
|
"pages": "2280--2283", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Filippo Petroni and Maurizio Serva. 2010. Mea- sures of lexical distance between languages. Phys- ica A: Statistical Mechanics and its Applications, 389(11):2280-2283.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Die ersten spaltungen des indogermanischen urvolkes", |
|
"authors": [], |
|
"year": null, |
|
"venue": "Allgemeine Monatsschrift f\u00fcr Wissenschaft und Literatur", |
|
"volume": "3", |
|
"issue": "", |
|
"pages": "786--787", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "August Schleicher. 1853. Die ersten spaltungen des indogermanischen urvolkes. Allgemeine Monatss- chrift f\u00fcr Wissenschaft und Literatur, 3:786-787.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Indo-European languages tree by Levenshtein distance", |
|
"authors": [ |
|
{ |
|
"first": "Maurizio", |
|
"middle": [], |
|
"last": "Serva", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Filippo", |
|
"middle": [], |
|
"last": "Petroni", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Europhysics Letters)", |
|
"volume": "81", |
|
"issue": "6", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Maurizio Serva and Filippo Petroni. 2008. Indo- European languages tree by Levenshtein distance. EPL (Europhysics Letters), 81(6):68005.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Phylogenetics of Indo-European language families via an algebrogeometric analysis of their syntactic structures", |
|
"authors": [ |
|
{ |
|
"first": "Kevin", |
|
"middle": [], |
|
"last": "Shu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andrew", |
|
"middle": [], |
|
"last": "Ortegaray", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Robert", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Matilde", |
|
"middle": [], |
|
"last": "Berwick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Marcolli", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2021, |
|
"venue": "Mathematics in Computer Science", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--55", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kevin Shu, Andrew Ortegaray, Robert C Berwick, and Matilde Marcolli. 2021. Phylogenetics of Indo-European language families via an algebro- geometric analysis of their syntactic structures. Mathematics in Computer Science, pages 1-55.", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "Salish internal relationships", |
|
"authors": [ |
|
{ |
|
"first": "Morris", |
|
"middle": [], |
|
"last": "Swadesh", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1950, |
|
"venue": "International Journal of American Linguistics", |
|
"volume": "16", |
|
"issue": "4", |
|
"pages": "157--167", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Morris Swadesh. 1950. Salish internal relation- ships. International Journal of American Linguis- tics, 16(4):157-167.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Huggingface's transformers: State-of-the-art natural language processing", |
|
"authors": [ |
|
{ |
|
"first": "Thomas", |
|
"middle": [], |
|
"last": "Wolf", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lysandre", |
|
"middle": [], |
|
"last": "Debut", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Victor", |
|
"middle": [], |
|
"last": "Sanh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Julien", |
|
"middle": [], |
|
"last": "Chaumond", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Clement", |
|
"middle": [], |
|
"last": "Delangue", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Anthony", |
|
"middle": [], |
|
"last": "Moi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Pierric", |
|
"middle": [], |
|
"last": "Cistac", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tim", |
|
"middle": [], |
|
"last": "Rault", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R\u00e9mi", |
|
"middle": [], |
|
"last": "Louf", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Morgan", |
|
"middle": [], |
|
"last": "Funtowicz", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"arXiv": [ |
|
"arXiv:1910.03771" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pier- ric Cistac, Tim Rault, R\u00e9mi Louf, Morgan Fun- towicz, et al. 2019. Huggingface's transformers: State-of-the-art natural language processing. arXiv preprint arXiv:1910.03771.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "Beto, bentz, becas: The surprising cross-lingual effectiveness of BERT", |
|
"authors": [ |
|
{ |
|
"first": "Shijie", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Dredze", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "833--844", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/D19-1077" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Shijie Wu and Mark Dredze. 2019. Beto, bentz, be- cas: The surprising cross-lingual effectiveness of BERT. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natu- ral Language Processing (EMNLP-IJCNLP), pages 833-844, Hong Kong, China. Association for Com- putational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Are all languages created equal in multilingual BERT?", |
|
"authors": [ |
|
{ |
|
"first": "Shijie", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Dredze", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the 5th Workshop on Representation Learning for NLP", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "120--130", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/2020.repl4nlp-1.16" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Shijie Wu and Mark Dredze. 2020. Are all languages created equal in multilingual BERT? In Proceedings of the 5th Workshop on Representation Learning for NLP, pages 120-130, Online. Association for Com- putational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Verb semantics and lexical selection", |
|
"authors": [ |
|
{ |
|
"first": "Zhibiao", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Proceedings of the 32nd Annual meeting of the Associations for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "133--138", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhibiao Wu and Martha Palmer. 1994. Verb semantics and lexical selection. In Proceedings of the 32nd An- nual meeting of the Associations for Computational Linguistics, pages 133-138.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"uris": null, |
|
"num": null, |
|
"text": "Figure 1: The Indo-European language branch; the graph was created by the igraph package for Python (Csardi and Nepusz, 2006). The languages are represented using their equivalent two-letter ISO 639-1 code.", |
|
"type_str": "figure" |
|
}, |
|
"FIGREF1": { |
|
"uris": null, |
|
"num": null, |
|
"text": "UPOS F1 scores compared with WP scores.", |
|
"type_str": "figure" |
|
}, |
|
"FIGREF2": { |
|
"uris": null, |
|
"num": null, |
|
"text": "A heatmap of the WP scores calculated for all language pairs, sorted according to F1 scores. For more information about this arrangement see the text.", |
|
"type_str": "figure" |
|
}, |
|
"FIGREF3": { |
|
"uris": null, |
|
"num": null, |
|
"text": "The 9 clusters resulted from k-means. The original 36-dimensional vectors are visualized using their first two principle components.", |
|
"type_str": "figure" |
|
}, |
|
"TABREF0": { |
|
"html": null, |
|
"content": "<table><tr><td>,</td></tr></table>", |
|
"type_str": "table", |
|
"num": null, |
|
"text": "" |
|
} |
|
} |
|
} |
|
} |