|
{ |
|
"paper_id": "Q14-1009", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:11:04.544307Z" |
|
}, |
|
"title": "Parallel Algorithms for Unsupervised Tagging", |
|
"authors": [ |
|
{ |
|
"first": "Sujith", |
|
"middle": [], |
|
"last": "Ravi", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Sergei", |
|
"middle": [], |
|
"last": "Vassilivitskii", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Vibhor", |
|
"middle": [], |
|
"last": "Rastogi", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "We propose a new method for unsupervised tagging that finds minimal models which are then further improved by Expectation Maximization training. In contrast to previous approaches that rely on manually specified and multi-step heuristics for model minimization, our approach is a simple greedy approximation algorithm DMLC (DISTRIBUTED-MINIMUM-LABEL-COVER) that solves this objective in a single step. We extend the method and show how to efficiently parallelize the algorithm on modern parallel computing platforms while preserving approximation guarantees. The new method easily scales to large data and grammar sizes, overcoming the memory bottleneck in previous approaches. We demonstrate the power of the new algorithm by evaluating on various sequence labeling tasks: Part-of-Speech tagging for multiple languages (including lowresource languages), with complete and incomplete dictionaries, and supertagging, a complex sequence labeling task, where the grammar size alone can grow to millions of entries. Our results show that for all of these settings, our method achieves state-of-the-art scalable performance that yields high quality tagging outputs. * * The research described herein was conducted while the author was working at Google.", |
|
"pdf_parse": { |
|
"paper_id": "Q14-1009", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "We propose a new method for unsupervised tagging that finds minimal models which are then further improved by Expectation Maximization training. In contrast to previous approaches that rely on manually specified and multi-step heuristics for model minimization, our approach is a simple greedy approximation algorithm DMLC (DISTRIBUTED-MINIMUM-LABEL-COVER) that solves this objective in a single step. We extend the method and show how to efficiently parallelize the algorithm on modern parallel computing platforms while preserving approximation guarantees. The new method easily scales to large data and grammar sizes, overcoming the memory bottleneck in previous approaches. We demonstrate the power of the new algorithm by evaluating on various sequence labeling tasks: Part-of-Speech tagging for multiple languages (including lowresource languages), with complete and incomplete dictionaries, and supertagging, a complex sequence labeling task, where the grammar size alone can grow to millions of entries. Our results show that for all of these settings, our method achieves state-of-the-art scalable performance that yields high quality tagging outputs. * * The research described herein was conducted while the author was working at Google.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Supervised sequence labeling with large labeled training datasets is considered a solved problem. For instance, state of the art systems obtain tagging accuracies over 97% for part-of-speech (POS) tagging on the English Penn Treebank. However, learning accurate taggers without labeled data remains a challenge. The accuracies quickly drop when faced with data from a different domain, language, or when there is very little labeled information available for training (Banko and Moore, 2004) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 468, |
|
"end": 491, |
|
"text": "(Banko and Moore, 2004)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Recently, there has been an increasing amount of research tackling this problem using unsupervised methods. A popular approach is to learn from POS-tag dictionaries (Merialdo, 1994) , where we are given a raw word sequence and a dictionary of legal tags for each word type. Learning from POStag dictionaries is still challenging. Complete wordtag dictionaries may not always be available for use and in every setting. When they are available, the dictionaries are often noisy, resulting in high tagging ambiguity. Furthermore, when applying taggers in new domains or different datasets, we may encounter new words that are missing from the dictionary. There have been some efforts to learn POS taggers from incomplete dictionaries by extending the dictionary to include these words using some heuristics (Toutanova and Johnson, 2008) or using other methods such as type-supervision (Garrette and Baldridge, 2012) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 165, |
|
"end": 181, |
|
"text": "(Merialdo, 1994)", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 804, |
|
"end": 833, |
|
"text": "(Toutanova and Johnson, 2008)", |
|
"ref_id": "BIBREF29" |
|
}, |
|
{ |
|
"start": 882, |
|
"end": 912, |
|
"text": "(Garrette and Baldridge, 2012)", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this work, we tackle the problem of unsupervised sequence labeling using tag dictionaries. The first reported work on this problem was on POS tagging from Merialdo (1994) . The approach involved training a standard Hidden Markov Model (HMM) using the Expectation Maximization (EM) algorithm (Dempster et al., 1977) , though EM does not perform well on this task (Johnson, 2007) . More recent methods have yielded better performance than EM (see (Ravi and Knight, 2009) for an overview).", |
|
"cite_spans": [ |
|
{ |
|
"start": 158, |
|
"end": 173, |
|
"text": "Merialdo (1994)", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 294, |
|
"end": 317, |
|
"text": "(Dempster et al., 1977)", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 365, |
|
"end": 380, |
|
"text": "(Johnson, 2007)", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 448, |
|
"end": 471, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "One interesting line of research introduced by Ravi and Knight (2009) explores the idea of performing model minimization followed by EM training to learn taggers. Their idea is closely related to the classic Minimum Description Length principle for model selection (Barron et al., 1998) . They (1) formulate an objective function to find the smallest model that explains the text (model minimization step), and then, (2) fit the minimized model to the data (EM step). For POS tagging, this method (Ravi and Knight, 2009) yields the best performance to date; 91.6% tagging accuracy on a standard test dataset from the English Penn Treebank. The original work from (Ravi and Knight, 2009) uses an integer linear programming (ILP) formulation to find minimal models, an approach which does not scale to large datasets. Ravi et al. (2010b) introduced a two-step greedy approximation to the original objective function (called the MIN-GREEDY algorithm) that runs much faster while maintaining the high tagging performance. Garrette and Baldridge (2012) showed how to use several heuristics to further improve this algorithm (for instance, better choice of tag bigrams when breaking ties) and stack other techniques on top, such as careful initialization of HMM emission models which results in further performance gains. Their method also works under incomplete dictionary scenarios and can be applied to certain low-resource scenarios (Garrette and Baldridge, 2013) by combining model minimization with supervised training.", |
|
"cite_spans": [ |
|
{ |
|
"start": 47, |
|
"end": 69, |
|
"text": "Ravi and Knight (2009)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 265, |
|
"end": 286, |
|
"text": "(Barron et al., 1998)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 497, |
|
"end": 520, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 663, |
|
"end": 686, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 816, |
|
"end": 835, |
|
"text": "Ravi et al. (2010b)", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 1018, |
|
"end": 1047, |
|
"text": "Garrette and Baldridge (2012)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 1431, |
|
"end": 1461, |
|
"text": "(Garrette and Baldridge, 2013)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this work, we propose a new scalable algorithm for performing model minimization for this task. By making an assumption on the structure of the solution, we prove that a variant of the greedy set cover algorithm always finds an approximately optimal label set. This is in contrast to previous methods that employ heuristic approaches with no guarantee on the quality of the solution. In addition, we do not have to rely on ad hoc tie-breaking procedures or careful initializations for unknown words. Finally, not only is the proposed method approximately optimal, it is also easy to distribute, allowing it to easily scale to very large datasets. We show empirically that our method, combined with an EM training step outperforms existing state of the art systems.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "\u2022 We present a new method, DISTRIBUTED MINIMUM LABEL COVER, DMLC, for model minimization that uses a fast, greedy algorithm with formal approximation guarantees to the quality of the solution.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Our Contributions", |
|
"sec_num": "1.1" |
|
}, |
|
{ |
|
"text": "\u2022 We show how to efficiently parallelize the algorithm while preserving approximation guarantees. In contrast, existing minimization approaches cannot match the new distributed algorithm when scaling from thousands to millions or even billions of tokens.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Our Contributions", |
|
"sec_num": "1.1" |
|
}, |
|
{ |
|
"text": "\u2022 We show that our method easily scales to both large data and grammar sizes, and does not require the corpus or label set to fit into memory. This allows us to tackle complex tagging tasks, where the tagset consists of several thousand labels, which results in more than one million entries in the grammar.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Our Contributions", |
|
"sec_num": "1.1" |
|
}, |
|
{ |
|
"text": "\u2022 We demonstrate the power of the new method by evaluating under several different scenarios-POS tagging for multiple languages (including low-resource languages), with complete and incomplete dictionaries, as well as a complex sequence labeling task of supertagging. Our results show that for all these settings, our method achieves state-of-the-art performance yielding high quality taggings.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Our Contributions", |
|
"sec_num": "1.1" |
|
}, |
|
{ |
|
"text": "Recently, there has been an increasing amount of research tackling this problem from multiple directions. Some efforts have focused on inducing POS tag clusters without any tags (Christodoulopoulos et al., 2010; Reichart et al., 2010; Moon et al., 2010) , but evaluating such systems proves difficult since it is not straightforward to map the cluster labels onto gold standard tags. A more popular approach is to learn from POS-tag dictionaries (Merialdo, 1994; Ravi and Knight, 2009) , incomplete dictionaries (Hasan and Ng, 2009; Garrette and Baldridge, 2012) and human-constructed dictionaries (Goldberg et al., 2008) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 178, |
|
"end": 211, |
|
"text": "(Christodoulopoulos et al., 2010;", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 212, |
|
"end": 234, |
|
"text": "Reichart et al., 2010;", |
|
"ref_id": "BIBREF27" |
|
}, |
|
{ |
|
"start": 235, |
|
"end": 253, |
|
"text": "Moon et al., 2010)", |
|
"ref_id": "BIBREF22" |
|
}, |
|
{ |
|
"start": 446, |
|
"end": 462, |
|
"text": "(Merialdo, 1994;", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 463, |
|
"end": 485, |
|
"text": "Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 512, |
|
"end": 532, |
|
"text": "(Hasan and Ng, 2009;", |
|
"ref_id": "BIBREF15" |
|
}, |
|
{ |
|
"start": 533, |
|
"end": 562, |
|
"text": "Garrette and Baldridge, 2012)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 598, |
|
"end": 621, |
|
"text": "(Goldberg et al., 2008)", |
|
"ref_id": "BIBREF12" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Another direction that has been explored in the past includes bootstrapping taggers for a new language based on information acquired from other languages (Das and Petrov, 2011) or limited annotation resources (Garrette and Baldridge, 2013) . Additional work focused on building supervised taggers for noisy domains such as Twitter (Gimpel et al., 2011) . While most of the relevant work in this area centers on POS tagging, there has been some work done for building taggers for more complex sequence labeling tasks such as supertagging (Ravi et al., 2010a) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 154, |
|
"end": 176, |
|
"text": "(Das and Petrov, 2011)", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 209, |
|
"end": 239, |
|
"text": "(Garrette and Baldridge, 2013)", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 331, |
|
"end": 352, |
|
"text": "(Gimpel et al., 2011)", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 537, |
|
"end": 557, |
|
"text": "(Ravi et al., 2010a)", |
|
"ref_id": "BIBREF25" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Other related work include alternative methods for learning sparse models via priors in Bayesian inference (Goldwater and Griffiths, 2007) and posterior regularization (Ganchev et al., 2010) . But these methods only encourage sparsity and do not explicitly seek to minimize the model size, which is the objective function used in this work. Moreover, taggers learned using model minimization have been shown to produce state-of-the-art results for the problems discussed here.", |
|
"cite_spans": [ |
|
{ |
|
"start": 107, |
|
"end": 138, |
|
"text": "(Goldwater and Griffiths, 2007)", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 168, |
|
"end": 190, |
|
"text": "(Ganchev et al., 2010)", |
|
"ref_id": "BIBREF8" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Following Ravi and Knight (2009) , we formulate the problem as that of label selection on the sentence graph. Formally, we are given a set of sequences,", |
|
"cite_spans": [ |
|
{ |
|
"start": 10, |
|
"end": 32, |
|
"text": "Ravi and Knight (2009)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "S = {S 1 , S 2 , . . . , S n } where each S i is a sequence of words, S i = w i1 , w i2 , . . . , w i,|S i | .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "With each word w ij we associate a set of possible tags T ij . We will denote by m the total number of (possibly duplicate) words (tokens) in the corpus.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Additionally, we define two special words w 0 and w \u221e with special tags start and end, and consider the modified sequences S i = w 0 , S i , w \u221e . To simplify notation, we will refer to w \u221e = w |S i |+1 . The sequence label problem asks us to select a valid tag t ij \u2208 T ij for each word w ij in the input to minimize a specific objective function.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We will refer to a tag pair (t i,j\u22121 , t ij ) as a label. Our aim is to minimize the number of distinct labels used to cover the full input. Formally, given a sequence S i and a tag t ij for each word w ij in S i , let the induced set of labels for sequence S i be", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "L i = |S i | j=1 {(t i,j\u22121 , t ij )}.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "The total number of distinct labels used over all sequences is then", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u03c6 = \u222a i L i | = i |S i |+1 j=1 {(t i,j\u22121 , t ij )}|.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Note that the order of the tokens in the label makes a difference as {(NN, VP)} and {(VP, NN)} are two distinct labels. Now we can define the problem formally, following (Ravi and Knight, 2009) . Problem 1 (Minimum Label Cover). Given a set S of sequences of words, where each word w ij has a set of valid tags T ij , the problem is to find a valid tag assignment t ij \u2208 T ij for each word that minimizes the number of distinct labels or tag pairs over all sequences, \u03c6 =", |
|
"cite_spans": [ |
|
{ |
|
"start": 170, |
|
"end": 193, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "i |S i |+1 j=1 {(t i,j\u22121 , t ij )}| .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "The problem is closely related to the classical Set Cover problem and is also NP-complete. To reduce Set Cover to the label selection problem, map each element i of the Set Cover instance to a single word sentence S i = w i1 , and let the valid tags T i1 contain the names of the sets that contain element i. Consider a solution to the label selection problem; every sentence S i is covered by two labels (w 0 , k i ) and (k i , w \u221e ), for some k i \u2208 T i1 , which corresponds to an element i being covered by set k i in the Set Cover instance. Thus any valid solution to the label selection problem leads to a feasible solution to the Set Cover problem ({k 1 , k 2 , . . .}) of exactly half the size.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Finally, we will use {{. . .}} notation to denote a multiset of elements, i.e. a set where an element may appear multiple times.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In this Section, we describe the DISTRIBUTED-MINIMUM-LABEL-COVER, DMLC, algorithm for approximately solving the minimum label cover problem. We describe the algorithm in a centralized setting, and defer the distributed implementation to Section 5. Before describing the algorithm, we briefly explain the relationship of the minimum label cover problem to set cover.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Algorithm", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "As we pointed out earlier, the minimum label cover problem is at least as hard as the Set Cover prob-1: Input: A set of sequences S with each words w ij having possible tags T ij . 2: Output: A tag assignment t ij \u2208 T ij for each word w ij approximately minimizing labels. 3: Let M be the multi set of all possible labels generated by choosing each possible tag t \u2208 T ij .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Modification of Set Cover", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "M = i \uf8eb \uf8ec \uf8ec \uf8ec \uf8ed |S i |+1 j=1 t \u2208T i,j\u22121 t\u2208T ij {{(t , t)}} \uf8f6 \uf8f7 \uf8f7 \uf8f7 \uf8f8", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Modification of Set Cover", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "(1) 4: Let L = \u2205 be the set of selected labels. 5: repeat", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Modification of Set Cover", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "Select the most frequent label not yet selected:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "6:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(t , t) = arg max (s ,s) / \u2208L |M \u2229 (s , s)|. 7: For each bigram (w i,j\u22121 , w ij ) where t \u2208 T i,j\u22121 and t \u2208 T ij tentatively assign t to w i,j\u22121 and t to w ij . Add (t , t) to L. 8:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "6:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "If a word gets two assignments, select one at random with equal probability.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "6:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "If a bigram (w i,j\u22121 , w ij ) is consistent with assignments in (t , t), fix the tentative assignments, and set T i,j\u22121 = {t } and T ij = {t}. Recompute M, the multiset of possible labels, with the updated T i,j\u22121 and T ij . 10: until there are no unassigned words Algorithm 1: MLC Algorithm 1: Input: A set of sequences S with each words w ij having possible tags T ij . 2: Output: A tag assignment t ij \u2208 T ij for each word w ij approximately minimizing labels. 3: (Graph Creation) Initialize each vertex v ij with the set of possible tags T ij and its neighbors v i,j+1 and v i,j\u22121 . 4: repeat 5:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "9:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(Message Passing) Each vertex v ij sends its possibly tags T ij to its forward neighbor v ij+1 .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "9:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(Counter Update) Each vertex receives the the tags T i,j\u22121 and adds all possible labels", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "6:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "{(s, s )|s \u2208 T i,j\u22121 , s \u2208 T ij } to a global counter (M ). 7:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "6:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(MaxLabel Selection) Each vertex queries the global counter M to find the maximum label (t, t ).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "6:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(Tentative Assignment) Each vertex v ij selects a tag tentatively as follows: If one of the tags t, t is in the feasible set T ij , it tentatively selects the tag.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "8:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(Random Assignment) If both are feasible it selects one at random. The vertex communicates its assignment to its neighbors.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "9:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(Confirmed Assignment) Each vertex receives the tentative assignment from its neighbors. If together with its neighbors it can match the selected label, the assignment is finalized. If the assigned tag is T , then the vertex v ij sets the valid tag set T ij to {t}. 11: until no unassigned vertices exist.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "10:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Algorithm 2: DMLC Implementation lem. An additional challenge comes from the fact that labels are tags for a pair of words, and hence are related. For example, if we label a word pair (w i,j\u22121 , w ij ) as (NN, VP), then the label for the next word pair (w ij , w i,j+1 ) has to be of the form (VP, *), i.e., it has to start with VP.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "10:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Previous work (Ravi et al., 2010a; Ravi et al., 2010b) recognized this challenge and employed two phase heuristic approaches. Eschewing heuristics, we will show that with one natural assumption, even with this extra set of constraints, the standard greedy algorithm for this problem results in a solution with a provable approximation ratio of O(log m). In practice, however, the algorithm performs far better than the worst case ratio, and similar to the work of (Gomes et al., 2006), we find that the greedy approach selects a cover approximately 11% worse than the optimum solution.", |
|
"cite_spans": [ |
|
{ |
|
"start": 14, |
|
"end": 34, |
|
"text": "(Ravi et al., 2010a;", |
|
"ref_id": "BIBREF25" |
|
}, |
|
{ |
|
"start": 35, |
|
"end": 54, |
|
"text": "Ravi et al., 2010b)", |
|
"ref_id": "BIBREF26" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "10:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We present in Algorithm 1 our MINIMUM LABEL COVER algorithm to approximately solve the minimum label cover problem. The algorithm is simple, efficient, and easy to distribute.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "The algorithm chooses labels one at a time, selecting a label that covers as many words as possible in every iteration. For this, it generates and maintains a multi-set of all possible labels M (Step 3). The multi-set contains an occurrence of each valid label, for example, if w i,j\u22121 has two possible valid tags N N and VP, and w ij has one possible valid tag VP, then M will contain two labels, namely (NN, VP) and (VP, VP). Since M is a multi-set it will contain duplicates, e.g. the label (NN, VP) will appear for each adjacent pair of words that have NN and VP as valid tags, respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "In each iteration, the algorithm picks a label with the most number of occurrences in M and adds it to the set of chosen labels (Step 6). Intuitively, this is a greedy step to select a label that covers the most number of word pairs.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Once the algorithm picks a label (t , t), it tries to assign as many words to tags t or t as possible (Step 7). A word can be assigned t if t is a valid tag for it, and t a valid tag for the next word in sequence. Similarly, a word can be assigned t, if t is a valid tag for it, and t a valid tag for the previous word. Some words can get both assignments, in which case we choose one tentatively at random (Step 8). If a word's tentative random tag, say t, is consistent with the choices of its adjacent words (say t from the previous word), then the tentative choice is fixed as a permanent one. Whenever a tag is selected, the set of valid tags T ij for the word is reduced to a singleton {t}. Once the set of valid tags T ij changes, the multi-set M of all possible labels also changes, as seen from Eq 1. The multi-set is then recomputed (Step 9) and the iterations repeated until all of words have been tagged.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "We can show that under a natural assumption this simple algorithm is approximately optimal.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Assumption 1 (c-feasibility). Let c \u2265 1 be any number, and k be the size of the optimal solution to the original problem. In each iteration, the MLC algorithm fixes the tags for some words. We say that the algorithm is c-feasible, if after each iteration there exists some solution to the remaining problem, consistent with the chosen tags, with size at most ck .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "The assumption encodes the fact that a single bad greedy choice is not going to destroy the overall structure of the solution, and a nearly optimal solution remains. We note that this assumption of cfeasibility is not only sufficient, as we will formally show, but is also necessary. Indeed, without any assumptions, once the algorithm fixes the tag for some words, an optimal label may no longer be consistent with the chosen tags, and it is not hard to find contrived examples where the size of the optimal solution doubles after each iteration of MLC.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Since the underlying problem is NP-complete, it is computationally hard to give direct evidence verifying the assumption on natural language inputs. However, on small examples we are able to show that the greedy algorithm is within a small constant factor of the optimum, specifically it is within 11% of the optimum model size for the POS tagging problem using the standard 24k dataset (Ravi and Knight, 2009) . Combined with the fact that the final method outperforms state of the art approaches, this leads us to conclude that the structural assumption is well justified. Proof. To prove the Lemma we will define an objective function\u03c6, counting the number of unlabeled word pairs, as a function of possible labels, and show that\u03c6 decreases by a factor of (1 \u2212 O( 1 /ck)) at every iteration.", |
|
"cite_spans": [ |
|
{ |
|
"start": 387, |
|
"end": 410, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "To define\u03c6, we first define \u03c6, the number of labeled word pairs. Consider a particular set of labels, L = {L 1 , L 2 , . . . , L k } where each label is a pair (t i , t j ). Call {t ij } a valid assignment of tokens if for each w ij , we have t ij \u2208 T ij . Then the score of L under an assignment t, which we denote by \u03c6 t , is the number of bigram labels that appear in", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "L. Formally, \u03c6 t (L) = | \u222a i,j {{(t i,j\u22121 , t ij ) \u2229 L}}|. Finally, we define \u03c6(L) to be the best such assign- ment, \u03c6(L) = max t \u03c6 t (L), and\u03c6(L) = m \u2212 \u03c6(L)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "the number of uncovered labels.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Consider the label selected by the algorithm in every step. By the c-feasibility assumption, there exists some solution having ck labels. Thus, some label from that solution covers at least a 1 /ck fraction of the remaining words. The selected label (t, t ) maximizes the intersection with the remaining feasible labels. The conflict resolution step ensures that in expectation the realized benefit is at least a half of the maximum, thereby reducing\u03c6 by at least a (1 \u2212 1 /2ck) fraction. Therefore, after O(kc log m) operations all of the labels are covered.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "MLC Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Once the greedy algorithm terminates and returns a minimized grammar of tag bigrams, we follow the approach of Ravi and Knight (2009) and fit the minimized model to the data using the alternating EM strategy.", |
|
"cite_spans": [ |
|
{ |
|
"start": 111, |
|
"end": 133, |
|
"text": "Ravi and Knight (2009)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Fitting the Model Using EM", |
|
"sec_num": "4.3" |
|
}, |
|
{ |
|
"text": "In this step, we run an alternating optimization procedure iteratively in phases. In each phase, we initialize (and prune away) parameters within the two HMM components (transition or emission model) using the output from the previous phase. We initialize this procedure by restricting the transition parameters to only those tag bigrams selected in the model minimization step. We train in conjunction with the original emission model using EM algorithm which prunes away some of the emission parameters. In the next phase, we alternate the initialization by choosing the pruned emission model along with the original transition model (with full set of tag bigrams) and retrain using EM. The alternating EM iterations are terminated when the change in the size of the observed grammar (i.e., the number of unique bigrams in the tagging output) is \u2264 5%. 1 We refer to our entire approach using greedy minimization followed by EM training as DMLC + EM.", |
|
"cite_spans": [ |
|
{ |
|
"start": 854, |
|
"end": 855, |
|
"text": "1", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Fitting the Model Using EM", |
|
"sec_num": "4.3" |
|
}, |
|
{ |
|
"text": "The DMLC algorithm is directly suited towards parallelization across many machines. We turn to Pregel (Malewicz et al., 2010) , and its open source version Giraph (Apa, 2013) . In these systems the computation proceeds in rounds. In every round, every machine does some local processing and then sends arbitrary messages to other machines. Semantically, we think of the communication graph as fixed, and in each round each vertex performs some local computation and then sends messages to its neighbors. This mode of parallel programming directs the programmers to \"Think like a vertex.\"", |
|
"cite_spans": [ |
|
{ |
|
"start": 102, |
|
"end": 125, |
|
"text": "(Malewicz et al., 2010)", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 156, |
|
"end": 174, |
|
"text": "Giraph (Apa, 2013)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Distributed Implementation", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "The specific systems like Pregel and Giraph build infrastructure that ensures that the overall system 1 For more details on the alternating EM strategy and how initialization with minimized models improve EM performance in alternating iterations, refer to (Ravi and Knight, 2009) . is fault tolerant, efficient, and fast. In addition, they provide implementation of commonly used distributed data structures, such as, for example global counters. The programmer's job is simply to specify the code that each vertex will run at every round.", |
|
"cite_spans": [ |
|
{ |
|
"start": 256, |
|
"end": 279, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Distributed Implementation", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "We implemented the DMLC algorithm in Pregel. The implementation is straightforward and given in Algorithm 2. The multi-set M of Algorithm 1 is represented as a global counter in Algorithm 2. The message passing ( Step 3) and counter update (Step 4) steps update this global counter and hence perform the role of Step 3 of Algorithm 1.", |
|
"cite_spans": [ |
|
{ |
|
"start": 211, |
|
"end": 212, |
|
"text": "(", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Distributed Implementation", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "Step 5 selects the label with largest count, which is equivalent to the greedy label picking step 6 of Algorithm 1. Finally steps 6, 7, and 8 update the tag assignment of each vertex performing the roles of steps 7, 8, and 9, respectively, of Algorithm 1.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Distributed Implementation", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "The implementation described above directly copies the sequential algorithm. Here we describe additional steps we took to further improve the parallel running times.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speeding up the Algorithm", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "Singleton Sets: As the parallel algorithm proceeds, the set of feasible sets associated with a node slowly decreases. At some point there is only one tag that a node can take on, however this tag is rare, and so it takes a while for it to be selected using the greedy strategy. Nevertheless, if a node and one of its neighbors have only a single tag left, then it is safe to assign the unique label 2 .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Speeding up the Algorithm", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "As is often the case, the bottleneck in parallel computations is the communication. To reduce the amount of communication we reduce the graph on the fly, removing nodes and edges once they no longer play a role in the computation. This simple modification decreases the communication time in later rounds as the total size of the problem shrinks.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Modifying the Graph:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In this Section, we describe the experimental setup for various tasks, settings and compare empirical performance of our method against several existing baselines. The performance results for all systems (on all tasks) are measured in terms of tagging accuracy, i.e. % of tokens from the test corpus that were labeled correctly by the system.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiments and Results", |
|
"sec_num": "6" |
|
}, |
|
{ |
|
"text": "6.1.1 Tagging Using a Complete Dictionary Data: We use a standard test set (consisting of 24,115 word tokens from the Penn Treebank) for the POS tagging task. The tagset consists of 45 distinct tag labels and the dictionary contains 57,388 word/tag pairs derived from the entire Penn Treebank. Per-token ambiguity for the test data is about 1.5 tags/token. In addition to the standard 24k dataset, we also train and test on larger data sets-973k tokens from the Penn Treebank, 3M tokens from PTB+Europarl (Koehn, 2005) data.", |
|
"cite_spans": [ |
|
{ |
|
"start": 505, |
|
"end": 518, |
|
"text": "(Koehn, 2005)", |
|
"ref_id": "BIBREF18" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Part-of-Speech Tagging Task", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "Methods: We evaluate and compare performance for POS tagging using four different methods that employ the model minimization idea combined with EM training:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Part-of-Speech Tagging Task", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "\u2022 EM: Training a bigram HMM model using EM algorithm (Merialdo, 1994) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 53, |
|
"end": 69, |
|
"text": "(Merialdo, 1994)", |
|
"ref_id": "BIBREF21" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Part-of-Speech Tagging Task", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "\u2022 ILP + EM: Minimizing grammar size using integer linear programming, followed by EM training (Ravi and Knight, 2009) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 94, |
|
"end": 117, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Part-of-Speech Tagging Task", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "\u2022 MIN-GREEDY + EM: Minimizing grammar size using the two-step greedy method (Ravi et al., 2010b ).", |
|
"cite_spans": [ |
|
{ |
|
"start": 76, |
|
"end": 95, |
|
"text": "(Ravi et al., 2010b", |
|
"ref_id": "BIBREF26" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Part-of-Speech Tagging Task", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "\u2022 DMLC + EM: This work. Results: Table 1 shows the results for POS tagging on English Penn Treebank data. On the smaller test datasets, all of the model minimization strategies (methods 2, 3, 4) tend to perform equally well, yielding state-of-the-art results and large improvement over standard EM. When training (and testing) on larger corpora sizes, DMLC yields the best reported performance on this task to date. A major advantage of the new method is that it can easily scale to large corpora sizes and the distributed nature of the algorithm still permits fast, efficient optimization of the global objective function. So, unlike the earlier methods (such as MIN-GREEDY) it is fast enough to run on several millions of tokens to yield additional performance gains (shown in last column).", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 33, |
|
"end": 40, |
|
"text": "Table 1", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Part-of-Speech Tagging Task", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "We also observe a significant speedup when using the parallelized version of the DMLC algorithm. Performing model minimization on the 24k tokens dataset takes 55 seconds on a single machine, whereas parallelization permits model minimization to be feasible even on large datasets. Fig 1 shows the running time for DMLC when run on a cluster of 100 machines. We vary the input data size from 1M word tokens to about 8M word tokens, while holding the resources constant. Both the algorithm and its distributed implementation in DMLC are linear time operations as evident by the plot. In fact, for comparison, we also plot a straight line passing through the first two runtimes. The straight line essentially plots runtimes corresponding to a linear speedup. DMLC clearly achieves better runtimes showing even better than linear speedup. The reason for this is that distributed version has a constant overhead for initialization, independent of the data size. While the running time for rest of the implementation is linear in data size. Thus, as the data size becomes larger, the constant overhead becomes less significant, and the distributed implementation appears to complete slightly faster as data size increases. Figure 1 : Runtime vs. data size (measured in # of word tokens) on 100 machines. For comparison, we also plot a straight line passing through the first two runtimes. The straight line essentially plots runtimes corresponding to a linear speedup. DMLC clearly achieves better runtimes showing a better than linear speedup.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 281, |
|
"end": 293, |
|
"text": "Fig 1 shows", |
|
"ref_id": "FIGREF0" |
|
}, |
|
{ |
|
"start": 1218, |
|
"end": 1226, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Speedups:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We also evaluate our approach for POS tagging under other resource-constrained scenarios. Obtain-", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Tagging Using Incomplete Dictionaries", |
|
"sec_num": "6.1.2" |
|
}, |
|
{ |
|
"text": "Tagging accuracy (%) te=24k te=973k tr=24k tr=973k tr=3.7M 1. EM 81.7 82.3 2. ILP + EM (Ravi and Knight, 2009) 91.6 3. MIN-GREEDY + EM (Ravi et al., 2010b) 91.6 87.1 4. DMLC + EM (this work) 91.4 87.5 87.8 ing a complete dictionary is often difficult, especially for new domains. To verify the utility of our method when the input dictionary is incomplete, we evaluate against standard datasets used in previous work (Garrette and Baldridge, 2012) and compare against the previous best reported performance for the same task. In all the experiments (described here and in subsequent sections), we use the following terminology-raw data refers to unlabeled text used by different methods (for model minimization or other unsupervised training procedures such as EM), dictionary consists of word/tag entries that are legal, and test refers to data over which tagging evaluation is performed.", |
|
"cite_spans": [ |
|
{ |
|
"start": 87, |
|
"end": 110, |
|
"text": "(Ravi and Knight, 2009)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 135, |
|
"end": 155, |
|
"text": "(Ravi et al., 2010b)", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 417, |
|
"end": 447, |
|
"text": "(Garrette and Baldridge, 2012)", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Method", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "English Data: For English POS tagging with incomplete dictionary, we evaluate on the Penn Treebank (Marcus et al., 1993) data. Following (Garrette and Baldridge, 2012), we extracted a word-tag dictionary from sections 00-15 (751,059 tokens) consisting of 39,087 word types, 45,331 word/tag entries, a per-type ambiguity of 1.16 yielding a pertoken ambiguity of 2.21 on the raw corpus (treating unknown words as having all 45 possible tags). As in their setup, we then use the first 47,996 tokens of section 16 as raw data and perform final evaluation on the sections 22-24. We use the raw corpus along with the unlabeled test data to perform model minimization and EM training. Unknown words are allowed to have all possible tags in both these procedures.", |
|
"cite_spans": [ |
|
{ |
|
"start": 99, |
|
"end": 120, |
|
"text": "(Marcus et al., 1993)", |
|
"ref_id": "BIBREF20" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Method", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Italian Data: The minimization strategy presented here is a general-purpose method that does not require any specific tuning and works for other languages as well. To demonstrate this, we also perform evaluation on a different language (Italian) us-ing the TUT corpus (Bosco et al., 2000) . Following (Garrette and Baldridge, 2012), we use the same data splits as their setting. We take the first half of each of the five sections to build the word-tag dictionary, the next quarter as raw data and the last quarter as test data. The dictionary was constructed from 41,000 tokens comprised of 7,814 word types, 8,370 word/tag pairs, per-type ambiguity of 1.07 and a per-token ambiguity of 1.41 on the raw data. The raw data consisted of 18,574 tokens and the test contained 18,763 tokens. We use the unlabeled corpus from the raw and test data to perform model minimization followed by unsupervised EM training.", |
|
"cite_spans": [ |
|
{ |
|
"start": 268, |
|
"end": 288, |
|
"text": "(Bosco et al., 2000)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Method", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Other Languages: In order to test the effectiveness of our method in other non-English settings, we also report the performance of our method on several other Indo-European languages using treebank data from CoNLL-X and CoNLL-2007 shared tasks on dependency parsing (Buchholz and Marsi, 2006; Nivre et al., 2007) . The corpus statistics for the five languages (Danish, Greek, Italian, Portuguese and Spanish) are listed below. For each language, we construct a dictionary from the raw training data. The unlabeled corpus from the raw training and test data is used to perform model minimization followed by unsupervised EM training. As before, unknown words are allowed to have all possible tags. We report the final tagging performance on the test data and compare it to baseline EM. Garrette and Baldridge (2012) treat unknown words (words that appear in the raw text but are missing from the dictionary) in a special manner and use several heuristics to perform better initialization for such words (for example, the probability that an unknown word is associated with a particular tag is conditioned on the openness of the tag). They also use an auto-supervision technique to smooth counts learnt from EM onto new words encountered during testing. In contrast, we do not apply any such technique for unknown words and allow them to be mapped uniformly to all possible tags in the dictionary. For this particular set of experiments, the only difference from the Garrette and Baldridge (2012) setup is that we include unlabeled text from the test data (but without any dictionary tag labels or special heuristics) to our existing word tokens from raw text for performing model minimization. This is a standard practice used in unsupervised training scenarios (for example, Bayesian inference methods) and in general for scalable techniques where the goal is to perform inference on the same data for which one wishes to produce some structured prediction. Results: Table 2 (column 2) compares previously reported results against our approach for English. We observe that our method obtains a huge improvement over standard EM and gets comparable results to the previous best reported scores for the same task from (Garrette and Baldridge, 2012) . It is encouraging to note that the new system achieves this performance without using any of the carefully-chosen heuristics employed by the previous method. However, we do note that some of these techniques can be easily combined with our method to produce further improvements. Table 2 (column 3) also shows results on Italian POS tagging. We observe that our method achieves significant improvements in tagging accuracy over all the baseline systems including the previous best system (+2.9%). This demonstrates that the method generalizes well to other languages and produces consistent tagging improvements over existing methods for the same task.", |
|
"cite_spans": [ |
|
{ |
|
"start": 266, |
|
"end": 292, |
|
"text": "(Buchholz and Marsi, 2006;", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 293, |
|
"end": 312, |
|
"text": "Nivre et al., 2007)", |
|
"ref_id": "BIBREF23" |
|
}, |
|
{ |
|
"start": 1465, |
|
"end": 1494, |
|
"text": "Garrette and Baldridge (2012)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 2216, |
|
"end": 2246, |
|
"text": "(Garrette and Baldridge, 2012)", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 1967, |
|
"end": 1974, |
|
"text": "Table 2", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 2529, |
|
"end": 2536, |
|
"text": "Table 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Method", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Results for POS tagging on CoNLL data in five different languages are displayed in Figure 2 . Note that the proportion of raw data in test versus train (from the standard CoNLL shared tasks) is much smaller compared to the earlier experimental settings. In general, we observe that adding more raw data for EM training improves the tagging quality (same trend observed earlier in Table 1 : column 2 versus column 3). Despite this, DMLC + EM still achieves significant improvements over the baseline EM system on multiple languages (as shown in Figure 2 ). An additional advantage of the new method is that it can easily scale to larger corpora and it produces a much more compact grammar that can be efficiently incorporated for EM training.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 83, |
|
"end": 91, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
}, |
|
{ |
|
"start": 380, |
|
"end": 387, |
|
"text": "Table 1", |
|
"ref_id": "TABREF0" |
|
}, |
|
{ |
|
"start": 544, |
|
"end": 552, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Method", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Learning part-of-speech taggers for severely lowresource languages (e.g., Malagasy) is very challenging. In addition to scarce (token-supervised) labeled resources, the tag dictionaries available for training taggers are tiny compared to other languages such as English. Garrette and Baldridge (2013) combine various supervised and semi-supervised learning algorithms into a common POS tagger training pipeline to address some of these challenges. They also report tagging accuracy improvements on low-resource languages when using the combined system over any single algorithm. Their system has four main parts, in order: (1) Tag dictionary expansion using label propagation algorithm, (2) Weighted model minimization, (3) Expectation maximization (EM) training of HMMs using auto-supervision, (4) MaxEnt Markov Model (MEMM) training. The entire procedure results in a trained tagger model that can then be applied to tag any raw data. 3 Step 2 in this procedure involves (Garrette and Baldridge, 2012) 88.52 72.86 4. DMLC + EM (this work) 88.11 75.79 Table 2 : Part-of-Speech tagging accuracy using PTB sections 00-15 and TUT to build the tag dictionary. For comparison, we also include the results for the previously reported state-of-the-art system (method 3) for the same task.", |
|
"cite_spans": [ |
|
{ |
|
"start": 271, |
|
"end": 300, |
|
"text": "Garrette and Baldridge (2013)", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 937, |
|
"end": 938, |
|
"text": "3", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 973, |
|
"end": 1003, |
|
"text": "(Garrette and Baldridge, 2012)", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 1029, |
|
"end": 1040, |
|
"text": "(this work)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1053, |
|
"end": 1060, |
|
"text": "Table 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Tagging for Low-Resource Languages", |
|
"sec_num": "6.1.3" |
|
}, |
|
{ |
|
"text": "Tagging accuracy (%) Total Known Unknown Low-resource tagging using (Garrette and Baldridge, 2013) 80.7 (70.2) 87.6 (90.3) 66.1 (45.1) Low-resource tagging using DMLC + EM (this work) 81.1 (70.8) 87.9 (90.3) 66.7 (46.5) Table 3 : Part-of-Speech tagging accuracy for a low-resource language (Malagasy) on All/Known/Unknown tokens in the test data. Tagging performance is shown for multiple experiments using different (incomplete) dictionary sizes: (a) small, (b) tiny (shown in parentheses). The new method (row 2) significantly outperforms the existing method with p < 0.01 for small dictionary and p < 0.05 for tiny dictionary.", |
|
"cite_spans": [ |
|
{ |
|
"start": 68, |
|
"end": 98, |
|
"text": "(Garrette and Baldridge, 2013)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 220, |
|
"end": 227, |
|
"text": "Table 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Method", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "a weighted version of model minimization which uses the multi-step greedy approach from Ravi et al. (2010b) enhanced with additional heuristics that uses tag weights learnt via label propagation (in Step 1) within the minimization process. We replace the model minimization procedure in their Step 2 with our method (DMLC + EM) and directly compare this new system with their approach in terms of tagging accuracy. Note for all other steps in the pipeline we follow the same procedure (and run the same code) as Garrette and Baldridge (2013) , including the same smoothing procedure for EM initialization in Step 3.", |
|
"cite_spans": [ |
|
{ |
|
"start": 88, |
|
"end": 107, |
|
"text": "Ravi et al. (2010b)", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 512, |
|
"end": 541, |
|
"text": "Garrette and Baldridge (2013)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Method", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We use the exact same setup as Garrette and Baldridge (2013) and run experiments on Malagasy, an Austronesian language spoken in Madagascar. We use the publicly available data 4 : 100k raw tokens for training, a word-tag dictionary acquired with 4 hours of human annotation effort (used for type-supervision), and a held-out test dataset (5341 tokens). We provide the unlabeled corpus from the raw training data along with the word-tag dictionary as input to model minimization and evaluate on the test corpus. We run multiple experiments for different (incomplete) dictionary scenarios: (a) small = 2773 word/tag pairs, (b) tiny = 329 word/tag pairs.", |
|
"cite_spans": [ |
|
{ |
|
"start": 54, |
|
"end": 60, |
|
"text": "(2013)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Results: Table 3 shows results on Malagasy data comparing a system that employs (unweighted) 4 github.com/ dhgarrette/low-resource-pos-tagging-2013 DMLC against the existing state-of-the-art system that incorporates a multi-step weighted model minimization combined with additional heuristics. We observe that switching to the new model minimization procedure alone yields significant improvement in tagging accuracy under both dictionary scenarios. It is encouraging that a better minimization procedure also leads to higher tagging quality on the unknown word tokens (column 4 in the table), even when the input dictionary is tiny.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 9, |
|
"end": 16, |
|
"text": "Table 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Data:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Compared to POS tagging, a more challenging task is learning supertaggers for lexicalized grammar formalisms such as Combinatory Categorial Grammar (CCG) (Steedman, 2000) . For example, CCGbank (Hockenmaier and Steedman, 2007) contains 1241 distinct supertags (lexical categories) and the most ambiguous word has 126 supertags. This provides a much more challenging starting point for the semi-supervised methods typically applied to the task. Yet, this is an important task since creating grammars and resources for CCG parsers for new domains and languages is highly labor-and knowledge-intensive.", |
|
"cite_spans": [ |
|
{ |
|
"start": 154, |
|
"end": 170, |
|
"text": "(Steedman, 2000)", |
|
"ref_id": "BIBREF28" |
|
}, |
|
{ |
|
"start": 194, |
|
"end": 226, |
|
"text": "(Hockenmaier and Steedman, 2007)", |
|
"ref_id": "BIBREF16" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Supertagging", |
|
"sec_num": "6.2" |
|
}, |
|
{ |
|
"text": "As described earlier, our approach scales easily to large datasets as well as label sizes. To evaluate it on the supertagging task, we use the same dataset from (Ravi et al., 2010a) and compare against their baseline method that uses an modified (two-step) version Method Supertagging accuracy (%) Ambiguous Total 1. EM 38.7 45.6 2. ILP * + EM (Ravi et al., 2010a) 52.1 57.3 3. DMLC + EM (this work) 55.9 59.3 Table 4 : Results for unsupervised supertagging with a dictionary. Here, we report the total accuracy as well as accuracy on just the ambiguous tokens (i.e., tokens which have more than one tagging possibility). * The baseline method 2 requires several pre-processing steps in order to run feasibly for this task (described in Section 6.2). In contrast, the new approach (DMLC) runs fast and also permits efficient parallelization.", |
|
"cite_spans": [ |
|
{ |
|
"start": 161, |
|
"end": 181, |
|
"text": "(Ravi et al., 2010a)", |
|
"ref_id": "BIBREF25" |
|
}, |
|
{ |
|
"start": 344, |
|
"end": 364, |
|
"text": "(Ravi et al., 2010a)", |
|
"ref_id": "BIBREF25" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 410, |
|
"end": 417, |
|
"text": "Table 4", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Supertagging", |
|
"sec_num": "6.2" |
|
}, |
|
{ |
|
"text": "of the ILP formulation for model minimization.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Supertagging", |
|
"sec_num": "6.2" |
|
}, |
|
{ |
|
"text": "We use the CCGbank data for this experiment. This data was created by semi-automatically converting the Penn Treebank to CCG derivations (Hockenmaier and Steedman, 2007) . We use the standard splits of the data used in semisupervised tagging experiments (Banko and Moore, 2004)-sections 0-18 for training (i.e., to construct the word-tag dictionary), and sections 22-24 for test.", |
|
"cite_spans": [ |
|
{ |
|
"start": 137, |
|
"end": 169, |
|
"text": "(Hockenmaier and Steedman, 2007)", |
|
"ref_id": "BIBREF16" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Results: Table 4 compares the results for two baseline systems-standard EM (method 1), and a previously reported system using model minimization (method 2) for the same task. We observe that DMLC produces better taggings than either of these and yields significant improvement in accuracy (+2% overall, +3.8% on ambiguous tokens).", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 9, |
|
"end": 16, |
|
"text": "Table 4", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Data:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Note that it is not feasible to run the ILP-based baseline (method 2 in the table) directly since it is very slow in practice, so Ravi et al. (2010a) use a set of pre-processing steps to prune the original grammar size (unique tag pairs) from >1M to several thousand entries followed by a modified twostep ILP minimization strategy. This is required to permit their model minimization step to be run in a feasible manner. On the other hand, the new approach DMLC (method 3) scales better even when the data/label sizes are large, hence it can be run with the full data using the original model minimization formulation (rather than a two-step heuristic). Ravi et al. (2010a) also report further improvements using an alternative approach involving an ILP-based weighted minimization procedure. In Section 7 we briefly discuss how the DMLC method can be extended to this setting and combined with other similar methods.", |
|
"cite_spans": [ |
|
{ |
|
"start": 130, |
|
"end": 149, |
|
"text": "Ravi et al. (2010a)", |
|
"ref_id": "BIBREF25" |
|
}, |
|
{ |
|
"start": 655, |
|
"end": 674, |
|
"text": "Ravi et al. (2010a)", |
|
"ref_id": "BIBREF25" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We present a fast, efficient model minimization algorithm for unsupervised tagging that improves upon previous two-step heuristics. We show that under a fairly natural assumption of c-feasibility the solution obtained by our minimization algorithm is O(c log m)-approximate to the optimal. Although in the case of two-step heuristics, the first step guarantees an O(log m)-approximation, the second step, which is required to get a consistent solution, can introduce many additional labels resulting in a solution arbitrarily away from the optimal. Our one step approach ensures consistency at each step of the algorithm, while the c-feasibility assumption means that the solution does not diverge too much from the optimal in each iteration.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusion", |
|
"sec_num": "7" |
|
}, |
|
{ |
|
"text": "In addition to proving approximation guarantees for the new algorithm, we show that it is parallelizable, allowing us to easily scale to larger datasets than previously explored. Our results show that the algorithm achieves state-of-the-art performance, outperforming existing methods on several different tasks (both POS tagging and supertagging) and works well even with incomplete dictionaries and extremely low-resource languages like Malagasy.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusion", |
|
"sec_num": "7" |
|
}, |
|
{ |
|
"text": "For future work, it would be interesting to apply a weighted version of the DMLC algorithm where labels (i.e., tag pairs) can have different weight distributions instead of uniform weights. Our algorithm can be extended to allow an input weight distribution to be specified for minimization. In order to initialize the weights we could use existing strategies such as grammar-informed initialization (Ravi et al., 2010a) or output distributions learnt via other methods such as label propagation (Garrette and Baldridge, 2013) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 400, |
|
"end": 420, |
|
"text": "(Ravi et al., 2010a)", |
|
"ref_id": "BIBREF25" |
|
}, |
|
{ |
|
"start": 496, |
|
"end": 526, |
|
"text": "(Garrette and Baldridge, 2013)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion and Conclusion", |
|
"sec_num": "7" |
|
}, |
|
{ |
|
"text": "Transactions of the Association for Computational Linguistics, 2 (2014) 105-118. Action Editor: Sharon Goldwater. Submitted 11/2013; Revised 2/2014; Published 4/2014. c 2014 Association for Computational Linguistics.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We must judiciously initialize the global counter to take care of this assignment, but this is easily accomplished.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "For more details, refer(Garrette and Baldridge, 2013).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Part-ofspeech tagging in context", |
|
"authors": [ |
|
{ |
|
"first": "Michele", |
|
"middle": [], |
|
"last": "Banko", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Robert", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Moore", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of COLING", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "556--561", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Michele Banko and Robert C. Moore. 2004. Part-of- speech tagging in context. In Proceedings of COLING, pages 556-561.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "The Minimum Description Length Principle in Coding and Modeling", |
|
"authors": [ |
|
{ |
|
"first": "Jorma", |
|
"middle": [], |
|
"last": "Andrew R Barron", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bin", |
|
"middle": [], |
|
"last": "Rissanen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Yu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "IEEE Transactions of Information Theory", |
|
"volume": "44", |
|
"issue": "6", |
|
"pages": "2743--2760", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Andrew R Barron, Jorma Rissanen, and Bin Yu. 1998. The Minimum Description Length Principle in Cod- ing and Modeling. IEEE Transactions of Information Theory, 44(6):2743-2760.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Building a Treebank for Italian: a data-driven annotation schema", |
|
"authors": [ |
|
{ |
|
"first": "Cristina", |
|
"middle": [], |
|
"last": "Bosco", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Vincenzo", |
|
"middle": [], |
|
"last": "Lombardo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daniela", |
|
"middle": [], |
|
"last": "Vassallo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Leonardo", |
|
"middle": [], |
|
"last": "Lesmo", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of the Second International Conference on Language Resources and Evaluation LREC-2000", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "99--105", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Cristina Bosco, Vincenzo Lombardo, Daniela Vassallo, and Leonardo Lesmo. 2000. Building a Treebank for Italian: a data-driven annotation schema. In Proceed- ings of the Second International Conference on Lan- guage Resources and Evaluation LREC-2000, pages 99-105.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Conll-x shared task on multilingual dependency parsing", |
|
"authors": [ |
|
{ |
|
"first": "Sabine", |
|
"middle": [], |
|
"last": "Buchholz", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Erwin", |
|
"middle": [], |
|
"last": "Marsi", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of CoNLL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "149--164", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sabine Buchholz and Erwin Marsi. 2006. Conll-x shared task on multilingual dependency parsing. In Proceed- ings of CoNLL, pages 149-164.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Two decades of unsupervised POS induction: How far have we come?", |
|
"authors": [ |
|
{ |
|
"first": "Christos", |
|
"middle": [], |
|
"last": "Christodoulopoulos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sharon", |
|
"middle": [], |
|
"last": "Goldwater", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Steedman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "575--584", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Christos Christodoulopoulos, Sharon Goldwater, and Mark Steedman. 2010. Two decades of unsupervised POS induction: How far have we come? In Proceed- ings of the Conference on Empirical Methods in Natu- ral Language Processing (EMNLP), pages 575-584.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Unsupervised partof-speech tagging with bilingual graph-based projections", |
|
"authors": [ |
|
{ |
|
"first": "Dipanjan", |
|
"middle": [], |
|
"last": "Das", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Slav", |
|
"middle": [], |
|
"last": "Petrov", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "600--609", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dipanjan Das and Slav Petrov. 2011. Unsupervised part- of-speech tagging with bilingual graph-based projec- tions. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Hu- man Language Technologies -Volume 1, pages 600- 609.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Maximum likelihood from incomplete data via the EM algorithm", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [ |
|
"P" |
|
], |
|
"last": "Dempster", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Laird", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [ |
|
"B" |
|
], |
|
"last": "Rubin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1977, |
|
"venue": "Journal of the Royal Statistical Society", |
|
"volume": "39", |
|
"issue": "1", |
|
"pages": "1--38", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "A. P. Dempster, N. M. Laird, and D. B. Rubin. 1977. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Se- ries B, 39(1):1-38.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Posterior regularization for structured latent variable models", |
|
"authors": [ |
|
{ |
|
"first": "Kuzman", |
|
"middle": [], |
|
"last": "Ganchev", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jo\u00e3o", |
|
"middle": [], |
|
"last": "Gra\u00e7a", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jennifer", |
|
"middle": [], |
|
"last": "Gillenwater", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ben", |
|
"middle": [], |
|
"last": "Taskar", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Journal of Machine Learning Research", |
|
"volume": "11", |
|
"issue": "", |
|
"pages": "2001--2049", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kuzman Ganchev, Jo\u00e3o Gra\u00e7a, Jennifer Gillenwater, and Ben Taskar. 2010. Posterior regularization for struc- tured latent variable models. Journal of Machine Learning Research, 11:2001-2049.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Typesupervised Hidden Markov Models for part-of-speech tagging with incomplete tag dictionaries", |
|
"authors": [ |
|
{ |
|
"first": "Dan", |
|
"middle": [], |
|
"last": "Garrette", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jason", |
|
"middle": [], |
|
"last": "Baldridge", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2012, |
|
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "821--831", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dan Garrette and Jason Baldridge. 2012. Type- supervised Hidden Markov Models for part-of-speech tagging with incomplete tag dictionaries. In Proceed- ings of the Conference on Empirical Methods in Nat- ural Language Processing and Computational Natu- ral Language Learning (EMNLP-CoNLL), pages 821- 831.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Learning a part-of-speech tagger from two hours of annotation", |
|
"authors": [ |
|
{ |
|
"first": "Dan", |
|
"middle": [], |
|
"last": "Garrette", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jason", |
|
"middle": [], |
|
"last": "Baldridge", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "138--147", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Dan Garrette and Jason Baldridge. 2013. Learning a part-of-speech tagger from two hours of annotation. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguis- tics: Human Language Technologies, pages 138-147.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Part-of-speech tagging for Twitter: annotation, features, and experiments", |
|
"authors": [ |
|
{ |
|
"first": "Kevin", |
|
"middle": [], |
|
"last": "Gimpel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Nathan", |
|
"middle": [], |
|
"last": "Schneider", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "O'", |
|
"middle": [], |
|
"last": "Brendan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dipanjan", |
|
"middle": [], |
|
"last": "Connor", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Das", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jacob", |
|
"middle": [], |
|
"last": "Mills", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Eisenstein", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dani", |
|
"middle": [], |
|
"last": "Heilman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jeffrey", |
|
"middle": [], |
|
"last": "Yogatama", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Noah", |
|
"middle": [ |
|
"A" |
|
], |
|
"last": "Flanigan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Smith", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "42--47", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kevin Gimpel, Nathan Schneider, Brendan O'Connor, Dipanjan Das, Daniel Mills, Jacob Eisenstein, Michael Heilman, Dani Yogatama, Jeffrey Flanigan, and Noah A. Smith. 2011. Part-of-speech tagging for Twitter: annotation, features, and experiments. In Pro- ceedings of the 49th Annual Meeting of the Associa- tion for Computational Linguistics: Human Language Technologies: short papers -Volume 2, pages 42-47.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "EM can find pretty good HMM POS-taggers (when given a good start)", |
|
"authors": [ |
|
{ |
|
"first": "Yoav", |
|
"middle": [], |
|
"last": "Goldberg", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Meni", |
|
"middle": [], |
|
"last": "Adler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Elhadad", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of ACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "746--754", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yoav Goldberg, Meni Adler, and Michael Elhadad. 2008. EM can find pretty good HMM POS-taggers (when given a good start). In Proceedings of ACL, pages 746-754.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "A fully Bayesian approach to unsupervised part-ofspeech tagging", |
|
"authors": [ |
|
{ |
|
"first": "Sharon", |
|
"middle": [], |
|
"last": "Goldwater", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Thomas", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Griffiths", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "ACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sharon Goldwater and Thomas L. Griffiths. 2007. A fully Bayesian approach to unsupervised part-of- speech tagging. In ACL.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Experimental analysis of approximation algorithms for the vertex cover and set covering problems", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Fernando", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Cludio", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Gomes", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Panos", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Meneses", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Gerardo", |
|
"middle": [], |
|
"last": "Pardalos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Valdisio", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Viana", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Fernando C. Gomes, Cludio N. Meneses, Panos M. Pardalos, and Gerardo Valdisio R. Viana. 2006. Ex- perimental analysis of approximation algorithms for the vertex cover and set covering problems.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Weakly supervised part-of-speech tagging for morphologically-rich, resource-scarce languages", |
|
"authors": [ |
|
{ |
|
"first": "Saidul", |
|
"middle": [], |
|
"last": "Kazi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Vincent", |
|
"middle": [], |
|
"last": "Hasan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Ng", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 12th Conference on the European Chapter of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "363--371", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kazi Saidul Hasan and Vincent Ng. 2009. Weakly super- vised part-of-speech tagging for morphologically-rich, resource-scarce languages. In Proceedings of the 12th Conference on the European Chapter of the Associa- tion for Computational Linguistics, pages 363-371.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "CCGbank: A corpus of CCG derivations and dependency structures extracted from the Penn Treebank", |
|
"authors": [ |
|
{ |
|
"first": "Julia", |
|
"middle": [], |
|
"last": "Hockenmaier", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Steedman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Computational Linguistics", |
|
"volume": "33", |
|
"issue": "3", |
|
"pages": "355--396", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Julia Hockenmaier and Mark Steedman. 2007. CCG- bank: A corpus of CCG derivations and dependency structures extracted from the Penn Treebank. Compu- tational Linguistics, 33(3):355-396.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Why doesn't EM find good HMM POS-taggers?", |
|
"authors": [ |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Johnson", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of the Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "296--305", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mark Johnson. 2007. Why doesn't EM find good HMM POS-taggers? In Proceedings of the Joint Conference on Empirical Methods in Natural Language Process- ing and Computational Natural Language Learning (EMNLP-CoNLL), pages 296-305.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "Europarl: A Parallel Corpus for Statistical Machine Translation", |
|
"authors": [ |
|
{ |
|
"first": "Philipp", |
|
"middle": [], |
|
"last": "Koehn", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Machine Translation Summit X", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "79--86", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Philipp Koehn. 2005. Europarl: A Parallel Corpus for Statistical Machine Translation. In Machine Transla- tion Summit X, pages 79-86.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Pregel: a system for largescale graph processing", |
|
"authors": [ |
|
{ |
|
"first": "Grzegorz", |
|
"middle": [], |
|
"last": "Malewicz", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Matthew", |
|
"middle": [ |
|
"H" |
|
], |
|
"last": "Austern", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Aart", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "James", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "Bik", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ilan", |
|
"middle": [], |
|
"last": "Dehnert", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Naty", |
|
"middle": [], |
|
"last": "Horn", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Grzegorz", |
|
"middle": [], |
|
"last": "Leiser", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Czajkowski", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the 2010 ACM SIGMOD International Conference on Management of data", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "135--146", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Grzegorz Malewicz, Matthew H. Austern, Aart J.C Bik, James C. Dehnert, Ilan Horn, Naty Leiser, and Grze- gorz Czajkowski. 2010. Pregel: a system for large- scale graph processing. In Proceedings of the 2010 ACM SIGMOD International Conference on Manage- ment of data, pages 135-146.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Building a large annotated corpus of English: The Penn Treebank", |
|
"authors": [ |
|
{ |
|
"first": "Mitchell", |
|
"middle": [ |
|
"P" |
|
], |
|
"last": "Marcus", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Beatrice", |
|
"middle": [], |
|
"last": "Santorini", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mary", |
|
"middle": [ |
|
"Ann" |
|
], |
|
"last": "Marcinkiewicz", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1993, |
|
"venue": "Computational Linguistics", |
|
"volume": "19", |
|
"issue": "2", |
|
"pages": "313--330", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mitchell P. Marcus, Beatrice Santorini, and Mary Ann Marcinkiewicz. 1993. Building a large annotated cor- pus of English: The Penn Treebank. Computational Linguistics, 19(2):313-330.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Tagging English text with a probabilistic model", |
|
"authors": [ |
|
{ |
|
"first": "Bernard", |
|
"middle": [], |
|
"last": "Merialdo", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "Computational Linguistics", |
|
"volume": "20", |
|
"issue": "2", |
|
"pages": "155--171", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bernard Merialdo. 1994. Tagging English text with a probabilistic model. Computational Linguistics, 20(2):155-171.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Crouching Dirichlet, Hidden Markov Model: Unsupervised POS tagging with context local tag generation", |
|
"authors": [ |
|
{ |
|
"first": "Taesun", |
|
"middle": [], |
|
"last": "Moon", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Katrin", |
|
"middle": [], |
|
"last": "Erk", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jason", |
|
"middle": [], |
|
"last": "Baldridge", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "196--206", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Taesun Moon, Katrin Erk, and Jason Baldridge. 2010. Crouching Dirichlet, Hidden Markov Model: Unsu- pervised POS tagging with context local tag genera- tion. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 196- 206.", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "The CoNLL 2007 shared task on dependency parsing", |
|
"authors": [ |
|
{ |
|
"first": "Joakim", |
|
"middle": [], |
|
"last": "Nivre", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Johan", |
|
"middle": [], |
|
"last": "Hall", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sandra", |
|
"middle": [], |
|
"last": "K\u00fcbler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ryan", |
|
"middle": [], |
|
"last": "Mcdonald", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jens", |
|
"middle": [], |
|
"last": "Nilsson", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sebastian", |
|
"middle": [], |
|
"last": "Riedel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Deniz", |
|
"middle": [], |
|
"last": "Yuret", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of the CoNLL Shared Task Session of EMNLP-CoNLL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "915--932", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Joakim Nivre, Johan Hall, Sandra K\u00fcbler, Ryan McDon- ald, Jens Nilsson, Sebastian Riedel, and Deniz Yuret. 2007. The CoNLL 2007 shared task on dependency parsing. In Proceedings of the CoNLL Shared Task Session of EMNLP-CoNLL, pages 915-932.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Minimized models for unsupervised part-of-speech tagging", |
|
"authors": [ |
|
{ |
|
"first": "Sujith", |
|
"middle": [], |
|
"last": "Ravi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kevin", |
|
"middle": [], |
|
"last": "Knight", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the Joint Conferenceof the 47th Annual Meeting of the Association for Computational Linguistics and the 4th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing (ACL-IJCNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "504--512", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sujith Ravi and Kevin Knight. 2009. Minimized models for unsupervised part-of-speech tagging. In Proceed- ings of the Joint Conferenceof the 47th Annual Meet- ing of the Association for Computational Linguistics and the 4th International Joint Conference on Natural Language Processing of the Asian Federation of Natu- ral Language Processing (ACL-IJCNLP), pages 504- 512.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "Minimized models and grammar-informed initialization for supertagging with highly ambiguous lexicons", |
|
"authors": [ |
|
{ |
|
"first": "Sujith", |
|
"middle": [], |
|
"last": "Ravi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jason", |
|
"middle": [], |
|
"last": "Baldridge", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kevin", |
|
"middle": [], |
|
"last": "Knight", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics (ACL)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "495--503", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sujith Ravi, Jason Baldridge, and Kevin Knight. 2010a. Minimized models and grammar-informed initializa- tion for supertagging with highly ambiguous lexicons. In Proceedings of the 48th Annual Meeting of the As- sociation for Computational Linguistics (ACL), pages 495-503.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Fast, greedy model minimization for unsupervised tagging", |
|
"authors": [ |
|
{ |
|
"first": "Sujith", |
|
"middle": [], |
|
"last": "Ravi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ashish", |
|
"middle": [], |
|
"last": "Vaswani", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kevin", |
|
"middle": [], |
|
"last": "Knight", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "David", |
|
"middle": [], |
|
"last": "Chiang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the 23rd International Conference on Computational Linguistics (COLING)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "940--948", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sujith Ravi, Ashish Vaswani, Kevin Knight, and David Chiang. 2010b. Fast, greedy model minimization for unsupervised tagging. In Proceedings of the 23rd In- ternational Conference on Computational Linguistics (COLING), pages 940-948.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Improved unsupervised POS induction using intrinsic clustering quality and a Zipfian constraint", |
|
"authors": [ |
|
{ |
|
"first": "Roi", |
|
"middle": [], |
|
"last": "Reichart", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Raanan", |
|
"middle": [], |
|
"last": "Fattal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ari", |
|
"middle": [], |
|
"last": "Rappoport", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the Fourteenth Conference on Computational Natural Language Learning", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "57--66", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Roi Reichart, Raanan Fattal, and Ari Rappoport. 2010. Improved unsupervised POS induction using intrinsic clustering quality and a Zipfian constraint. In Proceed- ings of the Fourteenth Conference on Computational Natural Language Learning, pages 57-66.", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "The Syntactic Process", |
|
"authors": [ |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Steedman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mark Steedman. 2000. The Syntactic Process. MIT Press, Cambridge, MA, USA.", |
|
"links": null |
|
}, |
|
"BIBREF29": { |
|
"ref_id": "b29", |
|
"title": "A Bayesian LDA-based model for semi-supervised partof-speech tagging", |
|
"authors": [ |
|
{ |
|
"first": "Kristina", |
|
"middle": [], |
|
"last": "Toutanova", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Johnson", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Advances in Neural Information Processing Systems (NIPS)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1521--1528", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kristina Toutanova and Mark Johnson. 2008. A Bayesian LDA-based model for semi-supervised part- of-speech tagging. In Advances in Neural Information Processing Systems (NIPS), pages 1521-1528.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "Under the assumption of c-feasibility, the MLC algorithm achieves a O(c log m) approximation to the minimum label cover problem, where m = i |S i | is the total number of tokens.", |
|
"num": null |
|
}, |
|
"FIGREF1": { |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "Part-of-Speech tagging accuracy for different languages on CoNLL data using incomplete dictionaries.", |
|
"num": null |
|
}, |
|
"TABREF0": { |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"text": "Results for unsupervised part-of-speech tagging on English Penn Treebank dataset. Tagging accuracies for different methods are shown on multiple datasets. te shows the size (number of tokens) in the test data, tr represents the size of the raw text used to perform model minimization.", |
|
"num": null, |
|
"html": null |
|
} |
|
} |
|
} |
|
} |