|
{ |
|
"paper_id": "C10-1033", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T12:55:32.620544Z" |
|
}, |
|
"title": "A Discriminative Latent Variable-Based \"DE\" Classifier for Chinese-English SMT", |
|
"authors": [ |
|
{ |
|
"first": "Jinhua", |
|
"middle": [], |
|
"last": "Du", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Dublin City University", |
|
"location": {} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Andy", |
|
"middle": [], |
|
"last": "Way", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Dublin City University", |
|
"location": {} |
|
}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "Syntactic reordering on the source-side is an effective way of handling word order differences. The (DE) construction is a flexible and ubiquitous syntactic structure in Chinese which is a major source of error in translation quality. In this paper, we propose a new classifier model-discriminative latent variable model (DPLVM)-to classify the DE construction to improve the accuracy of the classification and hence the translation quality. We also propose a new feature which can automatically learn the reordering rules to a certain extent. The experimental results show that the MT systems using the data reordered by our proposed model outperform the baseline systems by 6.42% and 3.08% relative points in terms of the BLEU score on PB-SMT and hierarchical phrase-based MT respectively. In addition, we analyse the impact of DE annotation on word alignment and on the SMT phrase table.", |
|
"pdf_parse": { |
|
"paper_id": "C10-1033", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "Syntactic reordering on the source-side is an effective way of handling word order differences. The (DE) construction is a flexible and ubiquitous syntactic structure in Chinese which is a major source of error in translation quality. In this paper, we propose a new classifier model-discriminative latent variable model (DPLVM)-to classify the DE construction to improve the accuracy of the classification and hence the translation quality. We also propose a new feature which can automatically learn the reordering rules to a certain extent. The experimental results show that the MT systems using the data reordered by our proposed model outperform the baseline systems by 6.42% and 3.08% relative points in terms of the BLEU score on PB-SMT and hierarchical phrase-based MT respectively. In addition, we analyse the impact of DE annotation on word alignment and on the SMT phrase table.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Syntactic structure-based reordering has been shown to be significantly helpful for handling word order issues in phrase-based machine translation (PB-SMT) (Xia and McCord, 2004; Collins et al., 2005; Wang et al., 2007; Li et al., 2007; Elming, 2008; Chang et al., 2009) . It is wellknown that in MT, it is difficult to translate between Chinese-English because of the different word orders (cf. the different orderings of head nouns and relative clauses). Wang et al. (2007) pointed out that Chinese differs from English in several important respects, such as relative clauses appearing before the noun being modified, prepositional phrases often appearing before the head they modify, etc. Chang et al. (2009) argued that many of the structural differences are related to the ubiquitous Chinese structural particle phrase (DE) construction, used for a wide range of noun modification constructions (both single word and clausal) and other uses. They pointed out that DE is a major source of word order error when a Chinese sentence is translated into English due to the different ways that the DE construction can be translated.", |
|
"cite_spans": [ |
|
{ |
|
"start": 156, |
|
"end": 178, |
|
"text": "(Xia and McCord, 2004;", |
|
"ref_id": "BIBREF15" |
|
}, |
|
{ |
|
"start": 179, |
|
"end": 200, |
|
"text": "Collins et al., 2005;", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 201, |
|
"end": 219, |
|
"text": "Wang et al., 2007;", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 220, |
|
"end": 236, |
|
"text": "Li et al., 2007;", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 237, |
|
"end": 250, |
|
"text": "Elming, 2008;", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 251, |
|
"end": 270, |
|
"text": "Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 457, |
|
"end": 475, |
|
"text": "Wang et al. (2007)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 692, |
|
"end": 711, |
|
"text": "Chang et al. (2009)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this paper, we focus on improving the classification accuracy of DE constructions in Chinese as well as investigating its impact on translation quality. From the grammatical perspective, the (DE) in Chinese represents the meaning of \"noun modification\" which generally is shown in the form of a Noun phrase (NP) [A DE B] . A includes all the words in the NP before DE and B contains all the words in the NP after DE. Wang et al. (2007) first introduced a reordering of the DE construction based on a set of rules which were generated manually and achieved significant improvements in translation quality. Chang et al. (2009) extended this work by classifying DE into 5 finer-grained categories using a log-linear classifier with rich features in order to achieve higher accuracy both in reordering and in lexical choice. Their experiments showed that a higher accuracy of the DE classification improved the accuracy of reordering component, and further indirectly improved the translation quality in terms of BLEU (Papineni et al., 2002) scores.", |
|
"cite_spans": [ |
|
{ |
|
"start": 315, |
|
"end": 323, |
|
"text": "[A DE B]", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 420, |
|
"end": 438, |
|
"text": "Wang et al. (2007)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 608, |
|
"end": 627, |
|
"text": "Chang et al. (2009)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 1017, |
|
"end": 1040, |
|
"text": "(Papineni et al., 2002)", |
|
"ref_id": "BIBREF12" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "We regard the DE classification as a labeling task, and hence propose a new model to label the DE construction using a discriminative latent variable algorithm (DPLVM) (Morency et al., 2007; Sun and Tsujii, 2009) , which uses latent variables to carry additional information that may not be expressed by those original labels and capture more complicated dependencies between DE and its corresponding features. We also propose a new feature defined as \"tree-pattern\" which can automatically learn the reordering rules rather than using manually generated ones.", |
|
"cite_spans": [ |
|
{ |
|
"start": 168, |
|
"end": 190, |
|
"text": "(Morency et al., 2007;", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 191, |
|
"end": 212, |
|
"text": "Sun and Tsujii, 2009)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The remainder of this paper is organised as follows. In section 2, we introduce the types of word order errors caused by the DE construction. Section 3 describes the closely related work on DE construction. In section 4, we detail our proposed DPLVM algorithm and its adaptation to our task. We also describe the feature templates as well as the proposed new feature used in our model. In section 5, the classification experiments are conducted to compare the proposed classification model with a log-linear model. Section 6 reports comparative experiments conducted on the NIST 2008 data set using two sets of reordered and non-reordered data. Meanwhile, in section 7, an analysis on how the syntactic DE reordering affects word alignment and phrase table is given. Section 8 concludes and gives avenues for future work.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Although syntactic reordering is an effective way of significantly improving translation quality, word order is still a major error source between Chinese and English translation. Take examples in Figure 1 as an illustration. The errors of three translation results in Figure 1 are from different MT systems, and many errors relate to incorrect reordering for the (DE) structure. These three translations are from different Hiero systems. Although Hiero has an inherent reordering capability, none of them correctly re-Source:", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 197, |
|
"end": 205, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 269, |
|
"end": 277, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "The Problem of Chinese DE Construction Translation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "(local) (a) (bad reputation) (with) (middle school) Reference: 'a local middle school with a bad reputation' Team 1: 'a bad reputation of the local secondary school' Team 2: 'the local a bad reputation secondary school' Team 3: 'a local stigma secondary schools' Figure 1 : Examples of DE construction translation errors from (Chang et al., 2009) ordered \"bad reputation\" and \"middle school\" around the DE. Chang et al. (2009) suggested that this is because it is not sufficient to have a formalism which supports phrasal reordering. They claimed it is necessary to have sufficient linguistic modeling, so that the system knows when and how much to rearrange. Figure 2 gives an example illustrating how the reordering of DE construction influences the translation of a Chinese sentence. We can see that if we can properly recognise the DE construction [A DE B] and correctly perform the reordering, we can achieve a closer word order with English and hence a good English translation even it is literal.", |
|
"cite_spans": [ |
|
{ |
|
"start": 326, |
|
"end": 346, |
|
"text": "(Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 407, |
|
"end": 426, |
|
"text": "Chang et al. (2009)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 263, |
|
"end": 271, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 660, |
|
"end": 668, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "The Problem of Chinese DE Construction Translation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Although the Hiero system has a strong reordering capability in its generalised phrases, it still cannot process some complicated and flexible cases of DE construction like those in Figure 1 . Therefore, a lot of work has gone into word reordering before decoding so that the Chinese sentences have a closer word order with corresponding English sentences.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 182, |
|
"end": 190, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "The Problem of Chinese DE Construction Translation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "To address the word order problems of the DE construction, Wang et al. (2007) proposed a syntactic reordering approach to deal with structural differences and to reorder source language sentences to be much closer to the order of target language sentences. They presented a set of manually generated syntactic rules to determine whether a (DE) construction should be reordered or not before translation, such as \"For DNPs consisting of 'XP+DEG', reorder if XP is PP or LCP\" etc. (cf. (Wang et al., 2007) ). The deficiency of their algorithm is that they did not fully consider the flexibility of the DE construction, as it can be translated in many different ways. (Wang et al., 2007) and characterised the DE structures into 5 finer-grained classes based on their syntactic behaviour. They argued that one possible reason why the (DE) construction remains problematic is that previous work has paid insufficient attention to the many ways that the (DE) construction can be translated, as well as the rich structural cues which exist for these translations.", |
|
"cite_spans": [ |
|
{ |
|
"start": 59, |
|
"end": 77, |
|
"text": "Wang et al. (2007)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 484, |
|
"end": 503, |
|
"text": "(Wang et al., 2007)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 665, |
|
"end": 684, |
|
"text": "(Wang et al., 2007)", |
|
"ref_id": "BIBREF14" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "For a Chinese noun phrase [A B], it can be categorized into one of the following five classes (cf. (Chang et al., 2009) for some real examples of each class):", |
|
"cite_spans": [ |
|
{ |
|
"start": 99, |
|
"end": 119, |
|
"text": "(Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 A B (label: DE AB )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In this category, A on the Chinese side is translated as a pre-modifier of B. In most cases A is an adjectival form.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 B preposition A (label: DE BprepA ) There are several cases that are translated into the form B preposition A.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 A's B (label: DE AsB )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In this class, the English translation is an explicit s-genitive case. This class occurs much less often but is still interesting because of the difference from the of-genitive.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 relative clause (label: DE relc )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In this class, the relative clause would be introduced by a relative pronoun or be a reduced relative clause.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 A preposition B (label: DE AprepB )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "This class is another small one. The English translations that fall into this class usually have some number, percentage or level word in the Chinese A. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work on DE Construction", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Based on the discussion so far, we can see that:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "\u2022 syntactic reordering of the DE construction in Chinese is an effective way to improve the translation quality;", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "\u2022 classifying the DE construction into finergrained categories could achieve better reordering and translation performance;", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "\u2022 classification accuracy of the DE construction in Chinese has a significant impact on SMT performance.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "Driven by these three points, especially the third one, we propose a DPLVM-based classifier to improve classification accuracy. In natural language processing (NLP) such as sequential labeling (Sun and Tsujii, 2009) , DPLVM demonstrated excellent capability of learning latent dependencies of the specific problems, and have outperformed several commonly-used conventional models, such as support vector machines, conditional random fields and hidden Markov models.", |
|
"cite_spans": [ |
|
{ |
|
"start": 193, |
|
"end": 215, |
|
"text": "(Sun and Tsujii, 2009)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "In this section, we theoretically introduce the definition and mathematical description of the DPLVM algorithm used in NLP tasks (Sun and Tsujii, 2009) . Given a sequence of observations x = {x 1 , x 2 , . . . , x m } and a sequence of labels y = {y 1 , y 2 , . . . , y m }, the task is to learn a mapping between x and y. y i is a class label and is a member of a set Y of possible class labels. DPLVM also assumes a sequence of latent variables h = {h 1 , h 2 , . . . , h m }, which is hidden in the training examples.", |
|
"cite_spans": [ |
|
{ |
|
"start": 129, |
|
"end": 151, |
|
"text": "(Sun and Tsujii, 2009)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "The DPLVM is defined as in (1) (Morency et al., 2007; Sun and Tsujii, 2009) :", |
|
"cite_spans": [ |
|
{ |
|
"start": 31, |
|
"end": 53, |
|
"text": "(Morency et al., 2007;", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 54, |
|
"end": 75, |
|
"text": "Sun and Tsujii, 2009)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "P (y|x, \u0398) = h P (y|h, x, \u0398)P (h|x, \u0398)", |
|
"eq_num": "(1)" |
|
} |
|
], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "where \u0398 are the parameters of the model. It can be seen that the DPLVM equates to a CRF model if it has only one latent variable for each label. For the sake of efficiency, the model is restricted to have disjoint sets of latent variables associated with each class label. Each h j is a member in a set H y j of possible latent variables for the class label y j . We define H as the union of all H y j sets, so sequences which have any h j \u2208 H y j will by definition have P (y|x, \u0398) = 0, so that the model can be rewritten as in (2):", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "P (y|x, \u0398) = h\u2208Hy 1 \u00d7...Hy m P (h|x, \u0398)", |
|
"eq_num": "(2)" |
|
} |
|
], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "where P (h|x, \u0398) is defined by the usual conditional random field formulation, as in (3):", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "P (h|x, \u0398) = exp\u0398 \u2022 f(h, x) \u2200h exp\u0398 \u2022 f(h, x)", |
|
"eq_num": "(3)" |
|
} |
|
], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "in which f(h, x) is a feature vector. Given a training set consisting of n labeled sequences (x i , y i ), for i = 1 . . . n, parameter estimation is performed by optimizing the objective function in (4):", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "L(\u0398) = n i=1 log P (y i |x i , \u0398) \u2212 R(\u0398)", |
|
"eq_num": "(4)" |
|
} |
|
], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "The first term of this equation is the conditional log-likelihood of the training data. The second term is a regularizer that is used for reducing overfitting in parameter estimation. For decoding in the test stage, given a test sequence x, we want to find the most probable label sequence y * , as in (5):", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "y * = arg max y P (y|x, \u0398 * )", |
|
"eq_num": "(5)" |
|
} |
|
], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Sun and Tsujii 2009argued that for latent conditional models like DPLVMs, the best label path y * cannot directly be generated by the Viterbi algorithm because of the incorporation of hidden states. They proposed a latent-dynamic inference (LDI) method based on A * search and dynamic programming to efficiently decode the optimal label sequence y * . For more details of the LDI algorithm, refer to (Sun and Tsujii, 2009) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 400, |
|
"end": 422, |
|
"text": "(Sun and Tsujii, 2009)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "In our experiments, we use the open source toolkit of DPLVM 1 and adapt it to our special requirements based on the different features and scenarios.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DPLVM Algorithm", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "We use the 5 classes of DE of (Chang et al., 2009) shown in Section 3 to label DE using our DPLVM model. In order to fairly compare the classification performance between that of Chang et al. (2009) and our proposed classifiers, we use the same data sets and conditions to train and test the classifier. The data sets are the Chinese Treebank 6.0 (LDC2007T36) and the English-Chinese Translation Treebank 1.0 (LDC2007T02). For more details about the data sets, refer to (Chang et al., 2009) . There are 3523 DEs in the data set, with 543 of them in the \"other\" category which do not belong to any of the 5 pre-defined classes. In the classification experiments, the \"other\" class is excluded 2 and 2980 DEs remain, each of which is manually annotated with DE labels for the purpose of classifier training and evaluation.", |
|
"cite_spans": [ |
|
{ |
|
"start": 30, |
|
"end": 50, |
|
"text": "(Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 179, |
|
"end": 198, |
|
"text": "Chang et al. (2009)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 470, |
|
"end": 490, |
|
"text": "(Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data and DE Annotation", |
|
"sec_num": "4.3" |
|
}, |
|
{ |
|
"text": "In order to match the training and testing conditions, we used a parser trained on CTB6 excluding files 1-325 to parse the data sets with DE annotation and extract parse-related features rather than using gold-standard parses (same conditions as in (Chang et al., 2009) ). It is worth noting that in the Chinese Treebank, there are two types of POS tag for DE in NPs, namely DEC and DEG. However, as a result of using a trained parser, the POS tags of DE might have other values than DEC and DEG. In our data set, there are four other POS tags, namely {AS, DER, DEV,SP}.", |
|
"cite_spans": [ |
|
{ |
|
"start": 249, |
|
"end": 269, |
|
"text": "(Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data and DE Annotation", |
|
"sec_num": "4.3" |
|
}, |
|
{ |
|
"text": "In our task, we use the 5 class labels of DE constructions in NPs, namely", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "DE AB , DE AprepB , DE AsB , DE BprepA , DE relc .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "Note that in the case of the DE construction in Chinese, it is different from traditional sequence labeling tasks such as POS tagging, parsing etc. We only need to label one word in the NP structure, i.e. the (DE) in a Chinese NP [A DE B] . Therefore the sequence labeling task becomes efficient and speedy using the DPLVM algorithm.", |
|
"cite_spans": [ |
|
{ |
|
"start": 230, |
|
"end": 238, |
|
"text": "[A DE B]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "Based on our task, the mathematical conditions for DE classification in a sequence of [A DE B] are denoted as follows:", |
|
"cite_spans": [ |
|
{ |
|
"start": 86, |
|
"end": 94, |
|
"text": "[A DE B]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "\u2022 Sequence of Observations: ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "x = x 1 , . . . , x l , x DE , x k , . . . , x m ,", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "h = h 1 , h 2 , .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": ". . , h m , where m = 3 in our task.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "We employ five features as well in the DPLVM model, namely DEPOS, POS-gram, lexical features, SemClass as well as a new feature: treepattern, which is discussed below.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "fair comparison, we did so too. For the SMT experiments, however, we kept it.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "We did not add the sixth feature used in (Chang et al., 2009 ) -topicality -in our classifier because we do not consider it to be a very useful in a data set in which the sentences which are randomly stored. In such a corpus, the content between any adjacent sentences are irrelevant in many cases.", |
|
"cite_spans": [ |
|
{ |
|
"start": 41, |
|
"end": 60, |
|
"text": "(Chang et al., 2009", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "The new feature and the templates of all features used in our task are defined as:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labels and Features in DPLVM Model", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "As mentioned in section 4.3, there are 6 kinds of POS tags of DE. Thus, the feature template is defined as in (5):", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DEPOS:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "T depos = {dDE|dDE \u2208 DP}, where DP = {AS, DEC, DEG,DER,DEV,SP}.", |
|
"eq_num": "(5)" |
|
} |
|
], |
|
"section": "DEPOS:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Tree-pattern: Chang (2009) used an A-pattern feature which is an indicator function that fires when some syntactic rules are satisfied, such as \"A is ADJP if A+DE is a DNP with the form of 'ADJP+DEG'\", etc. These rules are induced manually based on the grammatical phenomena at hand. Here we propose a more generalised feature defined as \"tree-pattern\" to automatically learn the reordering from the training data. We consider all the sub-tree structures around DE without any word POS tags. For example, consider the parse structure (an example in (Chang et al., 2009) ", |
|
"cite_spans": [ |
|
{ |
|
"start": 14, |
|
"end": 26, |
|
"text": "Chang (2009)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 549, |
|
"end": 569, |
|
"text": "(Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DEPOS:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "where the tree-pattern is \"NP NP CP IP VP ADVP VP DEC NP\". We do not use the word POS tag (except DE) in this feature, such as NR, AD, VA, etc. The intention of this feature is to enable the classifier to automatically learn the structural rules around DE. Given that the position of DE in the parsing of [A DE B] is i, then the feature template is defined as in 7:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DEPOS:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "T tree u = {t i\u2212l , . . . , t i\u22121 , t i , t i+1 , . . . , t i+m } (7) T tree b = {t i\u2212l t i\u2212l+1 , . . . , ti\u22121ti, titi+1, . . . , ti+m\u22121ti+m}", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DEPOS:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "where T tree u is the sequence of unigrams in connection with DE and T tree b is the sequence of bigrams related to DE; l and m are the window sizes of A and B respectively. Generally, we use all the unigrams and bigrams in the parsing of A and B in our experiments. We argue that the important advantage of this feature is that it does not depend on manually generated rules, but instead of learns and generalises the reordering rules from the training data directly.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "DEPOS:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "The POS-ngram feature adds all unigrams and bigrams in A and B. Given that the position of DE is i in [A DE B] , the feature template is defined as in (8): . . . , pi+m\u22121pi+m}(8) where T pos u and T pos b are uigrams and bigrams in A and B. In the unigrams, we exclude the POS of DE; in the bigrams, we include a bigram pair across DE.", |
|
"cite_spans": [ |
|
{ |
|
"start": 102, |
|
"end": 110, |
|
"text": "[A DE B]", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 156, |
|
"end": 178, |
|
"text": ". . . , pi+m\u22121pi+m}(8)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "POS-gram:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "T pos u = {p i\u2212l , . . . , p i\u22121 , p i+1 , . . . , p i+m } T pos b = {p i\u2212l p i\u2212l+1 , . . . , p i\u22121 p i+1 ,", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "POS-gram:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Some other features such as lexical features, SemClass (cf. (Chang et al., 2009) for details) can be defined using similar feature template.", |
|
"cite_spans": [ |
|
{ |
|
"start": 60, |
|
"end": 80, |
|
"text": "(Chang et al., 2009)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "POS-gram:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In this section, we compare the performance of DE classifiers between the DPLVM and log-linear methods.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiments on DPLVM DE Classifier", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "The Table 1 shows the comparison of accuracy, where \"5-A\" and \"2-A\" represent the accuracy of the 5-class and 2-class respectively. The 2-class is the categorised classes of DE in (Wang et al., 2007) which are defined as \"reordered\" and \"nonreordered\" categories. It can be seen that our DPLVM classifier outperforms the log-linear classifier by 1.4 absolute (1.86% and 1.61% relative respectively) points both on 5-class and 2class classifications. Furthermore, we see that the DPLVM achieves significantly better performance than the log-linear model only with the simple feature of \"DEPOS\". As to the new feature \"tree-pattern\", we can see that it achieves the improvement of 1.5% compared to the \"A-pattern\" in terms of the accuracy of \"2-A\". This improvement attributes to the good learning ability of DPLVM as well as the strong generalisation capability of the tree-pattern feature.", |
|
"cite_spans": [ |
|
{ |
|
"start": 180, |
|
"end": 199, |
|
"text": "(Wang et al., 2007)", |
|
"ref_id": "BIBREF14" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 4, |
|
"end": 11, |
|
"text": "Table 1", |
|
"ref_id": "TABREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experiments on DPLVM DE Classifier", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "In terms of speed, in our task we only need to label the Chinese character DE in the NP structure [A DE B] rather than label the whole sentence, so that we have a feature matrix of n \u00d7 1 for each DE. Accordingly, the DPLVM classifier can run efficiently with low memory usage.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiments on DPLVM DE Classifier", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "6 Experiments on SMT", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiments on DPLVM DE Classifier", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "For our SMT experiments, we used two systems, namely Moses and Moseschart. The former is the state-of-the-art PB-SMT system while the latter is a new extended system of the Moses toolkit re-implementing the hierarchical PB-SMT (HPB) model (Chiang, 2005) . The alignment is carried out by GIZA++ (Och and Ney, 2003) and then we symmetrized the word alignment using the grow-diag-final heuristic. Parameter tuning is performed using Minimum Error Rate Training (Och, 2003) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 239, |
|
"end": 253, |
|
"text": "(Chiang, 2005)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 295, |
|
"end": 314, |
|
"text": "(Och and Ney, 2003)", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 459, |
|
"end": 470, |
|
"text": "(Och, 2003)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Setting", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "The training data contains 2,159,232 sentence pairs.The 5-gram language model is trained on the English part of the parallel training data. The development set (devset) is the NIST MT2006 test set and the test set is the NIST MT2008 \"current\" test set. All the results are reported in terms of BLEU (Papineni et al., 2002) and METEOR (MTR) (Banerjee and Lavie, 2005) scores.", |
|
"cite_spans": [ |
|
{ |
|
"start": 299, |
|
"end": 322, |
|
"text": "(Papineni et al., 2002)", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 340, |
|
"end": 366, |
|
"text": "(Banerjee and Lavie, 2005)", |
|
"ref_id": "BIBREF0" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Setting", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "To run the DE classifiers, we use the Stanford Chinese parser (Levy and Manning, 2003) to parse the Chinese side of the MT training data, the devset and test set.", |
|
"cite_spans": [ |
|
{ |
|
"start": 62, |
|
"end": 86, |
|
"text": "(Levy and Manning, 2003)", |
|
"ref_id": "BIBREF7" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Setting", |
|
"sec_num": "6.1" |
|
}, |
|
{ |
|
"text": "For the DE-annotated MT experiments, after we parse the training data, the devset and the test set, we separately use the two DE classifiers to annotate the DE constructions in NPs in all of the parsed data. Once the DE data are labeled, we pre-process the Chinese data by reordering the sentences only with BprepA and relc annotations. Table 2 lists the statistics of the DE classes in the MT training data, devset and test set using our DPLVM classifier. \" non \" denotes the unlabeled (DE) which does not belong to any of the 5 classes.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 337, |
|
"end": 344, |
|
"text": "Table 2", |
|
"ref_id": "TABREF8" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Statistics of 5-class DE Annotation", |
|
"sec_num": "6.2" |
|
}, |
|
{ |
|
"text": "The experimental results from the PB-SMT and HPB systems separately using the DPLVM and log-linear classifiers are shown in Table 3 . Table 3 : Experimental results on PB-SMT and Moses-chart. \"BL\" are the baselines; \"LL\" indicates the log-linear model-based system; \"LV\" is our DPLVM method.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 124, |
|
"end": 131, |
|
"text": "Table 3", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 134, |
|
"end": 141, |
|
"text": "Table 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experimental Results", |
|
"sec_num": "6.3" |
|
}, |
|
{ |
|
"text": "The baseline systems indicate that the data is neither categorised into DE classes nor reordered on the Chinese side. We can see that (1) the \"LV\" method outperformed the \"BL\" and \"LL\" by 1.44 absolute (6.42% relative), 0.39 absolute (1.66% relative) BLEU points for PB-SMT, and by 0.75 absolute (3.08% relative), 0.36 absolute (1.45% relative) BLEU points for Moses-chart;", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Results", |
|
"sec_num": "6.3" |
|
}, |
|
{ |
|
"text": "(2) the \"LV\" method achieved the improvements for PB-SMT and Moses-chart in terms of MTR scores compared to the \"BL\" and \"LL\" systems. Therefore, using DE classification and reordering on the source-side is helpful in improving translation quality; (3) the results using DPLVM achieve better translation quality than that of the \"LL\" processed data in terms of BLEU and ME-TEOR (Banerjee and Lavie, 2005 ) scores, which indirectly shows that DPLVM outperforms the log-linear classification model; and (4) the improvements on both PB-SMT and Moses-chart show that the effectiveness of DE reordering is consistent for different types of MT systems. The results are verified by significance test on 95% confidence interval (Zhang and Vogel, 2004) . 3", |
|
"cite_spans": [ |
|
{ |
|
"start": 378, |
|
"end": 403, |
|
"text": "(Banerjee and Lavie, 2005", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 720, |
|
"end": 743, |
|
"text": "(Zhang and Vogel, 2004)", |
|
"ref_id": "BIBREF16" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Results", |
|
"sec_num": "6.3" |
|
}, |
|
{ |
|
"text": "In this section, we plan to evaluate how DE reordering contributes to the improvement of translation quality in two respects, namely word alignment and phrase table.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Analysis", |
|
"sec_num": "7" |
|
}, |
|
{ |
|
"text": "We create a word alignment test set which includes 500 sentences with human alignment annotation, and then add this test set into the MT training corpus. Accordingly, the DE-reordered test set is added into the reordered training corpus as well. Thus, we run GIZA++ using the same configurations for these two sets of data and symmetrize the bidirectional word alignment using grow-diag heuristic. The word alignment of the test set is evaluated with the human annotation using Precision, Recall, F1 and AER measures. The results are reported in Table 4 . We can see that in terms of the four measures, the word alignment produced by the reordered data is slightly better than that of the original data. In some sense, we might say that the DE reordering is helpful in improving the word alignment of the training data. Table Wang et al. (2007) with the consideration that if the reordered system learned a better phrase table, then it may outperform the baseline on non-reordered inputs despite the mismatch and vice versa. However, they did not settle the question as to whether the reordered system can learn better phrase tables. We also try to use the idea of Wang et al (2007) to carry out the phrase table evaluation on PB-SMT, 4 i.e. we tune the baseline on a reordered devset and then evaluate on a reordered test set; tune the reordered system on a non-reordered devset and then evaluate on a non-reordered test set. The results are shown in Table 5 . We find that (1) given the non-reordered test set, the DE reordered system performs better than the baseline system, which is consistent when different DE classifiers are applied; (2) given the reordered test set system, the reordered set produces a better result than the baseline, which is also consistent when different DE classifiers are applied; and (3) the results from the DPLVM-based reordered data are better than those from the LLbased reordered data. From the comparison, one might say that the reordered system was learned 4 The phrases in HPB systems are different from those in PB-SMT because they are variable-based, so we evaluate the hierarchical phrases in (Du and Way, 2010) a better phrase table and the reordered test set addresses the problem of word order.", |
|
"cite_spans": [ |
|
{ |
|
"start": 833, |
|
"end": 846, |
|
"text": "et al. (2007)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1167, |
|
"end": 1184, |
|
"text": "Wang et al (2007)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 1999, |
|
"end": 2000, |
|
"text": "4", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 2139, |
|
"end": 2157, |
|
"text": "(Du and Way, 2010)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 546, |
|
"end": 553, |
|
"text": "Table 4", |
|
"ref_id": "TABREF6" |
|
}, |
|
{ |
|
"start": 820, |
|
"end": 832, |
|
"text": "Table Wang", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1454, |
|
"end": 1461, |
|
"text": "Table 5", |
|
"ref_id": "TABREF10" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Evaluating the Word Alignment", |
|
"sec_num": "7.1" |
|
}, |
|
{ |
|
"text": "To sum up, from the SMT results and the evaluation results on the word alignment and the phrase table, we can conclude that the DE reordering methods contribute significantly to the improvements in translation quality, and it also implies that using DE reordered data can achieve better word alignment and phrase tables.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Evaluating the Phrase", |
|
"sec_num": "7.2" |
|
}, |
|
{ |
|
"text": "In this paper, we presented a new classifier: a DPLVM model to classify the Chinese (DE) constructions in NPs into 5 classes. We also proposed a new and effective feature -tree-pattern -to automatically learn the reordering rules using the DPLVM algorithm. The experimental results showed that our DPLVM classifier outperformed the log-linear model in terms of both the classification accuracy and MT translation quality. In addition, the evaluation of the experimental results in section 7 indicates that the DE-reordering approach is helpful in improving the accuracy of the word alignment, and can also produce better phrase pairs and thus generate better translations.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusions and Future Work", |
|
"sec_num": "8" |
|
}, |
|
{ |
|
"text": "As for future work, firstly we plan to examine and classify the DE constructions in other syntactic structures such as VP, LCP etc. Secondly, we plan to apply the DE-annotated approach in a syntax-based MT system (Zollmann and Venugopal, 2006) and examine the effects. We also intend to improve the classification accuracy of the DE classifier with richer features to further improve translation quality.", |
|
"cite_spans": [ |
|
{ |
|
"start": 213, |
|
"end": 243, |
|
"text": "(Zollmann and Venugopal, 2006)", |
|
"ref_id": "BIBREF17" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusions and Future Work", |
|
"sec_num": "8" |
|
}, |
|
{ |
|
"text": "http://www.ibis.t.u-tokyo.ac.jp/XuSun 2 In the classification experiments ofChang et al. (2009), the \"other\" class was excluded, so in order to carry out a", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "http://projectile.sv.cmu.edu/research/public/ tools/bootStrap/tutorial.htm.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "Many thanks to Dr. Pi-Chuan Chang for providing the source code of her DE classifier and manually DE-annotated training data as well as valuable instruction in their use. Thanks also to Dr. Xu Sun for the source code of his Latent Variable classifier together with help in their use. This research is supported by the Science Foundation Ireland (Grant 07/CE/I1142) as part of the Centre for Next Generation Localisation (www.cngl.ie) at Dublin City University.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgment", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "METEOR: an automatic metric for MT evaluation with improved correlation with human judgments", |
|
"authors": [ |
|
{ |
|
"first": "Satanjeev", |
|
"middle": [], |
|
"last": "Banerjee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Alon", |
|
"middle": [], |
|
"last": "Lavie", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, ACL-2005", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "65--72", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Satanjeev Banerjee and Alon Lavie. 2005. METEOR: an automatic metric for MT evaluation with im- proved correlation with human judgments. In Pro- ceedings of the Workshop on Intrinsic and Extrin- sic Evaluation Measures for Machine Translation and/or Summarization, ACL-2005, pages 65-72.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Disambiguating \"DE\" for Chinese-English machine translation", |
|
"authors": [ |
|
{ |
|
"first": "Pi-Chuan", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dan", |
|
"middle": [], |
|
"last": "Jurafsky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Christopher", |
|
"middle": [ |
|
"D" |
|
], |
|
"last": "Manning", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the Fourth Workshop on SMT", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "215--223", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Pi-Chuan Chang, Dan Jurafsky and Christopher D. Manning. 2009 Disambiguating \"DE\" for Chinese- English machine translation. In Proceedings of the Fourth Workshop on SMT, pages 215-223.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "A hierarchical phrase-based model for statistical machine translation", |
|
"authors": [ |
|
{ |
|
"first": "David", |
|
"middle": [], |
|
"last": "Chiang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of ACL'05", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "263--270", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "David Chiang. 2005. A hierarchical phrase-based model for statistical machine translation. In Pro- ceedings of ACL'05, pages 263-270.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Clause restructuring for statistical machine translation", |
|
"authors": [ |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Collins", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Philipp", |
|
"middle": [], |
|
"last": "Koehn", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ivona", |
|
"middle": [], |
|
"last": "Kucerova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of ACL05", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "531--540", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Michael Collins, Philipp Koehn, and Ivona Kucerova. newblock 2005. Clause restructuring for statisti- cal machine translation. In Proceedings of ACL05, pages 531-540.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "The impact of source-side syntactic reordering on hierarchical phrase-based SMT", |
|
"authors": [ |
|
{ |
|
"first": "Jinhua", |
|
"middle": [], |
|
"last": "Du", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andy", |
|
"middle": [], |
|
"last": "Way", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the 14th Annual conference of the European Association for Machine Translation, Saint-Raphael", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jinhua Du and Andy Way. 2010. The impact of source-side syntactic reordering on hierarchical phrase-based SMT. In Proceedings of the 14th An- nual conference of the European Association for Machine Translation, Saint-Raphael, France.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Syntactic reordering integrated with phrase-based SMT", |
|
"authors": [ |
|
{ |
|
"first": "Jakob", |
|
"middle": [], |
|
"last": "Elming", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of ACL-08 SSST-2", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "46--54", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jakob Elming. 2008. Syntactic reordering integrated with phrase-based SMT. In Proceedings of ACL-08 SSST-2, pages 46-54.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Moses: Open Source Toolkit for Statistical Machine Translation", |
|
"authors": [ |
|
{ |
|
"first": "Philipp", |
|
"middle": [], |
|
"last": "Koehn", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hieu", |
|
"middle": [], |
|
"last": "Hoang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Birch", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Callison-Burch", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Federico", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Bertoldi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Cowan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wade", |
|
"middle": [], |
|
"last": "Shen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Moran", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Zens", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Dyer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "O", |
|
"middle": [], |
|
"last": "Bojar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Constantin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Evan", |
|
"middle": [], |
|
"last": "Herbst", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "demonstration session of ACL'07", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "177--180", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Philipp Koehn, Hieu Hoang, A. Birch, C. Callison- Burch, M. Federico, N. Bertoldi, B. Cowan, Wade Shen, C. Moran, R. Zens, C. Dyer, O. Bojar, A. Constantin and Evan Herbst. 2007. Moses: Open Source Toolkit for Statistical Machine Translation. In demonstration session of ACL'07, pages 177- 180.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Is it harder to parse Chinese, or the Chinese treebank?", |
|
"authors": [ |
|
{ |
|
"first": "Roger", |
|
"middle": [], |
|
"last": "Levy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Christopher", |
|
"middle": [], |
|
"last": "Manning", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of ACL'03", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "439--446", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Roger Levy and Christopher Manning. 2003. Is it harder to parse Chinese, or the Chinese treebank? In Proceedings of ACL'03, pages 439-446.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "A probabilistic approach to syntax-based reordering for statistical machine translation", |
|
"authors": [ |
|
{ |
|
"first": "Chi-Ho", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dongdong", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mu", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Minghui", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yi", |
|
"middle": [], |
|
"last": "Guan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "proceedings of the ACL'07", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "720--727", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chi-Ho Li, Dongdong Zhang, Mu Li, Ming Zhou, Minghui Li and Yi Guan. 2007. A probabilistic approach to syntax-based reordering for statistical machine translation. In proceedings of the ACL'07, pages 720-727.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Latent-dynamic Discriminative Models for Continuous Gesture Recognition", |
|
"authors": [ |
|
{ |
|
"first": "Louis-Philippe", |
|
"middle": [], |
|
"last": "Morency", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ariadna", |
|
"middle": [], |
|
"last": "Quattoni", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Trevor", |
|
"middle": [], |
|
"last": "Darrell", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "proceedings of CVPR'07", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--8", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Louis-Philippe Morency, Ariadna Quattoni and Trevor Darrell. 2007. Latent-dynamic Discriminative Models for Continuous Gesture Recognition. In proceedings of CVPR'07, pages 1-8.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Minimum Error Rate Training in Statistical Machine Translation", |
|
"authors": [ |
|
{ |
|
"first": "Franz", |
|
"middle": [], |
|
"last": "Och", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of ACL'03", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "160--167", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Franz Och. 2003. Minimum Error Rate Training in Statistical Machine Translation. In Proceedings of ACL'03, pages 160-167.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "A systematic comparison of various statistical alignment models", |
|
"authors": [ |
|
{ |
|
"first": "Franz", |
|
"middle": [], |
|
"last": "Och", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hermann", |
|
"middle": [], |
|
"last": "Ney", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Computational Linguistics", |
|
"volume": "29", |
|
"issue": "1", |
|
"pages": "19--51", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Franz Och and Hermann Ney. 2003. A systematic comparison of various statistical alignment models. Computational Linguistics, 29(1):19-51.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "BLEU: a Method for Automatic Evaluation of Machine Translation", |
|
"authors": [ |
|
{ |
|
"first": "Kishore", |
|
"middle": [], |
|
"last": "Papineni", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Salim", |
|
"middle": [], |
|
"last": "Roukos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Todd", |
|
"middle": [], |
|
"last": "Ward", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wei-Jing", |
|
"middle": [], |
|
"last": "Zhu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of the ACL-02", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "311--318", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kishore Papineni, Salim Roukos, Todd Ward and Wei- Jing Zhu. 2002. BLEU: a Method for Automatic Evaluation of Machine Translation. In Proceedings of the ACL-02, pages 311-318.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Sequential Labeling with Latent Variables: An Exact Inference Algorithm and An Efficient Approximation", |
|
"authors": [ |
|
{ |
|
"first": "Xu", |
|
"middle": [], |
|
"last": "Sun", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jun'ichi", |
|
"middle": [], |
|
"last": "Tsujii", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of The European Chapter of the Association for Computational Linguistics (EACL'09)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "772--780", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Xu Sun and Jun'ichi Tsujii. 2009. Sequential Label- ing with Latent Variables: An Exact Inference Algo- rithm and An Efficient Approximation. In Proceed- ings of The European Chapter of the Association for Computational Linguistics (EACL'09), pages 772- 780.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Chinese syntactic reordering for statistical machine translation", |
|
"authors": [ |
|
{ |
|
"first": "Chao", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Collins", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Philipp", |
|
"middle": [], |
|
"last": "Koehn", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of EMNLP-CoNLL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "737--745", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chao Wang, Michael Collins, and Philipp Koehn. 2007. Chinese syntactic reordering for statistical machine translation. In Proceedings of EMNLP- CoNLL, pages 737-745.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Improving a statistical MT system with automatically learned rewrite patterns", |
|
"authors": [ |
|
{ |
|
"first": "Fei", |
|
"middle": [], |
|
"last": "Xia", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Mccord", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of Coling", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "508--514", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Fei Xia and Michael McCord. 2004. Improving a statistical MT system with automatically learned rewrite patterns. In Proceedings of Coling 2004, pages 508-514.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "Measuring Confidence Intervals for the Machine Translation Evaluation Metrics", |
|
"authors": [ |
|
{ |
|
"first": "Ying", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Stephan", |
|
"middle": [], |
|
"last": "Vogel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the 10th International Conference on Theoretical and Methodological Issues in Machine Translation (TMI)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "85--94", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ying Zhang and Stephan Vogel. 2004. Measuring Confidence Intervals for the Machine Translation Evaluation Metrics. In Proceedings of the 10th In- ternational Conference on Theoretical and Method- ological Issues in Machine Translation (TMI), pages 85-94.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Syntax augmented machine translation via chart parsing", |
|
"authors": [ |
|
{ |
|
"first": "Andreas", |
|
"middle": [], |
|
"last": "Zollmann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ashish", |
|
"middle": [], |
|
"last": "Venugopal", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of HLT-NAACL 2006: Proceedings of the Workshop on Statistical Machine Translation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "138--141", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Andreas Zollmann and Ashish Venugopal. 2006. Syn- tax augmented machine translation via chart pars- ing. In Proceedings of HLT-NAACL 2006: Proceed- ings of the Workshop on Statistical Machine Trans- lation, New York, pages 138-141.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"uris": null, |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "An example of DE construction reordering (extended from the original figure in (" |
|
}, |
|
"FIGREF1": { |
|
"uris": null, |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Chang et al. (2009) used 6 kinds of features for DE classification, namely part-of-speech tag of DE (DEPOS), Chinese syntactic patterns appearing before DE (A-pattern), unigrams and bigrams of POS tags(POS-ngram), suffix unigram and bigram of word (Lexical), Semantic class of words (SemClass) and Re-occurrence of nouns (Topicality). A conditional log-linear classifier(Chang et al., 2009) is trained to classify each DE based on features extracted from the parsed data." |
|
}, |
|
"TABREF3": { |
|
"content": "<table><tr><td>: Comparison between the two classifiers</td></tr><tr><td>on 5-class and 2-class accuracy</td></tr></table>", |
|
"type_str": "table", |
|
"html": null, |
|
"text": "", |
|
"num": null |
|
}, |
|
"TABREF6": { |
|
"content": "<table><tr><td>: Comparison of Precision, Recall, F1 and</td></tr><tr><td>AER scores of evaluating word alignment on orig-</td></tr><tr><td>inal and reordered data</td></tr></table>", |
|
"type_str": "table", |
|
"html": null, |
|
"text": "", |
|
"num": null |
|
}, |
|
"TABREF7": { |
|
"content": "<table><tr><td/><td colspan=\"2\">training</td><td colspan=\"2\">devset</td><td colspan=\"2\">testset</td></tr><tr><td>DE-class</td><td>count</td><td>percent (%)</td><td>count</td><td>percent (%)</td><td>count</td><td>percent (%)</td></tr><tr><td>AB</td><td>312,679</td><td>23.08</td><td>523</td><td>25.80</td><td>453</td><td>28.78</td></tr><tr><td>AprepB</td><td>6,975</td><td>0.51</td><td>9</td><td>0.44</td><td>7</td><td>0.44</td></tr><tr><td>AsB</td><td>13,205</td><td>0.97</td><td>23</td><td>1.13</td><td>14</td><td>0.89</td></tr><tr><td>BprepA</td><td>658,589</td><td>47.31</td><td>956</td><td>48.05</td><td>688</td><td>43.71</td></tr><tr><td>relc</td><td>316,772</td><td>23.38</td><td>419</td><td>20.67</td><td>341</td><td>21.66</td></tr><tr><td>non</td><td>46,547</td><td>3.44</td><td>97</td><td>4.79</td><td>71</td><td>4.51</td></tr><tr><td>Total</td><td>1,354,767</td><td>100</td><td>2027</td><td>100</td><td>1574</td><td>100</td></tr></table>", |
|
"type_str": "table", |
|
"html": null, |
|
"text": "proposed one way to indirectly evaluate the phrase table by giving the same type of input to the baseline and reordered systems,", |
|
"num": null |
|
}, |
|
"TABREF8": { |
|
"content": "<table/>", |
|
"type_str": "table", |
|
"html": null, |
|
"text": "The number of different DE classes labeled for training data, devset and testset using the DPLVM classifier", |
|
"num": null |
|
}, |
|
"TABREF10": { |
|
"content": "<table/>", |
|
"type_str": "table", |
|
"html": null, |
|
"text": "Comparison of BLEU scores in matched and mismatched conditions on PB-SMT.", |
|
"num": null |
|
} |
|
} |
|
} |
|
} |