|
{ |
|
"paper_id": "S15-1022", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:37:09.420623Z" |
|
}, |
|
"title": "Multi-Level Alignments As An Extensible Representation Basis for Textual Entailment Algorithms", |
|
"authors": [ |
|
{ |
|
"first": "Tae-Gil", |
|
"middle": [], |
|
"last": "Noh", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Lili", |
|
"middle": [], |
|
"last": "Kotlerman", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "A major problem in research on Textual Entailment (TE) is the high implementation effort for TE systems. Recently, interoperable standards for annotation and preprocessing have been proposed. In contrast, the algorithmic level remains unstandardized, which makes component re-use in this area very difficult in practice. In this paper, we introduce multi-level alignments as a central, powerful representation for TE algorithms that encourages modular, reusable, multilingual algorithm development. We demonstrate that a pilot open-source implementation of multi-level alignment with minimal features competes with state-of-theart open-source TE engines in three languages.", |
|
"pdf_parse": { |
|
"paper_id": "S15-1022", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "A major problem in research on Textual Entailment (TE) is the high implementation effort for TE systems. Recently, interoperable standards for annotation and preprocessing have been proposed. In contrast, the algorithmic level remains unstandardized, which makes component re-use in this area very difficult in practice. In this paper, we introduce multi-level alignments as a central, powerful representation for TE algorithms that encourages modular, reusable, multilingual algorithm development. We demonstrate that a pilot open-source implementation of multi-level alignment with minimal features competes with state-of-theart open-source TE engines in three languages.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "A key challenge of Natural Language Processing is to determine what conclusions can be drawn from a natural language text, a task known as Textual Entailment (TE, Dagan and Glickman 2004) . The ability to recognize TE helps dealing with surface variability in tasks like Question Answering (Harabagiu and Hickl, 2006) , Intelligent Tutoring (Nielsen et al., 2009) , or Text Exploration (Berant et al., 2012) . Open source implementations a number of TE algorithms have become available over the last years, including BIUTEE (Stern and Dagan, 2012) and EDITS (Kouylekov and Negri, 2010) , which has made it much easier for end users to utilize TE engines.", |
|
"cite_spans": [ |
|
{ |
|
"start": 163, |
|
"end": 187, |
|
"text": "Dagan and Glickman 2004)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 290, |
|
"end": 317, |
|
"text": "(Harabagiu and Hickl, 2006)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 332, |
|
"end": 363, |
|
"text": "Tutoring (Nielsen et al., 2009)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 386, |
|
"end": 407, |
|
"text": "(Berant et al., 2012)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 524, |
|
"end": 547, |
|
"text": "(Stern and Dagan, 2012)", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 558, |
|
"end": 585, |
|
"text": "(Kouylekov and Negri, 2010)", |
|
"ref_id": "BIBREF11" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "At the same time, the situation is still more difficult for researchers and developers. Even though recently a common platform for TE has been proposed (Pad\u00f3 et al., 2015) that standardizes important aspects like annotation types, preprocessing, and knowledge resources, it largely ignores the algorithmic level. In fact, TE algorithms themselves are generally not designed to be extensible or interoperable. Therefore, changes to the algorithms -like adding support for a new language or for new analysis aspect -are often very involved, if not impossible. This often forces the next generation of TE researchers to develop and implement their own core algorithms from scratch.", |
|
"cite_spans": [ |
|
{ |
|
"start": 152, |
|
"end": 171, |
|
"text": "(Pad\u00f3 et al., 2015)", |
|
"ref_id": "BIBREF16" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this paper, we address this problem by proposing a schema for TE algorithms that revolves around a central representation layer called multi-level alignment geared towards encoding the relevant information for deciding entailment. The use of multi-level alignments encourages a modular, extensible development of TE algorithms that can be partitioned into \"alignment producers\" and \"alignment consumers\". This enables for future researchers and developers to change analysis components or add new ones in a straightforward manner.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "We also present evaluation results for a very simple TE algorithm based on multi-level alignments for English, German and Italian. It utilizes a minimal set of analyzers and four basic language-independent features. It can thus be regarded as a baseline of the performance achievable with this approach. The results can already compete with the best open-source engines available for each of the languages.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The quality of the word alignment between a Text (T) and a Hypothesis (H) has been used very early as a simple feature to decide about TE. When it was found that alignment strength can be misleading (MacCartney et al., 2006) , alignment was understood as an intermediate step whose outcome is a set of correspondences between parts of T and H that can be used to define (mis-)match features. Alignments can be established at the word level, phrase level (MacCartney et al., 2008) , or dependency level (Dinu and Wang, 2009) . Dagan et al. (2013) generalized this practical use to an architectural principle: They showed that various TE algorithms can be mapped onto a universal alignment-based schema with six steps: preprocessing, enrichment, candidate alignment generation, alignment selection, and classification.", |
|
"cite_spans": [ |
|
{ |
|
"start": 199, |
|
"end": 224, |
|
"text": "(MacCartney et al., 2006)", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 454, |
|
"end": 479, |
|
"text": "(MacCartney et al., 2008)", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 502, |
|
"end": 523, |
|
"text": "(Dinu and Wang, 2009)", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 526, |
|
"end": 545, |
|
"text": "Dagan et al. (2013)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TE with Multi-Level Alignments", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Proposal. Our proposal is similar to, but simpler than, Dagan et al.'s. Figure 1 shows the data flow.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 72, |
|
"end": 80, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "TE with Multi-Level Alignments", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "First, the text and the hypothesis are linguistically pre-processed. Then, the annotated T-H pair becomes the input for various independent aligners, which have access to knowledge resources and can compute any evidence for or against entailment that can be represented as a weighted alignment between any linguistic levels of H and T. Note that this includes many analyses not normally treated as alignment, e.g. match or mismatch in negation or modality between parts of T and H. The union of all alignments forms the central data structure, the Multi-Level Alignments.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TE with Multi-Level Alignments", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The next step is feature extraction. Features can be extracted on the basis of individual alignments, or from sets of alignments. We assume that the features form a vector describing the T-H pair, and that the last step is supervised entailment classification.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TE with Multi-Level Alignments", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Discussion. The main difference to Dagan et al.'s schema is that we intentionally leave out the step of alignment selection which explicitly selects a single alignment for each part of H or T, typically the globally most probable one. Our decision to forgo selection is grounded in our design of multi-level alignments as a repository that supports coexistence of information from different sources. This has the following benefits: (a) aligners become decoupled in that adding a new aligner does not have a direct impact on other aligners; (b) alignments produced by different aligners can have different semantics, e.g. positive (match) or negative (mismatch); (c) interactions between alignments can still be captured by defining features in the feature extraction step.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TE with Multi-Level Alignments", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "In this manner, multi-level alignments serve as an abstraction layer that encourages the development of TE algorithms composed of small, self-contained modules that solve specialized tasks in TE recognition. Each of these modules consists of two parts: an aligner, and a set of feature extractors. A priori, each module can be defined independently; to introduce interactions with other modules, it should be sufficient to extend the feature extractors.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TE with Multi-Level Alignments", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The practical benefit for the developer is that even relatively complex TE algorithms use a small set of well-defined interfaces, which makes them easy to manage, even at the implementation level. The startup cost is getting acquainted with the common data structure of multi-level alignments. We believe that developers are willing to pay this cost, especially when this provides them with a platform that supports multilingual pre-processing and resources.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TE with Multi-Level Alignments", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "We describe an implementation of a pilot TE algorithm based on the Multi-Level Alignment approach and its evaluation in three languages (EN, DE, IT). The system is available as open-source. 1", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Implementation and Evaluation", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We implement the algorithm within an open source TE development platform (Pad\u00f3 et al., 2015) . The platform provides various multilingual preprocessing pipelines and knowledge resources such as WordNet, VerbOcean, etc., under a shared API. For pre-processing, we use TreeTagger-based pipelines for all three languages.", |
|
"cite_spans": [ |
|
{ |
|
"start": 73, |
|
"end": 92, |
|
"text": "(Pad\u00f3 et al., 2015)", |
|
"ref_id": "BIBREF16" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Technical Foundations", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Another important service provided by the platform is the ability of storing a wide range of linguistic annotations in a common, language-independent data representation. The platform uses UIMA CAS (Ferrucci and Lally, 2004) as the data container, adopts the DKPro type system (de Castilho and Gurevych, 2014), and defines annotation types which can be extended in a controlled manner. We used this capability to define a multilingual Multi-Level Alignment layer with little implementation effort.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Technical Foundations", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "The pilot algorithm restricts itself to three aligners. All three are fully language-independent, even if two use language-specific knowledge resources.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "A Minimal Set of Aligners", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Lexical Aligner. The lexical aligner adds an alignment link for a pair of lemmas in T and H if it finds some kind of semantic relationships between them in a set of lexical resources. The link is directed, labeled (by the semantic relation, e.g. \"synonym\", \"antonym\") and weighted, with the weight indicating the strength of the relationship. Note that this aligner can on its own already produce alignment links with inconsistent semantics (positive and negative). For English, WordNet and VerbOcean were used as lexical resources. Italian WordNet was used for Italian, and GermaNet and German DerivBase (Zeller et al., 2013) were used as lexical resources for German.", |
|
"cite_spans": [ |
|
{ |
|
"start": 605, |
|
"end": 626, |
|
"text": "(Zeller et al., 2013)", |
|
"ref_id": "BIBREF19" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "A Minimal Set of Aligners", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "1 As a part of Excitement Open Platform for Textual Entailment. https://github.com/hltfbk/EOP-1.2.1/ wiki/AlignmentEDAP1", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "A Minimal Set of Aligners", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Paraphrase Aligner. The paraphrase aligner concentrates on surface forms rather than lemmas and can align sequences of them rather than just individual tokens. It uses paraphrase tables, e.g. extracted from parallel corpora (Bannard and Callison-Burch, 2005 ). The alignment process is similar to the lexical aligner: any two sequences of tokens in T and H are aligned if the pair is listed in the resource. The alignment links created by this aligner instantiate only one relation (\"paraphrase\") but report the strength of the relation via the translation probability. We used the paraphrase tables provided by the METEOR MT evaluation package (Denkowski and Lavie, 2014) , which are available for numerous languages.", |
|
"cite_spans": [ |
|
{ |
|
"start": 224, |
|
"end": 257, |
|
"text": "(Bannard and Callison-Burch, 2005", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 645, |
|
"end": 672, |
|
"text": "(Denkowski and Lavie, 2014)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "A Minimal Set of Aligners", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Lemma Identity Aligner. This aligner does not use any resources. It simply aligns identical lemmas between T and H and plays an important role in practice to deal with named entities.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "A Minimal Set of Aligners", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Similar to the aligners, we concentrate on a small set of four features in the pilot algorithm. Again, the features are completely language independent, even at the implementation level. This is possible because the linguistic annotations and the alignments, use a language-independent type system (cf. Section 3.1).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "A Minimal Feature Set", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "All current features measure some form of coverage on the Hypothesis, i.e. the percentage of H that can be explained by T. The underlying hypothesis is that a higher coverage of H corresponds to a higher chance of entailment. Since parts-of-speech arguably differ in the importance of being covered, we compute coverage for four sets of words separately: (a), for all words; (b), for content words; (c), for verbs; (d), for proper names (according to the POS tagger). The features are defined on the union of all produced alignments: i.e., two words count as aligned if they were aligned by any aligner. Clearly, this is an overly simplistic (albeit surprisingly effective) strategy. It can be considered a baseline for our approach that can be extended with many features that suggest themselves from the literature.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "A Minimal Feature Set", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "Evaluation 1: RTE-3. RTE-3 was the third instance of the yearly benchmarking workshops of the Textual Entailment community (Giampiccolo et (Magnini et al., 2014) . It is the only Textual Entailment dataset in multiple languages with the same content. The task is binary TE recognition, with baseline of 50% accuracy (balanced classes). We trained and tested our Multi-Level Alignment approach (MultiAlign) on the RTE-3 dataset separately for each language. We compare against the other RTE systems from the platform by Pad\u00f3 et al. (2015) , namely BIUTEE (Stern and Dagan, 2012) , EDITS (Kouylekov and Negri, 2010) , and TIE (Wang and Zhang, 2009) . Each system is configured with its best known configurations. The pilot system supports all three languages, while others support one (BIUTEE) or two languages (EDITS, TIE).", |
|
"cite_spans": [ |
|
{ |
|
"start": 123, |
|
"end": 138, |
|
"text": "(Giampiccolo et", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 139, |
|
"end": 161, |
|
"text": "(Magnini et al., 2014)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 519, |
|
"end": 537, |
|
"text": "Pad\u00f3 et al. (2015)", |
|
"ref_id": "BIBREF16" |
|
}, |
|
{ |
|
"start": 554, |
|
"end": 577, |
|
"text": "(Stern and Dagan, 2012)", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 586, |
|
"end": 613, |
|
"text": "(Kouylekov and Negri, 2010)", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 624, |
|
"end": 646, |
|
"text": "(Wang and Zhang, 2009)", |
|
"ref_id": "BIBREF18" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Evaluation", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "The results are shown in Table 1 . The pilot system performs well in all three languages. It ties with BIU-TEE on English and it outperforms TIE and EDITS in their respective results on German and Italian. This is particularly notable since all three systems have gone through several years of development, while MultiAlign is only a pilot implementation.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 25, |
|
"end": 32, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experimental Evaluation", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Evaluation 2: T-H pairs from Application Data. We perform the second evaluation on real-world application data from two application datasets: an entailment graph dataset (for English and Italian), and an e-mail categorization dataset (for German). Entailment graph building is the task of constructing graphs that hierarchically structure the statements from a collection (Berant et al., 2012) for the application of Text Exploration. In TE-based e-mail categorization, the goal is to assign the right category to an email with TE, using the email as T and a category description as H. .", |
|
"cite_spans": [ |
|
{ |
|
"start": 372, |
|
"end": 393, |
|
"text": "(Berant et al., 2012)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Evaluation", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Due to space constraints, we cannot evaluate these applications end-to-end. Instead, we focus on the Table 2 : F 1 evaluation on application data respective first step, the binary decision of entailment for individual T-H pairs. This task corresponds to RTE-3, and the main difference to Evaluation 1 is that these pairs come from real-world interactions and were produced by native speakers. All T-H pairs are sampled from application gold data which were manually constructed on the basis of anonymized customer interactions for German; Kotlerman et al. (2015) for English and Italian 2 ). The sets are fairly large (5300 pairs for English, 1700 for Italian, 1274 for German), and were sampled to be balanced. We report F 1 for comparability with non-balanced setups (our random baseline is F 1 =50). Table 2 shows our evaluation results. MultiAlign system beats EDITS for Italian (+4), and ties with TIE for German. On English, BIUTEE still outperforms MultiAlign (-2). Thus, MultiAlign also performs acceptably on real-world data.", |
|
"cite_spans": [ |
|
{ |
|
"start": 539, |
|
"end": 562, |
|
"text": "Kotlerman et al. (2015)", |
|
"ref_id": "BIBREF10" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 101, |
|
"end": 108, |
|
"text": "Table 2", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 803, |
|
"end": 810, |
|
"text": "Table 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experimental Evaluation", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "In sum, we find that MultiAlign is already competitive with state-of-the-art open-source TE engines on three languages. MultiAlign is not only much less complex, but it is also a single system covering all three languages, without any language-specific optimizations. We interpret this as a positive sign for the future of the Multi-Level Alignment approach.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Evaluation", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Visualization. The platform also supports visualization of individual Text-Hypothesis pairs, showing the alignments that were created by the system as well as the features computed on the basis of the alignments. The visualization was built on the basis of the BRAT library. 3 Figure 2 shows an example for the Text The judges made an assessment of Peter's credibility and the Hypothesis The judges assessed if Peter was credible. The top line shows the final prediction, Entailment, and the confidence (75%). The main part shows the Text and the Hypothesis below each other, connected The three features currently used by the English system are are shown below. As can be seen, they aggregate very simple statistics about the alignments: 5 of 7 tokens in the hypothesis are covered, 4 out of 5 content words, and the one proper name is also aligned. This situation motivates nicely the use of those features: a relatively low alignment coverage on all tokens is still compatible with entailment as long as the crucial tokens are aligned.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 277, |
|
"end": 285, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Experimental Evaluation", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "This visualization enables end users to quickly take in the justification behind the system's decision. Developers can inspect alignments and features for plausibility and detect possible bugs and assess the limitations of aligners and their underlying resources. For example, the current example shows a wrong link produced by the VerbOcean resource between the noun judges in the Text and the verb assessed in the Hypothesis. The reason is that the noun judges is mistaken for an inflected form of the verb to judge which indeed stands in a Stronger-than relationship to to assess.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Evaluation", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "This paper proposed the use of multi-level alignments, a rich data structure allowing multiple alignments to co-exist. We argued that multi-level alignments are a suitable basis for developing Textual Entailment algorithms by virtue of providing a beneficial abstraction layer that supports extensible and modular entailment algorithms. A pilot TE algorithm developed in this schema showed performance comparable to much more sophisticated state-of-the-art open-source TE engines and is available as open source software.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "Both datasets are publicly available. 3 http://brat.nlplab.org/index.html", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "This work was partially supported by the EC-funded project EXCITEMENT (FP7ICT-287923).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgments", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Paraphrasing with bilingual parallel corpora", |
|
"authors": [ |
|
{ |
|
"first": "Colin", |
|
"middle": [], |
|
"last": "Bannard", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chris", |
|
"middle": [], |
|
"last": "Callison-Burch", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "597--604", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Colin Bannard and Chris Callison-Burch. 2005. Para- phrasing with bilingual parallel corpora. In Proceed- ings of the 43rd Annual Meeting of the Association for Computational Linguistics, pages 597-604, Ann Arbor, MI.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Learning entailment relations by global graph structure optimization", |
|
"authors": [ |
|
{ |
|
"first": "Jonathan", |
|
"middle": [], |
|
"last": "Berant", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ido", |
|
"middle": [], |
|
"last": "Dagan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jacob", |
|
"middle": [], |
|
"last": "Goldberger", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2012, |
|
"venue": "Computational Linguistics", |
|
"volume": "38", |
|
"issue": "1", |
|
"pages": "73--111", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jonathan Berant, Ido Dagan, and Jacob Goldberger. 2012. Learning entailment relations by global graph structure optimization. Computational Linguistics, 38(1):73- 111.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Probabilistic textual entailment: Generic applied modeling of language variability", |
|
"authors": [ |
|
{ |
|
"first": "Ido", |
|
"middle": [], |
|
"last": "Dagan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Oren", |
|
"middle": [], |
|
"last": "Glickman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Sammons, and Fabio Massimo Zanzotto. 2013. Recognizing Textual Entailment: Models and Applications. Number 17 in Synthesis Lectures on Human Language Technologies", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ido Dagan and Oren Glickman. 2004. Probabilistic tex- tual entailment: Generic applied modeling of language variability. In PASCAL workshop on Learning Methods for Text Understanding and Mining, Grenoble, France. Ido Dagan, Dan Roth, Mark Sammons, and Fabio Mas- simo Zanzotto. 2013. Recognizing Textual Entailment: Models and Applications. Number 17 in Synthesis Lec- tures on Human Language Technologies. Morgan & Claypool.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "A broad-coverage collection of portable NLP components for building shareable analysis pipelines", |
|
"authors": [ |
|
{ |
|
"first": "Richard", |
|
"middle": [], |
|
"last": "Eckart De Castilho", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Iryna", |
|
"middle": [], |
|
"last": "Gurevych", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the Workshop on Open Infrastructures and Analysis Frameworks for HLT at COLING 2014", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--11", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Richard Eckart de Castilho and Iryna Gurevych. 2014. A broad-coverage collection of portable NLP compo- nents for building shareable analysis pipelines. In Pro- ceedings of the Workshop on Open Infrastructures and Analysis Frameworks for HLT at COLING 2014, pages 1-11, Dublin, Ireland, August.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Meteor universal: Language specific translation evaluation for any target language", |
|
"authors": [ |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Denkowski", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Alon", |
|
"middle": [], |
|
"last": "Lavie", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the Ninth Workshop on Statistical Machine Translation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "376--380", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Michael Denkowski and Alon Lavie. 2014. Meteor uni- versal: Language specific translation evaluation for any target language. In Proceedings of the Ninth Workshop on Statistical Machine Translation, pages 376-380, Bal- timore, MD.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Inference rules and their application to recognizing textual entailment", |
|
"authors": [ |
|
{ |
|
"first": "Georgiana", |
|
"middle": [], |
|
"last": "Dinu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rui", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 12th Conference of the European Chapter of the ACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "211--219", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Georgiana Dinu and Rui Wang. 2009. Inference rules and their application to recognizing textual entailment. In Proceedings of the 12th Conference of the European Chapter of the ACL, pages 211-219, Athens, Greece.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "An analysis of textual inference in German customer emails", |
|
"authors": [ |
|
{ |
|
"first": "Kathrin", |
|
"middle": [], |
|
"last": "Eichler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Aleksandra", |
|
"middle": [], |
|
"last": "Gabryszak", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G\u00fcnter", |
|
"middle": [], |
|
"last": "Neumann", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the Third Joint Conference on Lexical and Computational Semantics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "69--74", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kathrin Eichler, Aleksandra Gabryszak, and G\u00fcnter Neu- mann. 2014. An analysis of textual inference in Ger- man customer emails. In Proceedings of the Third Joint Conference on Lexical and Computational Semantics, pages 69-74, Dublin, Ireland.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "UIMA: An architectural approach to unstructured information processing in the corporate research environment", |
|
"authors": [ |
|
{ |
|
"first": "David", |
|
"middle": [], |
|
"last": "Ferrucci", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Adam", |
|
"middle": [], |
|
"last": "Lally", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Natural Language Engineering", |
|
"volume": "10", |
|
"issue": "3-4", |
|
"pages": "327--348", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "David Ferrucci and Adam Lally. 2004. UIMA: An ar- chitectural approach to unstructured information pro- cessing in the corporate research environment. Natural Language Engineering, 10(3-4):327-348.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "The third PASCAL recognising textual entailment challenge", |
|
"authors": [ |
|
{ |
|
"first": "Danilo", |
|
"middle": [], |
|
"last": "Giampiccolo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bernardo", |
|
"middle": [], |
|
"last": "Magnini", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ido", |
|
"middle": [], |
|
"last": "Dagan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bill", |
|
"middle": [], |
|
"last": "Dolan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--9", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Danilo Giampiccolo, Bernardo Magnini, Ido Dagan, and Bill Dolan. 2007. The third PASCAL recognising textual entailment challenge. In Proceedings of the ACL-PASCAL Workshop on Textual Entailment and Paraphrasing, pages 1-9, Prague, Czech Republic.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Methods for using textual entailment in open-domain question answering", |
|
"authors": [ |
|
{ |
|
"first": "Sanda", |
|
"middle": [], |
|
"last": "Harabagiu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andrew", |
|
"middle": [], |
|
"last": "Hickl", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of the 21st International Conference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "905--912", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sanda Harabagiu and Andrew Hickl. 2006. Methods for using textual entailment in open-domain question an- swering. In Proceedings of the 21st International Con- ference on Computational Linguistics and 44th Annual Meeting of the Association for Computational Linguis- tics, pages 905-912, Sydney, Australia.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Textual entailment graphs", |
|
"authors": [ |
|
{ |
|
"first": "Lili", |
|
"middle": [], |
|
"last": "Kotlerman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ido", |
|
"middle": [], |
|
"last": "Dagan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bernardo", |
|
"middle": [], |
|
"last": "Magnini", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luisa", |
|
"middle": [], |
|
"last": "Bentivogli", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Natural Language Engineering", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lili Kotlerman, Ido Dagan, Bernardo Magnini, and Luisa Bentivogli. 2015. Textual entailment graphs. Natural Language Engineering.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "An opensource package for recognizing textual entailment", |
|
"authors": [ |
|
{ |
|
"first": "Milen", |
|
"middle": [], |
|
"last": "Kouylekov", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Matteo", |
|
"middle": [], |
|
"last": "Negri", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the ACL 2010 System Demonstrations", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "42--47", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Milen Kouylekov and Matteo Negri. 2010. An open- source package for recognizing textual entailment. In Proceedings of the ACL 2010 System Demonstrations, pages 42-47, Uppsala, Sweden.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Learning to recognize features of valid textual entailments", |
|
"authors": [ |
|
{ |
|
"first": "Bill", |
|
"middle": [], |
|
"last": "Maccartney", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Trond", |
|
"middle": [], |
|
"last": "Grenager", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Marie-Catherine", |
|
"middle": [], |
|
"last": "De Marneffe", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Cer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Christopher D", |
|
"middle": [], |
|
"last": "Manning", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of the Human Language Technology Conference of the NAACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "41--48", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bill MacCartney, Trond Grenager, Marie-Catherine de Marneffe, Daniel Cer, and Christopher D Manning. 2006. Learning to recognize features of valid textual entailments. In Proceedings of the Human Language Technology Conference of the NAACL, pages 41-48, New York City, USA.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "A phrase-based alignment model for natural language inference", |
|
"authors": [ |
|
{ |
|
"first": "Bill", |
|
"middle": [], |
|
"last": "Maccartney", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michel", |
|
"middle": [], |
|
"last": "Galley", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Christopher D", |
|
"middle": [], |
|
"last": "Manning", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "802--811", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bill MacCartney, Michel Galley, and Christopher D Man- ning. 2008. A phrase-based alignment model for nat- ural language inference. In Proceedings of the Con- ference on Empirical Methods in Natural Language Processing, pages 802-811, Honolulu, Hawaii.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "The EX-CITEMENT open platform for textual inferences", |
|
"authors": [ |
|
{ |
|
"first": "Bernardo", |
|
"middle": [], |
|
"last": "Magnini", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Roberto", |
|
"middle": [], |
|
"last": "Zanoli", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ido", |
|
"middle": [], |
|
"last": "Dagan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kathrin", |
|
"middle": [], |
|
"last": "Eichler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G\u00fcnter", |
|
"middle": [], |
|
"last": "Neumann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tae-Gil", |
|
"middle": [], |
|
"last": "Noh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sebastian", |
|
"middle": [], |
|
"last": "Pad\u00f3", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Asher", |
|
"middle": [], |
|
"last": "Stern", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Omer", |
|
"middle": [], |
|
"last": "Levy", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the ACL 2014 System Demonstrations", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "43--48", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bernardo Magnini, Roberto Zanoli, Ido Dagan, Kathrin Eichler, G\u00fcnter Neumann, Tae-Gil Noh, Sebastian Pad\u00f3, Asher Stern, and Omer Levy. 2014. The EX- CITEMENT open platform for textual inferences. In Proceedings of the ACL 2014 System Demonstrations, pages 43-48, Baltimore, MD.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Recognizing entailment in intelligent tutoring systems", |
|
"authors": [ |
|
{ |
|
"first": "Wayne", |
|
"middle": [], |
|
"last": "Rodney D Nielsen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "James H", |
|
"middle": [], |
|
"last": "Ward", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Martin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Journal of Natural Language Engineering", |
|
"volume": "15", |
|
"issue": "4", |
|
"pages": "479--501", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rodney D Nielsen, Wayne Ward, and James H Martin. 2009. Recognizing entailment in intelligent tutoring systems. Journal of Natural Language Engineering, 15(4):479-501.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "Design and realization of a modular architecture for textual entailment", |
|
"authors": [ |
|
{ |
|
"first": "Sebastian", |
|
"middle": [], |
|
"last": "Pad\u00f3", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tae-Gil", |
|
"middle": [], |
|
"last": "Noh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Asher", |
|
"middle": [], |
|
"last": "Stern", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rui", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Roberto", |
|
"middle": [], |
|
"last": "Zanoli", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Journal of Natural Language Engineering", |
|
"volume": "21", |
|
"issue": "2", |
|
"pages": "167--200", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sebastian Pad\u00f3, Tae-Gil Noh, Asher Stern, Rui Wang, and Roberto Zanoli. 2015. Design and realization of a modular architecture for textual entailment. Journal of Natural Language Engineering, 21(2):167-200.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "BIUTEE: A modular open-source system for recognizing textual entailment", |
|
"authors": [ |
|
{ |
|
"first": "Asher", |
|
"middle": [], |
|
"last": "Stern", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ido", |
|
"middle": [], |
|
"last": "Dagan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2012, |
|
"venue": "Proceedings of the ACL 2012 System Demonstrations", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "73--78", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Asher Stern and Ido Dagan. 2012. BIUTEE: A modular open-source system for recognizing textual entailment. In Proceedings of the ACL 2012 System Demonstra- tions, pages 73-78, Jeju Island, Korea.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "Recognizing textual relatedness with predicate-argument structures", |
|
"authors": [ |
|
{ |
|
"first": "Rui", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yi", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "784--792", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rui Wang and Yi Zhang. 2009. Recognizing textual relatedness with predicate-argument structures. In Pro- ceedings of the Conference on Empirical Methods in Natural Language Processing, pages 784-792, Singa- pore.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "DE-rivBase: Inducing and evaluating a derivational morphology resource for German", |
|
"authors": [ |
|
{ |
|
"first": "Britta", |
|
"middle": [], |
|
"last": "Zeller", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jan\u0161najder", |
|
"middle": [], |
|
"last": "", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sebastian", |
|
"middle": [], |
|
"last": "Pad\u00f3", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "Proceedings of ACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1201--1211", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Britta Zeller, Jan\u0160najder, and Sebastian Pad\u00f3. 2013. DE- rivBase: Inducing and evaluating a derivational mor- phology resource for German. In Proceedings of ACL, pages 1201-1211, Sofia, Bulgaria.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"text": "Dataflow for TE algorithms based on multi level alignment", |
|
"uris": null, |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"FIGREF1": { |
|
"text": "Screenshot of the Multi-Level Alignment Visualizer by alignment links that are labeled with their source and their score. Note that the alignments can link individual words (assessment and assess are aligned through a derivational link from WordNet) but also phrases (The two occurrences of The judges in Text and Hypothesis are linked by virtue of being identical lemmas).", |
|
"uris": null, |
|
"type_str": "figure", |
|
"num": null |
|
} |
|
} |
|
} |
|
} |