|
{ |
|
"paper_id": "2021", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T07:12:05.539771Z" |
|
}, |
|
"title": "Definition Extraction from Mathematical Texts on Graph Theory in German and English", |
|
"authors": [ |
|
{ |
|
"first": "Theresa", |
|
"middle": [], |
|
"last": "Kruse", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Fritz", |
|
"middle": [], |
|
"last": "Kliche", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "We extract definitions from text books and scientific publications on mathematics in both, German and English, from the sub-domain of graph theory. Mathematical texts differ from other domains because sentences which appear as definitions from a linguistic perspective are not necessarily definitions in the mathematical sense. For the English texts we train a neural network on existing training data (Vanetik et al., 2020). For the German texts we semiautomatically generate training data using patterns for the extraction of definitions. We show that this is a feasible approach for the domain of mathematical texts which generally makes extensive use of formalized language patterns. We measure precision and recall on a random sample to evaluate our results. The F-Score is similar for both languages but precision and recall are closer to each other for the German data. Further comparisons are made with a term list automatically extracted from the data. We conclude that our approach can be used to extract candidate sentences for further postprocessing.", |
|
"pdf_parse": { |
|
"paper_id": "2021", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "We extract definitions from text books and scientific publications on mathematics in both, German and English, from the sub-domain of graph theory. Mathematical texts differ from other domains because sentences which appear as definitions from a linguistic perspective are not necessarily definitions in the mathematical sense. For the English texts we train a neural network on existing training data (Vanetik et al., 2020). For the German texts we semiautomatically generate training data using patterns for the extraction of definitions. We show that this is a feasible approach for the domain of mathematical texts which generally makes extensive use of formalized language patterns. We measure precision and recall on a random sample to evaluate our results. The F-Score is similar for both languages but precision and recall are closer to each other for the German data. Further comparisons are made with a term list automatically extracted from the data. We conclude that our approach can be used to extract candidate sentences for further postprocessing.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "In this paper, we combine two domains where definitions play a significant role: lexicography and mathematics. In lexicography, definitions provide dictionary users with information about a term. In mathematics, definitions are crucial to ensure a common understanding of the domain's concepts. We extract definitions from texts on graph theory to use them in a domain-specific dictionary. In Section 2, we give an overview of related work on types and forms of definitions and on extraction methods. In Section 3.1, we describe our data. Sections 3.2 and 3.3 present our method for the extraction of definitions from the English and the German data. Section 4.1 gives a qualitative analysis of the results and Section 4.2 a quantitative evaluation. We conclude in Section 5.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Mathematical texts consist of corollaries, lemmas, propositions, theorems, proofs and definitions (Solow, 1990) . Usually, a numbering indicates the types (e.g. Theorem 2.1) but this is not necessarily the case for definitions. Some authors include them in the numbering and others simply give them in the text. Some authors highlight defined terms, e.g. by means of italics. Kruse and Heid (2020) analyze mathematical definitions for lexicographic purposes. They conclude that analytical definitions (or logical definitions as they are also called) in the Aristotelian scheme are mostly used. Single-clause when-definitions appear to define adjectives and verbs, contrary to usual practice in dictionaries for general language, as Dziemianko and Lew (2006) proposed. For definition extraction, however, it is more relevant to distinguish a definition from a non-definition rather than paying attention to the different types of definitions. Nevertheless, some sentences are definitions from a linguistic perspective but not from a mathematical perspective in content. This constitutes a special challenge for automatic definition extraction.", |
|
"cite_spans": [ |
|
{ |
|
"start": 98, |
|
"end": 111, |
|
"text": "(Solow, 1990)", |
|
"ref_id": "BIBREF29" |
|
}, |
|
{ |
|
"start": 376, |
|
"end": 397, |
|
"text": "Kruse and Heid (2020)", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 732, |
|
"end": 757, |
|
"text": "Dziemianko and Lew (2006)", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Different kinds of sentences may be regarded as definitions in mathematics: The first kind defines a term which is used in the rest of the text. Often, one finds similar definitions for the same terms in different works of a sub-domain. Definition 1 is an example of such a definition from the sub-domain of graph theory. It defines the term semiregular as a property of bipartite graphs. We want to find such definitions in our extraction experiments. They follow the Aristotelian scheme with a definiendum (the term defined) and the definiens (the part defining, cf. e.g. Meyer, 2001, p. 283 ).", |
|
"cite_spans": [ |
|
{ |
|
"start": 574, |
|
"end": 593, |
|
"text": "Meyer, 2001, p. 283", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The second kind of definitions follows the Aristotelian scheme on a syntactic level but requires anaphora resolution to be comprehensible as it refers to something mentioned earlier or later in the text. For this kind, the context is indispensable to understand its meaning. Definition 2 constitutes such an example. These definitions are not useful for our dictionary project because they do not contain any relevant semantic information.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The third kind of definitions defines a variable which also cannot be used for the dictionary. Even if some variables often refer to the same objects (e.g. G for graph) they are not considered as terminology in our dictionary project as they rather resemble named entities. For example, Definition 3 formally defines A(G) but is only relevant in the context of the particular paragraph, i.e., a proof or a construction, and not for the conceptual sphere of the domain.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "We thus aim to find definitions following the scheme of Definition 1. However, the structure of Definition 1 is ambiguous between definition and non-definition. As an example consider Definition 4. It appears to be a definition with tree as the definiendum and the subject as the definiens. This example is taken from Saha Ray (2013) where it actually is a theorem which requires a proof as tree has been defined before (cf. Definition 5). It is not obvious that a graph meets the criteria from Definition 5 if it already has the properties from Definition 4. We do not expect our system to differentiate between these two kinds of definitions because this would require an analysis of the context beyond sentence level.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Which aspects are used in a definition and which aspects are left to be proven depends on the author's preferences for introducing a concept. The decision seems arbitrary at first sight but depends on the author's intended target group of the text (Rey, 1995; Solow, 1990; van Dormolen and Arcavi, 2000) . Some general aspects can be considered for the decision because a mathematical definition should meet certain criteria (van Dormolen and Zaslavsky, 2003) : Hierarchy, existence, equivalence and axiomatization are necessary, whereas minimality, elegance and degeneration are common but not required.", |
|
"cite_spans": [ |
|
{ |
|
"start": 248, |
|
"end": 259, |
|
"text": "(Rey, 1995;", |
|
"ref_id": "BIBREF23" |
|
}, |
|
{ |
|
"start": 260, |
|
"end": 272, |
|
"text": "Solow, 1990;", |
|
"ref_id": "BIBREF29" |
|
}, |
|
{ |
|
"start": 273, |
|
"end": 303, |
|
"text": "van Dormolen and Arcavi, 2000)", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 425, |
|
"end": 459, |
|
"text": "(van Dormolen and Zaslavsky, 2003)", |
|
"ref_id": "BIBREF31" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "A hierarchy between the defined terms is inherent to the Aristotelian scheme. Further, a definition is only meaningful if the term defined does actually exist. Equivalence refers to the above-mentioned aspect that different definitions may exist for the same concept. It has to be shown that they are actually equivalent. The criterion of axiomatization is related to hierarchy: It is possible to define more and more general hypernyms. In order to stop this chain at one point axioms are needed, usually related to set theory.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The following criteria are not mandatory: minimality requires that only necessary properties are mentioned in a definition without redundancies. Elegance is hardly an objective property but can be taken into consideration when deciding which of several possible definitions is to be taken and which is left to be proven. \"Degenerations are instances of a concept which are not expected to be included when defining the concept. They are a logical consequence from the definition. One might not want the occurrence of such instances and therefore change the definition in order to exclude them. Describing an instance as a degeneration is, of course, highly subjective and there is no objective criterion for this decision\" (van Dormolen and Zaslavsky, 2003) . These criteria combined with the individual preferences and ideas of concepts sum up to the final definitions which a mathematician writes.", |
|
"cite_spans": [ |
|
{ |
|
"start": 723, |
|
"end": 757, |
|
"text": "(van Dormolen and Zaslavsky, 2003)", |
|
"ref_id": "BIBREF31" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Mathematical definitions are usually unambiguous within a certain sub-domain. Nevertheless, homonymy may occur between different subdomains. For example, the German K\u00f6rper is translated into English as solid figure in geometry but as field in algebra. Another example is the adjective complete used as an attribute to metric spaces or graphs. The definitions differ considerably in both cases, although the same mental concept underlies both. As we work only with one mathematical sub-domain we can neglect homonymy.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Background and Related Work 2.1 Definitions in Mathematical Texts", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "1. We call a bipartite graph semiregular if it has a proper 2-colouring such that all vertices with the same colour have the same valency.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Examples of definitions", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "2. We call the above procedure branchingsearch.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Examples of definitions", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "3. Let A(G) be an incidence matrix of a connected graph G with n vertices.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Examples of definitions", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "4. A connected graph with n vertices and n \u2212 1 edges is a tree.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Examples of definitions", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "5. A tree is a connected acyclic graph.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Examples of definitions", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "6. The floor function x , also called the greatest integer function or integer value, gives the largest integer less than or equal to x.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Examples of definitions", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "7. Similarly, define the points A c , B c , B a , C a , C b so that the points A c and B c lie on the extended segment AB, the points B a and C a lie on the extended segment BC, and the point C b lies on the extended segment CA, and we have AA c = a, BB c = b, BB a = b, CC a = c and CC b = c.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Examples of definitions", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Definition extraction originally started with patternbased approaches. The patterns were then combined with a grammar analysis for e.g. apposition and anaphora resolution or syntactic features. These methods have been applied to several languages like English (Klavans and Muresan, 2001 ), German (Storrer and Wellinghoff, 2006) , Spanish (Alarc\u00f3n et al., 2009) and Dutch (Fahmi and Bouma, 2006) . Examples for such patterns in English texts are is called, is the term used to describe, is defined as, is the term for. In German, the following patterns can be indicative for definitions: bedeuten, begreifen als, bekannt als, benennen, beschreiben, bestehen aus, bezeichnen als, charakterisieren als, definieren als, gebrauchen, hei\u00dfen, nennen, sein, spezifizieren als, sprechen von, Terminus einf\u00fchren, verstehen unter, verwenden als, vorstellen als 1 . Pattern-based approaches have been used in a wide range of applications. They were among others used by Meyer et al. (1999) , Meyer (2001) , or Pearson (1998) and are still applied today (Christensen, 2019). Barbaresi et al. (2018) extract \"definitory contexts\" for words from a broad range of domains (e.g. Auseinandersetzungsbilanz or Pelletheizung) in the context of lexicography using patterns such as a X 1 is a X 2 . In line with this approach, definitions in mathematical texts can be regarded as knowledge-rich contexts which can be used in pattern-based approaches for information extraction. (Meyer, 2001; Meyer et al., 1999) . Cramer (2011, 183 ff.) analyzes linguistic features of definitions. Schumann (2014) describes (corpus-)linguistic analyses for the detection of text passages containing description (thus not explicitly definitions) of terminologically relevant concepts.", |
|
"cite_spans": [ |
|
{ |
|
"start": 260, |
|
"end": 286, |
|
"text": "(Klavans and Muresan, 2001", |
|
"ref_id": "BIBREF15" |
|
}, |
|
{ |
|
"start": 297, |
|
"end": 328, |
|
"text": "(Storrer and Wellinghoff, 2006)", |
|
"ref_id": "BIBREF30" |
|
}, |
|
{ |
|
"start": 339, |
|
"end": 361, |
|
"text": "(Alarc\u00f3n et al., 2009)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 372, |
|
"end": 395, |
|
"text": "(Fahmi and Bouma, 2006)", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 959, |
|
"end": 978, |
|
"text": "Meyer et al. (1999)", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 981, |
|
"end": 993, |
|
"text": "Meyer (2001)", |
|
"ref_id": "BIBREF18" |
|
}, |
|
{ |
|
"start": 999, |
|
"end": 1013, |
|
"text": "Pearson (1998)", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 1063, |
|
"end": 1086, |
|
"text": "Barbaresi et al. (2018)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 1457, |
|
"end": 1470, |
|
"text": "(Meyer, 2001;", |
|
"ref_id": "BIBREF18" |
|
}, |
|
{ |
|
"start": 1471, |
|
"end": 1490, |
|
"text": "Meyer et al., 1999)", |
|
"ref_id": "BIBREF19" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition extraction", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Other approaches combine pattern-based extraction and machine learning (e.g., Westerhout, 2009 ). Boella and Di Caro (2013) combine syntactic dependencies with a Support Vector Machine classifier without using patterns. Fi\u0161er et al. (2010) combine morphosyntactic patterns, automatic terminology recognition and semantic tagging with Word-Net senses for their work on Slovene Wikipedia texts.", |
|
"cite_spans": [ |
|
{ |
|
"start": 78, |
|
"end": 94, |
|
"text": "Westerhout, 2009", |
|
"ref_id": "BIBREF33" |
|
}, |
|
{ |
|
"start": 220, |
|
"end": 239, |
|
"text": "Fi\u0161er et al. (2010)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition extraction", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Today, learning algorithms and neural networks are frequently used for definition extraction. Borg et al. (2010) use genetic programming and genetic algorithms to train their system on grammatical rules. Navigli and Velardi (2010) introduce Word-Class Lattices, an approach based on word lattices generalizing over lexico-syntactic definitional patterns which outperforms traditional extraction methods. Reiplinger et al. (2012) , however, compare two methods, one based on bootstrapping lexico-syntactic patterns and the other based on deep analysis, and do not find major differences in the performances. Espinosa-Anke et al. 2015use a weakly supervised bootstrapping approach and Espinosa-Anke and Schockaert (2018) combine Convolutional and Recurrent Neural Networks for definition extraction.", |
|
"cite_spans": [ |
|
{ |
|
"start": 94, |
|
"end": 112, |
|
"text": "Borg et al. (2010)", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 204, |
|
"end": 230, |
|
"text": "Navigli and Velardi (2010)", |
|
"ref_id": "BIBREF20" |
|
}, |
|
{ |
|
"start": 404, |
|
"end": 428, |
|
"text": "Reiplinger et al. (2012)", |
|
"ref_id": "BIBREF22" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition extraction", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Del Gaudio and Branco (2009) suggest that definition extraction is language and domain independent. But Vanetik et al. (2020) show that this does not hold for definition extraction from mathematical texts. They work on a corpus crawled from Wolfram MathWorld 2 and indicate whether a certain sentence is a definition. 3 They conclude \"that mathematical definitions require special treatment, and that using cross-domain learning for detection of mathematical definitions is inefficient\".", |
|
"cite_spans": [ |
|
{ |
|
"start": 4, |
|
"end": 28, |
|
"text": "Gaudio and Branco (2009)", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 104, |
|
"end": 125, |
|
"text": "Vanetik et al. (2020)", |
|
"ref_id": "BIBREF32" |
|
}, |
|
{ |
|
"start": 318, |
|
"end": 319, |
|
"text": "3", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition extraction", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Our work is based on two comparable corpora, one in German, one in English, with texts from the mathematical sub-domain of graph theory. The German corpus contains about 700, 000 tokens with about 30, 000 types and consists of lecture notes and (parts of) nine text books. Parts of books, as opposed to the entire book, were used when only some chapters cover graph theory. The English corpus consists of eight text books and 26 scientific papers, totaling about one million tokens with about 30, 000 types.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data preprocessing", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Our goal was to create corpora of a similar size. The exact number of tokens depends on how formulas are counted. We chose material from text books and literature students at our institution work with, as students are the target group of our dictionary. The choice of texts was based on a survey we carried out with the students (Kruse and Giacomini, 2019) . Although many students indicated that they use Wikipedia for their studies we decided against including it into our corpus because we have less control on its quality from an academic perspective. Due to copyright restrictions we cannot make our corpus publicly available but in can be reproduced as we used published material.", |
|
"cite_spans": [ |
|
{ |
|
"start": 329, |
|
"end": 356, |
|
"text": "(Kruse and Giacomini, 2019)", |
|
"ref_id": "BIBREF16" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data preprocessing", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Our source files are machine-readable PDF documents, scans and plain texts. As the data is not homogeneous, we had to use different workflows to integrate them into the corpus depending on the source file. We used inftyreader 4 and Tesseract (Smith, 2007) to convert PDF documents into plain text. The mathematical formulas produced some obstacles, e.g., Tesseract had difficulties to convert fractions into plain text as it works line by line. Inftyreader is specialized in processing mathematical texts and converts formulas according to the W3C standard MathML 5 but has difficulties with low quality scans. In the latter cases we used Tesseract which ignored the formulas. Thus, some errors remain in the texts due to errors in the optical character recognition (OCR). Afterwards, we did some semi-manual post-processing to eliminate the most common errors but could not cover all of them. Thus, some errors remain as can be seen in Examples 8 and 9.", |
|
"cite_spans": [ |
|
{ |
|
"start": 242, |
|
"end": 255, |
|
"text": "(Smith, 2007)", |
|
"ref_id": "BIBREF28" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data preprocessing", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We remove Latex commands for typesetting and document layout, while commands for mathematical formulas (e.g. \\sum) are kept to preserve parts of the formulas in the input for the classifier. We split the data into sentences using the tokenizer described by Schmid (2000) . Some issues with the automatic split into sentences remain, e.g., the exclamation mark is used for calculating factorials, or sentences with the following structure appear: We can say that G is bipartite (why?) and continue the following way..., where (why?) should motivate the reader to realize the truth of the given statement. As it would cost too much effort to go through these cases manually we leave them unchanged but they should be kept in mind when discussing quantitative results such as the number of sentences in the corpus because a different tokenizer might yield different results.", |
|
"cite_spans": [ |
|
{ |
|
"start": 257, |
|
"end": 270, |
|
"text": "Schmid (2000)", |
|
"ref_id": "BIBREF26" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data preprocessing", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We use the training data compiled by Vanetik et al. (2020) for the extraction of definitions in the English corpus. The training data consists of 1, 793 sentences of which 811 are definitions. We count the sentences with domain-specific definition patterns in the training data using the following pattern indicators: abbreviate, termed, determine, definition, refer, name, the term, associate, consist, said to be, then .* is, denote, known as, given by, is a(n), define, call, is the. 72.87 % of the definition sentences and 28.21 % of the non-definition sentences contain at least one of the patterns. This legitimates our workflow to semi-automatically create the German training data by extracting sentences containing definition patterns. We further analyze the training data and find definitions in which none of the patterns appears. They often contain the verbs is/are, has/have not followed by an article and thus deviate from the standard pattern. We exclude these verbs in our set of defining verbs to avoid too many false positives.", |
|
"cite_spans": [ |
|
{ |
|
"start": 37, |
|
"end": 58, |
|
"text": "Vanetik et al. (2020)", |
|
"ref_id": "BIBREF32" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition Extraction from the English Corpus", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "As mentioned above, some non-definitions also contain the patterns. One of these cases is the verb call which in the non-definitions is frequently combined with also, as in Definition 6 where it is followed by a synonym but not an actual definition. In our lexicographic application, synonyms are dealt with separately from definitions. Similar reasons hold for the other definition patterns in the non-definitions. We give an example for the indicator define in Definition 7 which constitutes a typical example of defining a variable, as described in Section 2 (cf. Definition 3).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition Extraction from the English Corpus", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "After the pre-processing as described in Section 3.1, our corpus contains 56, 978 English sentences to be classified. We use the SimpleTransformers implementation 6 of BERT (Devlin et al., 2019) 7 with one epoch. 11, 936 (20.95 %) of the sentences were classified as definitions and 45, 042 sentences (79.05 %) as non-definitions.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition Extraction from the English Corpus", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Again, we count the sentences containing at least one definition pattern: 51.42 % of the sentences classified as definitions contain a pattern but only 15.98 % of those classified as non-definitions. Like in the training data, more sentences classified as definitions contain one of the patterns. Yet, this holds for only half of the sentences unlike the 72.87 % in the training data. The amount of sentences classified as non-definitions containing a pattern is significantly lower which might be also a consequence of noise in the data.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition Extraction from the English Corpus", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "We measure precision and recall on an exemplary random sample. 8 We manually collect 100 definitions and 100 non-definitions from our data set. To that end, we randomly sequence the sentences in the corpus and find definition sentences with help of patterns. For the non-definitions we randomly extract 200 sentences from the corpus and manually annotate if they are definitions. We take the first 100 of them for the evaluation. Thus, we have a random sample of 100 definitions and 100 non-definitions.", |
|
"cite_spans": [ |
|
{ |
|
"start": 63, |
|
"end": 64, |
|
"text": "8", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Definition Extraction from the English Corpus", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "We measure precision and recall for the labels these 200 sentences were assigned with by the BERT classifier. The results for the definition sentences are given in Table 1 . If we evaluate, in turn, the classification of non-definitions we get a precision of 0.8857 and a recall of 0.62 resulting in an F-Score of 0.7229. The higher precision of the non-definitions can probably be explained with the much higher number of non-definitions in the data compared to the number of definitions. Likewise, the high recall for the definitions can be explained by the fact that we calculate the values on a balanced random sample. We would get more realistic results if we would select 200 sentences completely random for this evaluation but in that case we run the risk of having almost no definitions in the sample which would not give reliable results.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 164, |
|
"end": 171, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Definition Extraction from the English Corpus", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "For the German corpus we create our own training data. To that end, we collect sentences containing at least one form for the following lemmas: bestehen, bezeichnen, definieren, hei\u00dfen, nennen, sagen, sprechen, verstehen. 9 We randomly extract a maximum of 100 sentences for each indicator verb and manually annotate them as definitions or non-definitions following the criteria detailed in Section 2. Additionally, we manually search the corpus for examples of definitions which do not contain an indicator verb, e.g., because they contain the verb sein (engl. be). Again, we did not include all sentences containing sein to avoid false positives. Further, non-definitions without any indicator verbs are added to the data set. In sum, we collect 799 sentences of which 256 are definitions. We use the pre-trained model bert-based-german-cased 10 from the Hugging Face library and one epoch of training. All results are summarized in Table 1 . 90.54 % of the sentences are labeled as non-definitions and 9.46 % as definitions. 47.79 % of the sentences labeled as definitions contain at least one of the patterns whereas this is the case for only 4.75 % of the sentences labeled as non-definitions which matches the expectation as this percentage is higher for definitions. For measuring precision, recall and F-Scores we evaluate again a random sample of 100 sentences for each category. We yield a similar F-Score as for the English data. But precision and recall for German are closer to each other, i.e., the precision is slightly higher and the recall slightly lower. This might be explained by the fact that the percentage of sentences labeled as definitions is lower in the German data set. However, this comparison is only valid if we expect the same percentage of definitions in both corpora.", |
|
"cite_spans": [ |
|
{ |
|
"start": 222, |
|
"end": 223, |
|
"text": "9", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 935, |
|
"end": 942, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Definition Extraction from the German Corpus", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "Both, the English and German results have lower values for precision but higher values for recall. Thus, the definitions are usually found but false positives need to be filtered. We take a closer look at the false negatives in our evaluation samples. The German sample contains nine and the English sample only 15 false negatives (cf. Examples 8 to 11). Example 8 repeats the distributive law which is not defined in this sentence. Example 9 states that two already defined terms describe the same concept. This is another example where definitions and theorems are not distinguishable. The same holds for Example 10. Four of the nine false negatives in the random sample contain the phrase we say that. We searched for this phrase in the training data and found that no example containing this phrase is included. This might be because the data was extracted from Wolfram MathWorld and not from scientific publications or textbooks. This might hint at differences in the \"language for definitions\" in different resources. Examples 12 to 14 are false positives. Example 12 contains tokens which could also occur in definitions (e.g. ist eine Zahl, Engl. is a number). Example 13 is a similar case (nennt man, Engl. is called). Example 14 is an example from the English evaluation sample. It contains the expression is defined which is also indicative for a definition. Furthermore, the English sample includes several false positives beginning with If. In the whole data set, 2, 719 sentences contain this feature; 66.50 % of them are classified as a definition. This ratio may be a result from the training data which contains 52 sentences with an initial If which are labeled in 78.85 % of the cases as definitions. 9. Damit beschreiben die Ausdr\u00fccke { Ecken)-3-panzyklisch und { Ecken)-panzyklisch den gleichen Sachverh\u00e4lt. Thus, the expressions { node)-3-pancyclic and { node)-pancyclic describe the same state of affairs.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "10. Die einzigen 3-kritischen Graphen sind Kreise ungerader L\u00e4nge.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "The only 3-critical graphs are circles of odd length.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "11. We say that a graph G is reconstructible if every reconstruction of G ' is isomorphic to G, in other words, if G can be ' reconstructed up to isomorphism from its vertex-deleted subgraphs.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "12. Jeder Buchstabe ist eine Zahl zwischen 1 und n.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "Each letter is a number between 1 and n.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "13. In diesem speziellen Fall nennt man die Menge {x,y} auch das ungeordnete Paar von x und y. In this particular case, the set {x,y} is also called the unordered pair of x and y.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "14. The matrix Miq is defined dually.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Qualitative Analysis", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "For a quantitative analysis we extract 1,000 keywords and 1,000 multiword terms from our data for each language using the corpus web tool Sketch Engine (Kilgarriff et al., 2014) which includes a function for keyword extraction. One rater evaluates in two rounds if these automatically found \"keywords\" are terminologically relevant. In the German list, 0.4705% of the keywords were relevant in this sense, and 0.5350% in the English list. These values are quite similar. Most of the false positives are variables and multiword expressions like following graph.", |
|
"cite_spans": [ |
|
{ |
|
"start": 152, |
|
"end": 177, |
|
"text": "(Kilgarriff et al., 2014)", |
|
"ref_id": "BIBREF14" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "The chosen terms are manually divided into nine semantic categories:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "\u2022 ACTIVITY: events which can be performed in graph theory, mostly verbs, e.g. connect", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "\u2022 ALGORITHM: domain-specific algorithms having a given name, e.g. Dijkstra's algorithm", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "\u2022 GENERAL: mathematical terminology which is not particularly attributed to the domain of graph theory, e.g. disjoint", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "\u2022 MAPPING: mappings in the mathematical sense, e.g. edge contraction", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "\u2022 PART: elements a graph is composed of, e.g. edges Table 2 shows the percentage of lemmas in the sentences classified as definitions. The value is higher for the English data which can be explained with the higher amount of definition sentences and the slightly lower precision indicated by the random sample. Thus, the probability to find a word in this set is generally higher. Figures 1 and 2 show which lemmas are found grouped by category. This matches our hypothesis that definitions mostly lack for the categories PERSON, PROBLEM, THEOREM and ALGORITHM. The results are much clearer for the German data which matches the results for precision and recall (cf. Table 1). We conclude that the definition extraction worked well for the majority of sentences which is reflected by the values for recall.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 52, |
|
"end": 59, |
|
"text": "Table 2", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 381, |
|
"end": 396, |
|
"text": "Figures 1 and 2", |
|
"ref_id": "FIGREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "\u2022", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Still, some aspects affect the results, e.g., we did not exclude variants in our simple search. So, there is for example a definition for 1-Faktor-Satz but not for 1-Faktorsatz; and some multiword terms appear in the lemma list as a compound but are separated in the definition.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Quantitative Analysis", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Our approach yields higher values for recall but lower values for precision. We conclude that our semi-automatic approach can be used for finding candidates for mathematical definitions but they require a subsequent manual or automatic postprocessing in order to distinguish definitions from sentences with a similar syntactic structure and vocabulary. An active learning approach in which parts of the results are evaluated in order to increase the training data iteratively could improve the approach.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion and future work", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "We get different results for the English and the German data. We see several reasons for that: The German training data was semi-automatically generated using sentences from the sample on which the trained model was subsequently applied. Therefore, the same rules for annotating definitions were used for the generation of training data and for the evaluation of the results. For our English training data provided by Vanetik et al. (2020) we only had few indications of the annotation guidelines. Furthermore, the German training data contained only half as many sentences as the English data. In combination with the fact that the training data and evaluation data stem from the same corpus, there might be some over-specification to the data set. It might be interesting to train a network on this data and to apply the model on mathematical texts from different sub-domains.", |
|
"cite_spans": [ |
|
{ |
|
"start": 418, |
|
"end": 439, |
|
"text": "Vanetik et al. (2020)", |
|
"ref_id": "BIBREF32" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion and future work", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "About 20 % of the English sentences are classified as a definition, but only about 10 % of the German sentences. A reason for this difference may be the number of sources: The German corpus comprises of only ten texts while the English corpus contains 34 texts which are shorter. A reason for the different lengths are the text types as we used more text books for German and more scientific papers for English. The number of definitions in a mathematical text also depends on its type. In general, we would expect scientific papers to contain less definitions when compared to textbooks because they can pick up prior knowledge of their readers whereas textbooks are mostly targeted at learners with less prior domain knowledge. However, our results do not confirm this hypothesis as there are more sentences classified as definitions in German Data English Data number of definitions 3,417 11,936 number of lemmas 1,070 933 percentage of lemmas found in definitions 70.63% 90.47% the English data than in the German. This might be related to the lower precision for the experiments on English texts. It would be interesting to investigate empirically if the percentage of definitions varies across different mathematical text types. Furthermore, for our English corpus we had to rely more on OCR than for the German data. This may result in more mistakes which cause difficulties for the classifier. Interesting further research would be to analyze if the English extraction results differ when the training data is taken from the same corpus or from a corpus of the same sub-domain or type of resource. Maybe the results of Vanetik et al. (2020) can be interpreted in the more general way that the quality of definition extraction increases with the similarity between training data and evaluation data even for a highly formalized language like mathematics.", |
|
"cite_spans": [ |
|
{ |
|
"start": 1634, |
|
"end": 1655, |
|
"text": "Vanetik et al. (2020)", |
|
"ref_id": "BIBREF32" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 839, |
|
"end": 981, |
|
"text": "German Data English Data number of definitions 3,417 11,936 number of lemmas 1,070 933 percentage of lemmas found in definitions 70.63%", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Conclusion and future work", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "We can conclude that patterns are good indicators for mathematical definitions in German and English and can be used to generate training data. Nevertheless, automatic solutions are still needed for definition extraction in mathematics as some sentences are definitions from a linguistic perspective but not intended as such by their author.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion and future work", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "Engl. mean, understand as, known as, designate, describe, consist of, refer to as, characterize as, define as, use, be called, state, be, specify as, speak of, introduce a term, understand by, use as, conceive as", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "https://mathworld.wolfram.com 3 The data ofVanetik et al. (2020) is publicly available on GitHub https://github.com/uplink007/ FinalProject/tree/master/data/wolfram", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "https://www.inftyreader.org 5 https://www.w3.org/TR/MathML3/", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "https://github.com/ThilinaRajapakse/ simpletransformers 7 https://huggingface.co/ bert-base-cased", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We thank the anonymous reviewers for their useful comments on the evaluation and discussion sections.9 Engl. consist, denote, define, call, name, (to) name, say, speak, understand", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "https://huggingface.co/ bert-base-german-cased", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Description and evaluation of a pattern based approach for definition extraction", |
|
"authors": [ |
|
{ |
|
"first": "Rodrigo", |
|
"middle": [], |
|
"last": "Alarc\u00f3n", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Gerardo", |
|
"middle": [], |
|
"last": "Sierra", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Carme", |
|
"middle": [], |
|
"last": "Bach", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 1st Workshop on Definition Extraction", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "7--13", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rodrigo Alarc\u00f3n, Gerardo Sierra, and Carme Bach. 2009. Description and evaluation of a pattern based approach for definition extraction. In Proceedings of the 1st Workshop on Definition Extraction, pages 7-13, Borovets, Bulgaria. Association for Computa- tional Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "A database of German definitory contextsfrom selected web sources", |
|
"authors": [ |
|
{ |
|
"first": "Adrien", |
|
"middle": [], |
|
"last": "Barbaresi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lothar", |
|
"middle": [], |
|
"last": "Lemnitzer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Alexander", |
|
"middle": [], |
|
"last": "Geyken", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC'10)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Adrien Barbaresi, Lothar Lemnitzer, and Alexander Geyken. 2018. A database of German definitory contextsfrom selected web sources. In Proceed- ings of the Eleventh International Conference on Language Resources and Evaluation (LREC'10), Miyazaki, Japan.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Extracting definitions and hypernym relations relying on syntactic dependencies and support vector machines", |
|
"authors": [ |
|
{ |
|
"first": "Guido", |
|
"middle": [], |
|
"last": "Boella", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luigi", |
|
"middle": [ |
|
"Di" |
|
], |
|
"last": "Caro", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "532--537", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Guido Boella and Luigi Di Caro. 2013. Extracting def- initions and hypernym relations relying on syntactic dependencies and support vector machines. In Pro- ceedings of the 51st Annual Meeting of the Associa- tion for Computational Linguistics (Volume 2: Short Papers), pages 532-537, Sofia, Bulgaria.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Automatic grammar rule extraction and ranking for definitions", |
|
"authors": [ |
|
{ |
|
"first": "Claudia", |
|
"middle": [], |
|
"last": "Borg", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mike", |
|
"middle": [], |
|
"last": "Rosner", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Gordon", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Pace", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Claudia Borg, Mike Rosner, and Gordon J. Pace. 2010. Automatic grammar rule extraction and ranking for definitions. In Proceedings of the Seventh Interna- tional Conference on Language Resources and Eval- uation (LREC'10), Valletta, Malta.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Danish knowledge patterns and word sketches for semi-automatic extraction of terminological information", |
|
"authors": [ |
|
{ |
|
"first": "Lotte Weilgaard", |
|
"middle": [], |
|
"last": "Christensen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "New challenges for Research on Language for Special Purposes. Selected Proceedings from the 21st LSP-conference 28-30", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "173--189", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lotte Weilgaard Christensen. 2019. Danish knowledge patterns and word sketches for semi-automatic ex- traction of terminological information. In Ingrid Si- monnaes, \u00d8ivin Andersen, and Klaus Schubert, ed- itors, New challenges for Research on Language for Special Purposes. Selected Proceedings from the 21st LSP-conference 28-30 June 2017 Bergen, Nor- way, pages 173-189. Frank & Timme, Berlin.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Definitionen in W\u00f6rterbuch und Text: Zur manuellen Annotation, korpusgest\u00fctzten Analyse und automatischen Extraktion definitorischer Textsegmente im Kontext der computergest\u00fctzten Lexikographie. Dissertation, Technische Universit\u00e4t Dortmund", |
|
"authors": [ |
|
{ |
|
"first": "Irene", |
|
"middle": [], |
|
"last": "Cramer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Irene Cramer. 2011. Definitionen in W\u00f6rterbuch und Text: Zur manuellen Annotation, korpus- gest\u00fctzten Analyse und automatischen Extraktion definitorischer Textsegmente im Kontext der comput- ergest\u00fctzten Lexikographie. Dissertation, Technis- che Universit\u00e4t Dortmund, Dortmund.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Language independent system for definition extraction: First results using learning algorithms", |
|
"authors": [ |
|
{ |
|
"first": "Rosa", |
|
"middle": [], |
|
"last": "Del Gaudio", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ant\u00f3nio", |
|
"middle": [], |
|
"last": "Branco", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 1st Workshop on Definition Extraction", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "33--39", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rosa Del Gaudio and Ant\u00f3nio Branco. 2009. Lan- guage independent system for definition extraction: First results using learning algorithms. In Proceed- ings of the 1st Workshop on Definition Extraction, pages 33-39, Borovets, Bulgaria. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
|
"authors": [ |
|
{ |
|
"first": "Jacob", |
|
"middle": [], |
|
"last": "Devlin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming-Wei", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kenton", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kristina", |
|
"middle": [], |
|
"last": "Toutanova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "4171--4186", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/N19-1423" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171-4186, Minneapolis, Minnesota. Associ- ation for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "What is a circle?", |
|
"authors": [ |
|
{ |
|
"first": "Abraham", |
|
"middle": [], |
|
"last": "Joop Van Dormolen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Arcavi", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Mathematics in School", |
|
"volume": "29", |
|
"issue": "5", |
|
"pages": "15--19", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Joop van Dormolen and Abraham Arcavi. 2000. What is a circle? Mathematics in School, 29(5):15-19.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "When you are explaining the meaning of a word: The effect of abstract noun definition format on syntactic class identification", |
|
"authors": [ |
|
{ |
|
"first": "Anna", |
|
"middle": [], |
|
"last": "Dziemianko", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Robert", |
|
"middle": [], |
|
"last": "Lew", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of the 12th EURALEX International Congress", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "857--863", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Anna Dziemianko and Robert Lew. 2006. When you are explaining the meaning of a word: The ef- fect of abstract noun definition format on syntac- tic class identification. In Proceedings of the 12th EURALEX International Congress, pages 857-863, Torino, Italy. Edizioni dell'Orso.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Weakly supervised definition extraction", |
|
"authors": [ |
|
{ |
|
"first": "Luis", |
|
"middle": [], |
|
"last": "Espinosa-Anke", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Horacio", |
|
"middle": [], |
|
"last": "Saggion", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Francesco", |
|
"middle": [], |
|
"last": "Ronzano", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the International Conference Recent Advances in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "176--185", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Luis Espinosa-Anke, Horacio Saggion, and Francesco Ronzano. 2015. Weakly supervised definition ex- traction. In Proceedings of the International Con- ference Recent Advances in Natural Language Pro- cessing, pages 176-185, Hissar, Bulgaria. INCOMA Ltd. Shoumen.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Syntactically aware neural architectures for definition extraction", |
|
"authors": [ |
|
{ |
|
"first": "Luis", |
|
"middle": [], |
|
"last": "Espinosa-Anke", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Steven", |
|
"middle": [], |
|
"last": "Schockaert", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "378--385", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/N18-2061" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Luis Espinosa-Anke and Steven Schockaert. 2018. Syntactically aware neural architectures for defini- tion extraction. In Proceedings of the 2018 Confer- ence of the North American Chapter of the Associ- ation for Computational Linguistics: Human Lan- guage Technologies, Volume 2 (Short Papers), pages 378-385, New Orleans, Louisiana.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Learning to identify definitions using syntactic features", |
|
"authors": [ |
|
{ |
|
"first": "Ismail", |
|
"middle": [], |
|
"last": "Fahmi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Gosse", |
|
"middle": [], |
|
"last": "Bouma", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of the Workshop on Learning Structured Information in Natural Language Applications", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "64--71", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ismail Fahmi and Gosse Bouma. 2006. Learning to identify definitions using syntactic features. In Pro- ceedings of the Workshop on Learning Structured In- formation in Natural Language Applications, pages 64-71.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Learning to mine definitions from slovene structured and unstructured knowledge-rich resources", |
|
"authors": [ |
|
{ |
|
"first": "Darja", |
|
"middle": [], |
|
"last": "Fi\u0161er", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Senja", |
|
"middle": [], |
|
"last": "Pollak", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Vintar", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Darja Fi\u0161er, Senja Pollak, and\u0160pela Vintar. 2010. Learning to mine definitions from slovene structured and unstructured knowledge-rich resources. In Pro- ceedings of the Seventh International Conference on Language Resources and Evaluation (LREC'10), Valletta, Malta.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "The sketch engine: ten years on. Lexicography", |
|
"authors": [ |
|
{ |
|
"first": "Adam", |
|
"middle": [], |
|
"last": "Kilgarriff", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "V\u00ect", |
|
"middle": [], |
|
"last": "Baisa", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jan", |
|
"middle": [], |
|
"last": "Bu\u0161ta", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Milo\u0161", |
|
"middle": [], |
|
"last": "Jakub\u00ed\u010dek", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Vojt\u011bch", |
|
"middle": [], |
|
"last": "Kov\u00e1\u0159", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "7--36", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1007/s40607-014-0009-9" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Adam Kilgarriff, V\u00ect Baisa, Jan Bu\u0161ta, Milo\u0161 Jakub\u00ed\u010dek, Vojt\u011bch Kov\u00e1\u0159, Jan Michelfeit, Pavel Rychl\u00fd, and V\u00ect Suchomel. 2014. The sketch engine: ten years on. Lexicography, pages 7-36.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Evaluation of the definder system for fully automatic glossary construction", |
|
"authors": [ |
|
{ |
|
"first": "Judith", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Klavans", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Smaranda", |
|
"middle": [], |
|
"last": "Muresan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Proceedings AMIA Symposium", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "324--328", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Judith L. Klavans and Smaranda Muresan. 2001. Eval- uation of the definder system for fully automatic glossary construction. In Proceedings AMIA Sym- posium, pages 324-328.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "Planning a domain-specific electronic dictionary for the mathematical field of graph theory: definitional patterns and term variation", |
|
"authors": [ |
|
{ |
|
"first": "Theresa", |
|
"middle": [], |
|
"last": "Kruse", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Laura", |
|
"middle": [], |
|
"last": "Giacomini", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Electronic lexicography in the 21st century. Proceedings of the eLex 2019 conference", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "676--693", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Theresa Kruse and Laura Giacomini. 2019. Planning a domain-specific electronic dictionary for the math- ematical field of graph theory: definitional patterns and term variation. In Electronic lexicography in the 21st century. Proceedings of the eLex 2019 confer- ence, pages 676-693, Brno. Lexical Computing CZ, s.r.o.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Lemma selection and microstructure: Definitions and semantic relations of a domain-specific e-dictionary of the mathematical domain of graph theory", |
|
"authors": [ |
|
{ |
|
"first": "Theresa", |
|
"middle": [], |
|
"last": "Kruse", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ulrich", |
|
"middle": [], |
|
"last": "Heid", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Euralex Proceedings", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "227--233", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Theresa Kruse and Ulrich Heid. 2020. Lemma selec- tion and microstructure: Definitions and semantic re- lations of a domain-specific e-dictionary of the math- ematical domain of graph theory. In Euralex Pro- ceedings, volume 1, pages 227-233.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "Extracting knowledge-rich contexts for terminography: A conceptual and methodological framework", |
|
"authors": [ |
|
{ |
|
"first": "Ingrid", |
|
"middle": [], |
|
"last": "Meyer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "Didier Bourigault, Christian Jacquemin, and Marie-Claude L'Homme", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "279--302", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ingrid Meyer. 2001. Extracting knowledge-rich con- texts for terminography: A conceptual and method- ological framework. In Didier Bourigault, Chris- tian Jacquemin, and Marie-Claude L'Homme, ed- itors, Recent Advances in Computational Termi- nology, volume 2 of Natural Language Process- ing, pages 279-302. John Benjamins, Amster- dam/Philadelphia.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Conceptual sampling for terminographical corpus analysis", |
|
"authors": [ |
|
{ |
|
"first": "Ingrid", |
|
"middle": [], |
|
"last": "Meyer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kristen", |
|
"middle": [], |
|
"last": "Mackintosh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Caroline", |
|
"middle": [], |
|
"last": "Barri\u00e8re", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tricia", |
|
"middle": [], |
|
"last": "Morgan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "Terminology and Knowledge engineering", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "256--267", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ingrid Meyer, Kristen Mackintosh, Caroline Barri\u00e8re, and Tricia Morgan. 1999. Conceptual sampling for terminographical corpus analysis. In Terminol- ogy and Knowledge engineering, pages 256-267, Vi- enna. TermNet.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Learning word-class lattices for definition and hypernym extraction", |
|
"authors": [ |
|
{ |
|
"first": "Roberto", |
|
"middle": [], |
|
"last": "Navigli", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paola", |
|
"middle": [], |
|
"last": "Velardi", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1318--1327", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Roberto Navigli and Paola Velardi. 2010. Learning word-class lattices for definition and hypernym ex- traction. In Proceedings of the 48th Annual Meet- ing of the Association for Computational Linguistics, pages 1318-1327, Uppsala, Sweden.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Terms in Context, volume 1 of Studies in Corpus Linguistics", |
|
"authors": [ |
|
{ |
|
"first": "Jennifer", |
|
"middle": [], |
|
"last": "Pearson", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "John Benjamins", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jennifer Pearson. 1998. Terms in Context, volume 1 of Studies in Corpus Linguistics. John Benjamins, Amsterdam / Philadelphia.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Extracting glossary sentences from scholarly articles: A comparative evaluation of pattern bootstrapping and deep analysis", |
|
"authors": [ |
|
{ |
|
"first": "Melanie", |
|
"middle": [], |
|
"last": "Reiplinger", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ulrich", |
|
"middle": [], |
|
"last": "Sch\u00e4fer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Magdalena", |
|
"middle": [], |
|
"last": "Wolska", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2012, |
|
"venue": "Proceedings of the ACL-2012 Special Workshop on Rediscovering 50 Years of Discoveries", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "55--65", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Melanie Reiplinger, Ulrich Sch\u00e4fer, and Magdalena Wolska. 2012. Extracting glossary sentences from scholarly articles: A comparative evaluation of pat- tern bootstrapping and deep analysis. In Proceed- ings of the ACL-2012 Special Workshop on Redis- covering 50 Years of Discoveries, pages 55-65, Jeju Island, Korea.", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "Essays on Terminology", |
|
"authors": [ |
|
{ |
|
"first": "Alain", |
|
"middle": [], |
|
"last": "Rey", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1995, |
|
"venue": "Benjamins Translation Library. John Benjamins", |
|
"volume": "9", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Alain Rey. 1995. Essays on Terminology, volume 9 of Benjamins Translation Library. John Benjamins, Amsterdam.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Graph Theory with Algorithms and its Applications", |
|
"authors": [ |
|
{ |
|
"first": "Ray", |
|
"middle": [], |
|
"last": "Santanu Saha", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "In Applied Science and Technology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Santanu Saha Ray. 2013. Graph Theory with Al- gorithms and its Applications. In Applied Sci- ence and Technology. Springer India, New", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Unsupervised learning of period disambiguation for tokenisation", |
|
"authors": [ |
|
{ |
|
"first": "Helmut", |
|
"middle": [], |
|
"last": "Schmid", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Helmut Schmid. 2000. Unsupervised learning of pe- riod disambiguation for tokenisation. Technical re- port, IMS, University of Stuttgart.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Linguistische Analyse und korpusbasierte Extraktion deutscher und russischer wissenshaltiger Kontexte. Dissertation", |
|
"authors": [ |
|
{ |
|
"first": "Anne-Kathrin", |
|
"middle": [], |
|
"last": "Schumann", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Anne-Kathrin Schumann. 2014. Linguistische Analyse und korpusbasierte Extraktion deutscher und russis- cher wissenshaltiger Kontexte. Dissertation, Univer- sit\u00e4t Wien, Wien.", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "An overview of the tesseract ocr engine", |
|
"authors": [ |
|
{ |
|
"first": "Ray", |
|
"middle": [], |
|
"last": "Smith", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Ninth International Conference on Document Analysis and Recognition", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "629--633", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1109/ICDAR.2007.4376991" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ray Smith. 2007. An overview of the tesseract ocr engine. In Ninth International Conference on Doc- ument Analysis and Recognition, pages 629-633, Parana, Brasilien.", |
|
"links": null |
|
}, |
|
"BIBREF29": { |
|
"ref_id": "b29", |
|
"title": "How to read and do proofs. An introduction to mathematical thought processes", |
|
"authors": [ |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Solow", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1990, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Daniel Solow. 1990. How to read and do proofs. An introduction to mathematical thought processes, sec- ond edition edition. John Wiley & sons, New York.", |
|
"links": null |
|
}, |
|
"BIBREF30": { |
|
"ref_id": "b30", |
|
"title": "Automated detection and annotation of term definitions in german text corpora", |
|
"authors": [ |
|
{ |
|
"first": "Angelika", |
|
"middle": [], |
|
"last": "Storrer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sandra", |
|
"middle": [], |
|
"last": "Wellinghoff", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of The 5th Language Resources and Evaluation Conference", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "2373--2376", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Angelika Storrer and Sandra Wellinghoff. 2006. Auto- mated detection and annotation of term definitions in german text corpora. In Proceedings of The 5th Lan- guage Resources and Evaluation Conference, pages 2373-2376.", |
|
"links": null |
|
}, |
|
"BIBREF31": { |
|
"ref_id": "b31", |
|
"title": "The many facets of a definition: The case of periodicity", |
|
"authors": [ |
|
{ |
|
"first": "Orit", |
|
"middle": [], |
|
"last": "Joop Van Dormolen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Zaslavsky", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "The Journal of Mathematical Behavior", |
|
"volume": "22", |
|
"issue": "1", |
|
"pages": "91--106", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1016/S0732-3123(03)00006-3" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Joop van Dormolen and Orit Zaslavsky. 2003. The many facets of a definition: The case of periodicity. The Journal of Mathematical Behavior, 22(1):91 - 106.", |
|
"links": null |
|
}, |
|
"BIBREF32": { |
|
"ref_id": "b32", |
|
"title": "Automated discovery of mathematical definitions in text", |
|
"authors": [ |
|
{ |
|
"first": "Natalia", |
|
"middle": [], |
|
"last": "Vanetik", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Marina", |
|
"middle": [], |
|
"last": "Litvak", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sergey", |
|
"middle": [], |
|
"last": "Shevchuk", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lior", |
|
"middle": [], |
|
"last": "Reznik", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the 12th Language Resources and Evaluation Conference", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "2086--2094", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Natalia Vanetik, Marina Litvak, Sergey Shevchuk, and Lior Reznik. 2020. Automated discovery of math- ematical definitions in text. In Proceedings of the 12th Language Resources and Evaluation Confer- ence, pages 2086-2094, Marseille, France. Euro- pean Language Resources Association.", |
|
"links": null |
|
}, |
|
"BIBREF33": { |
|
"ref_id": "b33", |
|
"title": "Definition extraction using linguistic and structural features", |
|
"authors": [ |
|
{ |
|
"first": "Eline", |
|
"middle": [], |
|
"last": "Westerhout", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the 1st Workshop on Definition Extraction", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "61--67", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Eline Westerhout. 2009. Definition extraction using lin- guistic and structural features. In Proceedings of the 1st Workshop on Definition Extraction, pages 61-67, Borovets, Bulgaria. Association for Computational Linguistics.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"num": null, |
|
"type_str": "figure", |
|
"text": "Es gilt das Distributivgesetz: a-(b +c ) = ( a-b ) + ( a-c ) f\u00fcr alle a, b, c eK. The distributive law holds: a-(b +c ) = ( a-b ) + ( a-c ) for all a, b, c eK.", |
|
"uris": null |
|
}, |
|
"FIGREF1": { |
|
"num": null, |
|
"type_str": "figure", |
|
"text": "Distribution of English lemmas in definitions over categories", |
|
"uris": null |
|
}, |
|
"TABREF0": { |
|
"num": null, |
|
"text": "(20.95 %) ...classified as non-definition 32, 686 (90.54 %) 45, 042 (79.05 %)", |
|
"type_str": "table", |
|
"content": "<table><tr><td/><td/><td>German</td><td>English</td></tr><tr><td/><td>...of sentences</td><td>36, 103</td><td>56, 978</td></tr><tr><td colspan=\"4\">Number... 11, 936 Patterns in sentences... ...classified as definition 3, 417 (9.46 %) ...classified as definitions 47.79 % 51.42 % ...classified as non-definitions 4.75 % 15.98 %</td></tr><tr><td/><td>precision</td><td>0.7522</td><td>0.7054</td></tr><tr><td>Evaluation of random sample</td><td>recall</td><td>0.8500</td><td>0.9100</td></tr><tr><td/><td>F-Score</td><td>0.7981</td><td>0.7948</td></tr><tr><td/><td colspan=\"2\">Table 1: Overview of extraction results</td><td/></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF2": { |
|
"num": null, |
|
"text": "", |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"html": null |
|
} |
|
} |
|
} |
|
} |