|
{ |
|
"paper_id": "2020", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T10:23:31.950982Z" |
|
}, |
|
"title": "Mitigating Silence in Compliance Terminology during Parsing of Utterances", |
|
"authors": [ |
|
{ |
|
"first": "Esme", |
|
"middle": [], |
|
"last": "Manandise", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Intuit Futures", |
|
"location": { |
|
"addrLine": "2600 Marine Way Mountain View", |
|
"postCode": "94043", |
|
"region": "CA", |
|
"country": "USA esme" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Conrad", |
|
"middle": [], |
|
"last": "De Peuter", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "This paper reports on an approach to increase multi-token-term recall in a parsing task. We use a compliance-domain parser to extract, during the process of parsing raw text, terms that are unlisted in the terminology. The parser uses a similarity measure (Generalized Dice Coefficient) between listed terms and unlisted term candidates to (i) determine term status, (ii) serve putative terms to the parser, (iii) decrease parsing complexity by glomming multi-tokens as lexical singletons, and (iv) automatically augment the terminology. We illustrate a small experiment with examples from the tax-and-regulations domain. Bootstrapping the parsing process to detect outof-vocabulary terms at runtime increases parsing accuracy in addition to producing other benefits to a natural-language-processing pipeline, which translates arithmetic calculations written in English into computer-executable operations.", |
|
"pdf_parse": { |
|
"paper_id": "2020", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "This paper reports on an approach to increase multi-token-term recall in a parsing task. We use a compliance-domain parser to extract, during the process of parsing raw text, terms that are unlisted in the terminology. The parser uses a similarity measure (Generalized Dice Coefficient) between listed terms and unlisted term candidates to (i) determine term status, (ii) serve putative terms to the parser, (iii) decrease parsing complexity by glomming multi-tokens as lexical singletons, and (iv) automatically augment the terminology. We illustrate a small experiment with examples from the tax-and-regulations domain. Bootstrapping the parsing process to detect outof-vocabulary terms at runtime increases parsing accuracy in addition to producing other benefits to a natural-language-processing pipeline, which translates arithmetic calculations written in English into computer-executable operations.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "The task of extracting multi-token terms 1 , i.e. terminological units which denote concepts and entities in a domain, is a core task of Natural Language Processing (NLP). Within the tax-and-regulations domain, some terms are compositional (Nunberg et al., 1994; Baldwin, 2006; Krcmar et al., 2013; Boguraev et al., 2015) 2 in meaning and/or in form, such as unmarried college student or estimated tax payment; others are mixed instances of compositionality such as taxable sick leave pay or cannabis duty payable. In addition, terms can correspond either to the canonical form of the concept or to variant forms of concepts' names as in spouse or common-law partner credit versus spouse's or common-law partner's credit, or spouse amount versus spousal amount (Park et al., 2002) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 240, |
|
"end": 262, |
|
"text": "(Nunberg et al., 1994;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 263, |
|
"end": 277, |
|
"text": "Baldwin, 2006;", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 278, |
|
"end": 298, |
|
"text": "Krcmar et al., 2013;", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 299, |
|
"end": 321, |
|
"text": "Boguraev et al., 2015)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 761, |
|
"end": 780, |
|
"text": "(Park et al., 2002)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "From the perspective of parsing raw text, having multi-token terms not only simplifies the input by grouping multi-tokens as singletons but also removes syntactic complexity as the internal structure of these expressions remains opaque to parsing (Korkontzelos et al., 2010; Wehrli, 2014; Boguraev et al., 2015; Nerima et al., 2017) . In addition, having multi-token terms allows parsers to output structurallysimilar parses for sentences that are constituent-wise similar even if the intra-phrasal complexity of multi-token terms vary.", |
|
"cite_spans": [ |
|
{ |
|
"start": 247, |
|
"end": 274, |
|
"text": "(Korkontzelos et al., 2010;", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 275, |
|
"end": 288, |
|
"text": "Wehrli, 2014;", |
|
"ref_id": "BIBREF15" |
|
}, |
|
{ |
|
"start": 289, |
|
"end": 311, |
|
"text": "Boguraev et al., 2015;", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 312, |
|
"end": 332, |
|
"text": "Nerima et al., 2017)", |
|
"ref_id": "BIBREF11" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "One major issue in term extraction, known as silence, is the failure to extract terms that appear infrequently in a domain-corpus but that domain-specialists would include in a term lexicon. An example of silence is when terms such as inventory valuation or combination money purchase do not make it into the terminology because they occur infrequently in our domain corpus 3 . In the tax-and-regulations domain, the problem with terminology silence is particularly acute, as there are many instances of rules and arithmetic calculations concerning a specific entity which is mentioned once in the entire corpus. While This work is licensed under a Creative Commons Attribution 4.0 International License. License details: http:// creativecommons.org/licenses/by/4.0/.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "1 We are not focused on single-token terms. These were extracted in a separate base lexicon.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "2 Compositionality whereby the meaning of an expression is a function of the meaning of its immediate constituents and the syntactic rules used to combine them.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "3 Our corpus consists of a collection of tax forms and instructions for Canada (https://www.canada.ca/en/revenueagency/services/forms-publications.html) and the United States (https://www.irs.gov/forms-instructions). Approximately 20,000 multi-token terms were automatically extracted. The terms are not vetted by human experts. The parser itself is tasked with choosing the best listed terms for the raw utterance being parsed. the entity is not mentioned enough to be highly scored by collocation-based measures during the term extraction process, for the purposes of automatically interpreting and representing tax-and-regulations content, it is essential the entity be treated the same as other items in its class by the parser.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "This paper describes a simple method for extracting automatically multi-token terms that are not listed in the compliance terminology at the start of the parsing process of unlabelled utterances. The compliance terminology for taxes and regulations is the result of prior work on identifying, given the domain corpus, concepts and entities by means of co-occurrence/collocation-based surface statistical measures. In addition, linguistic-rule-based filters exclude a set of ill-formed terms from the final list. Because out-of-vocabulary (OOV) terms are tax-and-regulations-specific, we cannot rely on external lexical resources as these expressions are unlikely to be listed in general-purpose, financial or business lexicons. In our approach, the detection of OOV terms takes place during parsing. We use Generalized-Dice-Coefficient-based (GDC) metrics 4 to estimate the degree of similarity between established terms and OOV term candidates. When the GDC-based detection of multi-token terms is enabled during parsing of utterances, experiments show improvements of 93% in parsing accuracy for utterances with OOV terms at the start of the parsing process.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The goal of the compliance-domain parser is to output, in a simplified logical form (SLF) (Wang et al., 2015 , Constant et al., 2016 , a semantic representation of utterances taken from tax forms written in English 5 . Of particular interest, are utterances that express entire or partial arithmetic calculations. Downstream components of our NLP pipeline interpret the SLFs and automatically transform them into executable operations.", |
|
"cite_spans": [ |
|
{ |
|
"start": 90, |
|
"end": 108, |
|
"text": "(Wang et al., 2015", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 109, |
|
"end": 132, |
|
"text": ", Constant et al., 2016", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "In our compliance domain, the language of arithmetic calculations written in English is distributed along a continuum of syntactic complexity. ifte(or(eq(filing status(taxpayer),married filing jointly), eq(filing status(taxpayer),qualifying widow(er))),75300) Table 1 : Calculation utterances and corresponding SLF.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 260, |
|
"end": 267, |
|
"text": "Table 1", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The SLF of utterance 1 is an addition between the amount denoted by the multi-token term exclusion of income and the amount on a specific line of a specific form. Utterance 2 is about a choice: the smallest amount of two amounts (min operator). The notion of smallest is conveyed by a discontinuous dependency as a relative clause at the end of the utterance, namely, whichever is less. The first amount for the choice is a dollar constant amount; the second is an addition of the amounts denoted by the term employment income, expressed by total ... on lines 101 and 104. In utterance 3, inputting the constant amount of $75,300 to a calculation is conditional on satisfying one of the disjuncts (or operator) for the filing status of the taxpayer, expressed by the multi-token terms married filing jointly and qualifying widow(er).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "When multi-token terms are missing from the terminology, the task of the parser is to determine the dependencies between the individual tokens that make up the utterances. With more available tokens to parse, the chance of an inaccurate parse increases. Note that, with a large terminology (over 20,000 lexical entries), the prior acquisition of multi-token expressions that are in fact not multi-token domainconcepts is a possibility 7 . If multi-token terms contain pieces of calculations that should be discrete, the parser will not accurately break down the calculation into its constituents, since multi-token terms are monolithic literal strings to the parser-regardless of how many spaces there are between the tokens of a term.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Consider the expressions married and retired versus tuition and fees. The first expression is not a term; it describes a Boolean operation, where the NLP pipeline must check separately whether the taxpayer is both married and retired. In contrast, the second expression is a term describing a single entity which is only specific to educational-related forms and instructions. Tuition and fees does not occur frequently-enough in our corpus to have made it into the terminology. Extracting, during parsing, the OOV expression tuition and fees as a GDC-based term candidate prevents interpreting tuition and fees as a Boolean operation. The method identifies the OOV term tuition and fees as similar to pensions and annuities-a term in the existing terminology. Additional examples are provided in Utterances 1 and 2 differ by the tokens eligible versus qualified where eligible is part of the term eligible dependant in utterance 1. While 3 contains the single four-token term single parent qualified dependant, utterance 4 counts two separate two-token terms, namely, single parent and eligible dependant. In the case of utterances 2 and 4 with OOV terms, the parser parses qualified and single parent as left modifiers of the head of the noun phrase. In SLFs, the modifiers are enclosed in parentheses as arguments to the predicates, respectively dependant in utterance 2 and eligible dependant in utterance 4.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Intuitively, each of the utterances in Table 3 should be of the form if X, Y where X is a multi-token term with no internal structure for the parser to consume. Otherwise, there is a discrepancy in predicateargument relations across syntactically-and/or semantically-similar utterances 9 .", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 39, |
|
"end": 46, |
|
"text": "Table 3", |
|
"ref_id": "TABREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The SLFs of utterances 2 and 4 of Table 3 should be as in Table 4 below such that their SLFs align with those of utterances 1 and 3 of Even with an increase in the size of the domain corpus, there is no guarantee that the OOV terms qualified dependant or single parent eligible dependant of Table 3 will occur frequently-enough to make it into a term lexicon (as the result of a terminology extraction process) 10 . When processing arithmetic calculations automatically from utterances in English, our compliance system has one shot at outputting SLFs for each calculation such that downstream components can interpret them and transform them into executable operations. SLFs need be accurate. Bootstrapping parsing with a method that detects, on-the-fly, OOV term candidates increases parsing accuracy for these utterances.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 34, |
|
"end": 41, |
|
"text": "Table 3", |
|
"ref_id": "TABREF3" |
|
}, |
|
{ |
|
"start": 58, |
|
"end": 65, |
|
"text": "Table 4", |
|
"ref_id": "TABREF5" |
|
}, |
|
{ |
|
"start": 291, |
|
"end": 298, |
|
"text": "Table 3", |
|
"ref_id": "TABREF3" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Motivation", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "In order to measure the impact of detecting GDC-based terms at runtime on the success of the NLP pipeline that extracts and interprets arithmetic calculations in the tax-and-regulations domain, we ran the following experiment.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiment", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We used, as a baseline lexical resource, the latest version of the terminology created by our termextraction process. Separately, we fitted the parser with a preprocessing module to generate multi-token terms from each input utterance. In effect, the preprocessor functions as a chunker with a small grammar to define nominal phrases in English. The grammar sanctions the sequences of adjacent single-token words, which are grouped on the basis of linguistic properties (POS, semantic features, and/or WordNetbased similarity traits); in other words, the grammar rules out non-linguistic n-gram word sequences. For instance, for the utterance Capital gains on gifts of property to qualified donees, the preprocessor will generate two nominal multi-token term candidates capital gain and qualified donees 11 .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "For each term candidate generated by the preprocessor, the process retrieves, from the existing terminology, all terms that share a similar POS pattern. POS patterns need not be identical to allow for complimentary variance like spouse versus spouse's versus spousal in terms such as spouse amount, spouse's amount, and spousal amount. For instance, the POS pattern [NN, NN] is a permissible match with [JJ, NN] . used the GDC statistical measure (Park et al.2002 , Kozakov et al. 2004 13 . The idea was to measure the termness of the n-gram candidate terms formed at the onset of parsing with the term-generation preprocessor. Further, the GDC-values could help indicate whether, even weakly-associated n-gram word sequences with at least one identical anchor-token between the candidate and related terms in the terminology, can be extracted as candidate terms to use for parsing of the utterance. We can extract candidate terms whose combined tokens are lexically-cohesive to a certain limit. The GDC metrics used:", |
|
"cite_spans": [ |
|
{ |
|
"start": 403, |
|
"end": 411, |
|
"text": "[JJ, NN]", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 447, |
|
"end": 463, |
|
"text": "(Park et al.2002", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 464, |
|
"end": 485, |
|
"text": ", Kozakov et al. 2004", |
|
"ref_id": "BIBREF8" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "1. all of the terms listed in the terminology that have similar POS patterns.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "2. all of the POS patterns from the terms retrieved from the terminology.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "3. all of the terms with at least one token shared between a preprocessor-generated term candidate and the terms retrieved from the terminology.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "4. on all of the single tokens of the terms retrieved from the terminology.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "5. on all of the dictionary values associated with each of the single tokens that make up the terms retrieved from the terminology. Table 6 below lists some examples of GDC-based terms, scores and closest-related term from the terminology.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 132, |
|
"end": 139, |
|
"text": "Table 6", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "When parsing of an utterance completes, GDC-based terms, if any, are added to the terminology. Note that the compliance terminology does not consist merely of a list of terms, but rather the terminology is a dictionary of pairs like [key:values] . The GDC-based terms are added as new entries augmented with a set of default values from established related terms already in the terminology. Adding GDC-based terms to the terminology upon completion of the parsing of an utterance makes the terms immediately available to the utterances that remain, if any, in the parser's queue.", |
|
"cite_spans": [ |
|
{ |
|
"start": 233, |
|
"end": 245, |
|
"text": "[key:values]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Functional Setup", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "1. Tokenize utterance.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "2. For each single token, do morphological stemming and return base forms of individual tokens.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "3. Look up each base form in the single-token base lexicon. Retrieve all lexical data of base form. 4. Generate POS-based term candidates (all legal combinations from left to right) that resolve in a nominal phrase of preset length.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "eligible in a two-token term. With qualified, it is the reverse-a four-token term but no two-token term.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "11 The preprocessor rules out patterns that include multiple prepositions. For a discussion on term extraction and the nature of terms, see Boguraev et al., 2015. 12 The Penn tags NN, NNP, NNPS, JJ, JJR, JJS, VBG and VBN correspond to, respectively, singular noun, singular proper noun, plural proper noun, adjective, comparative adjective, superlative adjective, gerund/present participle and past participle.", |
|
"cite_spans": [ |
|
{ |
|
"start": 140, |
|
"end": 165, |
|
"text": "Boguraev et al., 2015. 12", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "13 Park et al., 2002 and Kozakov et al., 2004 NN,NN] 1.00 0.33 qualified retirement plan Table 6 : New terms containing qualified detected at runtime. 5. For each term candidate generated in step 4, check current baseline terminology for possible matches. If term candidate exists in terminology, remove candidate from term-candidate list generated in step 4.", |
|
"cite_spans": [ |
|
{ |
|
"start": 3, |
|
"end": 24, |
|
"text": "Park et al., 2002 and", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 25, |
|
"end": 45, |
|
"text": "Kozakov et al., 2004", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 46, |
|
"end": 52, |
|
"text": "NN,NN]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 89, |
|
"end": 96, |
|
"text": "Table 6", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "6. For each term candidate generated in step 4, retrieve all terms that have similar POS patterns. Retrieve the terms as [key:values] pairs.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "7. Compute all the GDC metrics as described in subsection 3.1 above.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "The terms in the leftmost column of Table 6 are terms produced by our approach.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 36, |
|
"end": 43, |
|
"text": "Table 6", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Steps for Detecting OOV Terms", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "The goal of our evaluation was two-fold:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "\u2022 Determine the number of new multi-token terms extracted on-the-fly during parsing.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "\u2022 Evaluate the impact of GDC-based terms on parsing accuracy.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "In one experiment, we created a corpus of unlabelled utterances collected from tax forms and instructions published in 2017 by the Internal Revenue Service of the United States. We further narrowed the focus to utterances that included the token qualified. The test set counted 6,553 utterances with the token qualified .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "First, we parsed each utterance in the test set only with the latest version of the terminology to create a baseline consisting of utterances paired with their corresponding SLF; the output name is BTPrun 14 . Second, we parsed each utterance in the test set with the baseline terminology used in BTPrun but with GDC-based-extraction enabled; the output is called BTGDCrun 15 .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "In the BTPrun, 1099 multi-token terms with qualified were detected. In contrast, at the end of BT-GDCrun, 1,380 terms were extracted-a gain of 281 new multi-token terms extracted during parsing (or 4.3%) (see Table 7 ).", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 209, |
|
"end": 216, |
|
"text": "Table 7", |
|
"ref_id": "TABREF10" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "Finally, we had two human annotators perform a quality assessment of the SLFs with GDC-based terms acquired during parsing. A comparison of the SLFs output by each of the runs yielded 724 changes in the output of the BTGDCrun. The annotation schema was simple: given the utterance, is the new SLF correct? The evaluation was binary: 1 for correct SLF, 0 for ill-formed SLF. Even with a stringent schema for manual evaluation, annotators may judge outcomes differently. However, for this task, judgments were close. Averaging the numbers of SLFs judged as having been corrected through the detection of terms during parsing, we saw a 93% improvement in parsing accuracy for the test set (see Table 8 below). In addition, for each improvement, the annotators were asked to classify the changes in the modified SLFs according to at least one of the following three categories:", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 691, |
|
"end": 698, |
|
"text": "Table 8", |
|
"ref_id": "TABREF11" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "\u2022 In examples 1-3 of the BTGDCrun, the SLFs include novel multi-token terms. The SLF of example 2 is more accurate as activities is included as a content-bearing token of the term. In example 3 of the BTGDCrun, the detection of the OOV term qualified retirement plan enables the parser to push one possible interpretation for the content in the parenthetical material. Note that, in the BTPrun, the parser did not output any SLF, which is indicated by the convention UNSPECIFIED. Table 10 summarizes the classification by each of the human annotators. Interestingly, the annotators' judgments about the class of improvements by GDC-based terms on SLFs align.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 480, |
|
"end": 488, |
|
"text": "Table 10", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "An additional advantage for this method occurs further downstream in the NLP pipeline, where elements of parsed phrases are matched to an internal data model. When executing the calculation for ifte(qualified dependant,85.00), we use a custom-built entity-recognition system to determine the value of qualified dependant. One of the features of this entity recognition system is consecutive token matches. As the internal named entity has a description of IsQualifiedDependant, having the parser output ifte(qualified dependent,85.00) instead of ifte(dependent(qualified),85.00) increases the likelihood of predicting the correct entity.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Additional Downstream Benefits", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "In this paper, we have described a method which increases term recall and improves parsing accuracy of utterances with OOV terms at the start of the parsing process. In addition, multi-token terms detected at parsing runtime are automatically added to the existing terminology. We use a parser fitted with a term-generation preprocessor to identify similarity between OOV multi-token term candidates and multi-token terms listed in the terminology. We observe improvements not only in the interpretation and representation of the utterances by our parser but also in the transformation of the SLFs into executable operations in downstream components of our NLP pipeline.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "This method is best suited for domains where precision in a term lexicon, which has been automatically extracted, is important and where the problem with term silence can be severe 16 In the future, we would like to experiment with beginning the parsing process by using a terminology manually curated by domain experts. Because the method is domain-agnostic, we do not believe the discovery of OOV terms and augmentation of a domain-specific terminology (as long as there is a preexisting terminology at outset of a parsing task) is constrained to our example domain.", |
|
"cite_spans": [ |
|
{ |
|
"start": 181, |
|
"end": 183, |
|
"text": "16", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "This work has shown that it is possible to overcome term silence by adding functionality to a parser with a preprocessor to discover OOV terms on the fly.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "The Generalized Dice coefficient is a similarity measure used in lexicography and term extraction to compute the lexical cohesion of multi-token term candidates.Park et al., 2002 andKozakov et al., 2004 discuss at length the Dice-Coefficient and the Generalized-Dice-Coefficient statistical measures.5 We use a lexical-functional-based parser developed in-house for the compliance domain. In particular, it is tailored to interpret unannotated utterances that express arithmetic calculations written in English 6 ifte in SLFs is a token to denote Boolean branching conditions.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Extraction of invalid term candidates is called noise-the opposite of silence. 8 The part-of-speech (POS) names are Penn tags where JJ stands for adjective, NN for singular noun, and NNP for proper noun.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Boguraev et al (2015) make a similar observation about ESG, the parser used in IBM Watson. 10 Note that, in the examples discussed inTable 3, term extraction did not detect a four-token term with eligible but did detect", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "BTPrun stands for parsing with baseline terminology and baseline parser with no GDC-built-in method.15 BTGDCrun stands for parsing with baseline terminology and with parser where GDC-based extraction is enabled.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Our method can introduce noise, i.e. wrongly glom a multi-token expression as a single term. However, the individual tokens of a candidate term created by the preprocessor remain available to the parser, which relies on additional corpus-based decision heuristics to rank SLF parses with or without GDC-based terms.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "We gratefully acknowledge the comments by three anonymous reviewers. Not everyone agreed with everything but each comment and suggestion helped the co-authors clarify points in the paper and start a TODO list for future development and testing.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgements", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Compositionality and multiword expressions: Six of one, half a dozen of the other?", |
|
"authors": [ |
|
{ |
|
"first": "Timothy", |
|
"middle": [], |
|
"last": "Baldwin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of the Workshop on Multiword Expressions: Identifying and Exploiting Underlying Properties", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Timothy Baldwin. 2006. Compositionality and multiword expressions: Six of one, half a dozen of the other? In Proceedings of the Workshop on Multiword Expressions: Identifying and Exploiting Underlying Properties, page 1, Sydney, Australia, July. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "The bare necessities: Increasing lexical coverage for multi-word domain terms with less lexical data", |
|
"authors": [ |
|
{ |
|
"first": "Branimir", |
|
"middle": [], |
|
"last": "Boguraev", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Esme", |
|
"middle": [], |
|
"last": "Manandise", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Benjamin", |
|
"middle": [], |
|
"last": "Segal", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the 11th Workshop on Multiword Expressions", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "60--64", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Branimir Boguraev, Esme Manandise, and Benjamin Segal. 2015. The bare necessities: Increasing lexical cov- erage for multi-word domain terms with less lexical data. In Proceedings of the 11th Workshop on Multiword Expressions, pages 60-64, Denver, Colorado, June. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Words! words! words!: Teaching the language of tax", |
|
"authors": [ |
|
{ |
|
"first": "Stephen", |
|
"middle": [], |
|
"last": "Cohen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Journal of legal education", |
|
"volume": "55", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Stephen Cohen. 2005. Words! words! words!: Teaching the language of tax. Journal of legal education, 55, 12.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Deep lexical segmentation and syntactic parsing in the easy-first dependency framework", |
|
"authors": [ |
|
{ |
|
"first": "Matthieu", |
|
"middle": [], |
|
"last": "Constant", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Joseph", |
|
"middle": [ |
|
"Le" |
|
], |
|
"last": "Roux", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Nadi", |
|
"middle": [], |
|
"last": "Tomeh", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of NAACL-HLT 2016", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1095--1101", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Matthieu Constant, Joseph Le Roux, and Nadi Tomeh. 2016. Deep lexical segmentation and syntactic parsing in the easy-first dependency framework. In Proceedings of NAACL-HLT 2016, pages 1095-1101. Association for Computational Linguistics, June.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "A corpus of Australian contract language: Description, profiling and analysis", |
|
"authors": [ |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Curtotti", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Eric", |
|
"middle": [], |
|
"last": "Mccreath", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Michael Curtotti and Eric McCreath. 2011. A corpus of Australian contract language: Description, profiling and analysis. pages 199-208, 06.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Bert: Pre-training of deep bidirectional transformers for language understanding", |
|
"authors": [ |
|
{ |
|
"first": "Jacob", |
|
"middle": [], |
|
"last": "Devlin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming-Wei", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kenton", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kristina", |
|
"middle": [], |
|
"last": "Toutanova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirec- tional transformers for language understanding.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "A well-founded ontological framework for modeling personal income tax", |
|
"authors": [ |
|
{ |
|
"first": "Isabella", |
|
"middle": [], |
|
"last": "Distinto", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Nicola", |
|
"middle": [], |
|
"last": "Guarino", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Claudio", |
|
"middle": [], |
|
"last": "Masolo", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Isabella Distinto, Nicola Guarino, and Claudio Masolo. 2013. A well-founded ontological framework for model- ing personal income tax. 06.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Can recognising multiword expressions improve shallow parsing?", |
|
"authors": [ |
|
{ |
|
"first": "Ioannis", |
|
"middle": [], |
|
"last": "Korkontzelos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Suresh", |
|
"middle": [], |
|
"last": "Manandhar", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "636--644", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ioannis Korkontzelos and Suresh Manandhar. 2010. Can recognising multiword expressions improve shallow parsing? In Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pages 636-644, Los Angeles, California, June. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Glossary extraction and utilization in the information search and delivery system for ibm technical support", |
|
"authors": [ |
|
{ |
|
"first": "Lev", |
|
"middle": [], |
|
"last": "Kozakov", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Youngja", |
|
"middle": [], |
|
"last": "Park", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Youssef", |
|
"middle": [], |
|
"last": "Tong-Haing Fin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yurdaer", |
|
"middle": [], |
|
"last": "Drissi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Thomas", |
|
"middle": [], |
|
"last": "Doganata", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Cofino", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "IBM SYSTEMS JOURNAL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "546--653", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lev Kozakov, Youngja Park, Tong-Haing Fin, Youssef Drissi, Yurdaer Doganata, and Thomas Cofino. 2004. Glossary extraction and utilization in the information search and delivery system for ibm technical support. In IBM SYSTEMS JOURNAL, pages 546-653. IBM.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Determining compositionality of word expressions using word space models", |
|
"authors": [ |
|
{ |
|
"first": "Lubom\u00edr", |
|
"middle": [], |
|
"last": "Krcm\u00e1r", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Karel", |
|
"middle": [], |
|
"last": "Jezek", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Pavel", |
|
"middle": [], |
|
"last": "Pecina", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "MWE@NAACL-HLT", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lubom\u00edr Krcm\u00e1r, Karel Jezek, and Pavel Pecina. 2013. Determining compositionality of word expressions using word space models. In MWE@NAACL-HLT.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Towards unlocking the narrative of the United States income tax forms", |
|
"authors": [ |
|
{ |
|
"first": "Esme", |
|
"middle": [], |
|
"last": "Manandise", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the Second Financial Narrative Processing Workshop (FNP 2019)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "33--41", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Esme Manandise. 2019. Towards unlocking the narrative of the United States income tax forms. In Proceedings of the Second Financial Narrative Processing Workshop (FNP 2019), pages 33-41, Turku, Finland, September. Link\u00f6ping University Electronic Press.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Parsing and mwe detection: Fips at the parseme shared task", |
|
"authors": [ |
|
{ |
|
"first": "Luka", |
|
"middle": [], |
|
"last": "Nerima", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Vasiliki", |
|
"middle": [], |
|
"last": "Foufi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Eric", |
|
"middle": [], |
|
"last": "Wehrli", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Proceedings of the 13th Workshop on Multiword Expressions (MWE 2017)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "54--59", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Luka Nerima, Vasiliki Foufi, and Eric Wehrli. 2017. Parsing and mwe detection: Fips at the parseme shared task. In Proceedings of the 13th Workshop on Multiword Expressions (MWE 2017), pages 54-59, Valencia, Spain, April. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Automatic glossary extraction: Beyond terminology identification", |
|
"authors": [ |
|
{ |
|
"first": "Youngja", |
|
"middle": [], |
|
"last": "Park", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Roy", |
|
"middle": [], |
|
"last": "Byrd", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Branimir", |
|
"middle": [], |
|
"last": "Boguraev", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of the 19th International Conference on Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "60--64", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Youngja Park, Roy Byrd, and Branimir Boguraev. 2002. Automatic glossary extraction: Beyond terminology identification. In Proceedings of the 19th International Conference on Computational Linguistics, pages 60-64, Denver, Colorado, January. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Building a semantic parser overnight", |
|
"authors": [ |
|
{ |
|
"first": "Yushi", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jonathan", |
|
"middle": [], |
|
"last": "Berant", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Percy", |
|
"middle": [], |
|
"last": "Liang", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "1332--1342", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yushi Wang, Jonathan Berant, and Percy Liang. 2015. Building a semantic parser overnight. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1332-1342, Beijing, China, July. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "The relevance of collocations for parsing", |
|
"authors": [ |
|
{ |
|
"first": "Eric", |
|
"middle": [], |
|
"last": "Wehrli", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 10th Workshop on Multiword Expressions (MWE)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "26--32", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Eric Wehrli. 2014. The relevance of collocations for parsing. In Proceedings of the 10th Workshop on Multiword Expressions (MWE), pages 26-32, Gothenburg, Sweden, April. Association for Computational Linguistics.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF0": { |
|
"num": null, |
|
"text": "lists some pairs of utterances and their corresponding SLF (terms are denoted by the underscore) 6 .", |
|
"type_str": "table", |
|
"content": "<table><tr><td colspan=\"2\">No. Calculation Utterances</td><td>SLF</td></tr><tr><td>1</td><td>Enter the sum of exclusion of in-</td><td>add(exclusion of income(puerto rico),form(4563,line(15)))</td></tr><tr><td/><td>come from Puerto Rico and form</td><td/></tr><tr><td/><td>4563 line 15.</td><td/></tr><tr><td>2</td><td>Enter $1,195 or the total of</td><td>min(1195, employment income(add(line(101),line(104))))</td></tr><tr><td/><td>your employment income you</td><td/></tr><tr><td/><td>reported on lines 101 and 104 of</td><td/></tr><tr><td/><td>your return, whichever is less.</td><td/></tr><tr><td>3</td><td>Enter $75,300 if married filing</td><td/></tr><tr><td/><td>jointly or qualifying widow(er)</td><td/></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF1": { |
|
"num": null, |
|
"text": "", |
|
"type_str": "table", |
|
"content": "<table><tr><td>8 .</td></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF2": { |
|
"num": null, |
|
"text": "GDC-based term candidates and similarity to terms in terminology.Finally, the interpretation of utterances with and without OOV terms is reflected in the corresponding SLF. Contrast utterances inTable 3.", |
|
"type_str": "table", |
|
"content": "<table><tr><td>No. Term Utterances</td><td>SLF</td></tr><tr><td>1 Yes amount for an eligible dependant,</td><td>ifte(eligible dependant,85.00)</td></tr><tr><td>claim $85.00</td><td/></tr><tr><td>2 OOV amount for a qualified dependant,</td><td>ifte(dependant(qualified),85.00)</td></tr><tr><td>claim $85.00</td><td/></tr><tr><td>3 Yes amount for a single parent's qualified</td><td>ifte(single parent qualified dependant,64.00)</td></tr><tr><td>dependant, claim $64.00</td><td/></tr><tr><td>4 OOV amount for a single parent's eligible de-</td><td>ifte(eligible dependant(single parent),64.00)</td></tr><tr><td>pendant, claim $64.00</td><td/></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF3": { |
|
"num": null, |
|
"text": "Utterances and SLF with/without terms.", |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"html": null |
|
}, |
|
"TABREF4": { |
|
"num": null, |
|
"text": "(repeated in Table 4).", |
|
"type_str": "table", |
|
"content": "<table><tr><td>No. Term Utterances</td><td>SLF</td></tr><tr><td>1 Yes amount for an eligible dependant,</td><td>ifte(eligible dependant,85.00)</td></tr><tr><td>claim $85.00</td><td/></tr><tr><td>2 Yes amount for a qualified dependant,</td><td>ifte(qualified dependant,85.00)</td></tr><tr><td>claim $85.00</td><td/></tr><tr><td>3 Yes amount for a single parent's qualified</td><td>ifte(single parent qualified dependant,64.00)</td></tr><tr><td>dependant, claim $64.00</td><td/></tr><tr><td>4 Yes amount for a single parent's eligible de-</td><td>ifte(single parent eligible dependant,64.00)</td></tr><tr><td>pendant, claim $64.00</td><td/></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF5": { |
|
"num": null, |
|
"text": "", |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"html": null |
|
}, |
|
"TABREF6": { |
|
"num": null, |
|
"text": "", |
|
"type_str": "table", |
|
"content": "<table><tr><td>POS Pattern</td><td>Permissible Matches</td></tr><tr><td>[NN, NN]</td><td>[NN. NN]</td></tr><tr><td/><td>[NNP, NN]</td></tr><tr><td/><td>[NNPS, NN]</td></tr><tr><td/><td>[JJ, NN]</td></tr><tr><td/><td>[JJR, NN]</td></tr><tr><td/><td>[JJS, NN]</td></tr><tr><td/><td>[VBG, NN]</td></tr><tr><td/><td>[VBN, NN]</td></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF7": { |
|
"num": null, |
|
"text": "Permissible POS-pattern matches.", |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"html": null |
|
}, |
|
"TABREF10": { |
|
"num": null, |
|
"text": "Totals and percentages of terms detected during parsing of utterances in test set.", |
|
"type_str": "table", |
|
"content": "<table><tr><td>Total of Utterances in</td><td>Total of modified SLFs</td><td>Number of Improved</td></tr><tr><td>Test Set</td><td>in BTGDCrun</td><td>SLFs</td></tr><tr><td>6,553</td><td>724</td><td/></tr><tr><td>Annotator 1</td><td/><td>669</td></tr><tr><td>Annotator 2</td><td/><td>675</td></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF11": { |
|
"num": null, |
|
"text": "Manual evaluation of modified SLFs in BTGDCrun.", |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"html": null |
|
}, |
|
"TABREF13": { |
|
"num": null, |
|
"text": "SLF improvements with GDC-based terms.", |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"html": null |
|
}, |
|
"TABREF14": { |
|
"num": null, |
|
"text": "Classification of modified SLFs in BTGDCrun.", |
|
"type_str": "table", |
|
"content": "<table/>", |
|
"html": null |
|
} |
|
} |
|
} |
|
} |