|
{ |
|
"paper_id": "W19-0111", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T06:17:56.261180Z" |
|
}, |
|
"title": "Learning complex inflectional paradigms through blended gradient inputs", |
|
"authors": [ |
|
{ |
|
"first": "Eric", |
|
"middle": [], |
|
"last": "Rosen", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Johns Hopkins University", |
|
"location": {} |
|
}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "Through Gradient Symbolic Computation (Smolensky and Goldrick, 2016), in which input forms can consist of gradient blends of more than one phonological realization, we propose a way of deriving surface forms in complex inflectional paradigms that dispenses with direct references to inflectional classes and relies solely on relatively simple blends of input expressions.", |
|
"pdf_parse": { |
|
"paper_id": "W19-0111", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "Through Gradient Symbolic Computation (Smolensky and Goldrick, 2016), in which input forms can consist of gradient blends of more than one phonological realization, we propose a way of deriving surface forms in complex inflectional paradigms that dispenses with direct references to inflectional classes and relies solely on relatively simple blends of input expressions.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "In languages whose inflectional systems have a highly complex paradigmatic structure, it becomes a challenge to explain how a speaker can produce a correct inflectional form when the number of possible forms is exceedingly large. As Ackerman and Malouf (2013, p. 429 ) (henceforth A&M) comment: \"That speakers are able to do this is a truly puzzling accomplishment given the extraordinary variation and complexity attested in the morphological systems of the world.\" When it becomes extremely difficult for a speaker to memorize every possible inflectional form for every lexeme in their language, it becomes attractive to posit some means by which they can produce a correct unmemorized form.", |
|
"cite_spans": [ |
|
{ |
|
"start": 233, |
|
"end": 266, |
|
"text": "Ackerman and Malouf (2013, p. 429", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Three interacting types of paradigmatic complexity that put a burden on learning are addressed here: (a) differences in inflectional material across lexemes, which, descriptively, result in division of lexemes into inflectional classes; (b) syncretism, where the same inflectional material occurs in different paradigm cells; (c) independent subsystems of inflectional class behaviour, evident in Mazatec (see \u00a73.) These complexities are all ways in which inflectional patterns depart from canonicity as defined in detail by Corbett (2009) ; Baerman and Corbett (2010) ; Stump (2016) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 525, |
|
"end": 539, |
|
"text": "Corbett (2009)", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 542, |
|
"end": 568, |
|
"text": "Baerman and Corbett (2010)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 571, |
|
"end": 583, |
|
"text": "Stump (2016)", |
|
"ref_id": "BIBREF23" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Through Gradient Symbolic Computation (Smolensky and Goldrick, 2016) , a type of Harmonic Grammar, and with examples from Russian and Mazatec 1 , we propose a system that enables a speaker to produce correct forms of complex paradigms through learnable input representations without indexing to inflectional classes.", |
|
"cite_spans": [ |
|
{ |
|
"start": 38, |
|
"end": 68, |
|
"text": "(Smolensky and Goldrick, 2016)", |
|
"ref_id": "BIBREF21" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "The paper is organized as follows. \u00a71 introduces GSC and how it can be applied to learning exponents of inflectional paradigms. \u00a72 shows how Russian noun inflection can be acquired through this framework. \u00a72.1 shows how GSC limits the kinds of inflectional patterning that are possible.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "\u00a73 analyses complex paradigms in Mazatec, where inflectional classes vary in three cross-cutting dimensions. \u00a74 discusses testing and comparison with other models. \u00a75 summarizes how this framework can both explain departures from canonicity in inflectional paradigms and also constrain the degree of departure from the canonical.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We adopt here Gradient Symbolic Computation (henceforth GSC), in which gradient inputs are given numerical activation levels, which we shall show to be learnable, and are evaluated with weighted constraints that calculate the Harmony of input candidates, where the candidate with the greatest Harmony surfaces. This formalism is part of a larger research program in which computation derives outputs from gradient representations in phonology, syntax and semantics (Cho et al., 2017; Faust and Smolensky, 2017; Faust, 2017; Goldrick et al., 2016; Hsu, 2018; M\u00fcller, 2017; Rosen, 2016 Rosen, , 2018 Smolensky et al., 2014; Smolensky and Goldrick, 2016; Van Hell et al., 2016; Zimmermann, 2017b,a, forthcoming). 2 In the examples in \u00a72 from Russian noun inflection, the surface exponents that can represent genitive singular among the descriptive inflectional classes, form a blend of input segments, which we shall call 'inflectional input': {a, i}. A lexeme (basic element of the lexicon) in descriptive class 1 whose genitive sg. is vina, 'wine' contributes an input we shall call 'base input' that consists of what is traditionally thought of as a stem plus a blend of word-final segments {a, u, o} that mirrors segments that can occur in surface-final position. We propose that it derives as follows, with numerical details shown later in table 4.", |
|
"cite_spans": [ |
|
{ |
|
"start": 465, |
|
"end": 483, |
|
"text": "(Cho et al., 2017;", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 484, |
|
"end": 510, |
|
"text": "Faust and Smolensky, 2017;", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 511, |
|
"end": 523, |
|
"text": "Faust, 2017;", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 524, |
|
"end": 546, |
|
"text": "Goldrick et al., 2016;", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 547, |
|
"end": 557, |
|
"text": "Hsu, 2018;", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 558, |
|
"end": 571, |
|
"text": "M\u00fcller, 2017;", |
|
"ref_id": "BIBREF16" |
|
}, |
|
{ |
|
"start": 572, |
|
"end": 583, |
|
"text": "Rosen, 2016", |
|
"ref_id": "BIBREF19" |
|
}, |
|
{ |
|
"start": 584, |
|
"end": 597, |
|
"text": "Rosen, , 2018", |
|
"ref_id": "BIBREF20" |
|
}, |
|
{ |
|
"start": 598, |
|
"end": 621, |
|
"text": "Smolensky et al., 2014;", |
|
"ref_id": "BIBREF22" |
|
}, |
|
{ |
|
"start": 622, |
|
"end": 651, |
|
"text": "Smolensky and Goldrick, 2016;", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 652, |
|
"end": 674, |
|
"text": "Van Hell et al., 2016;", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 675, |
|
"end": 711, |
|
"text": "Zimmermann, 2017b,a, forthcoming). 2", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Gradient Symbolic Computation", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Input:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lexical base Inflectional affix", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "root vin {a, u, o } GEN.SG. = {a, i} Output:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lexical base Inflectional affix", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "vin a root exponent affixal exponent In this framework, a discrete output form is chosen from possible candidates by the action of Faithfulness constraints that are familiar from Optimality Theory (Prince and Smolensky, 1993) , but which, in GSC, have weighted values and evaluate gradient input activations. We assume a highly-weighted quantization constraint quantified by an equation in Cho et al. (2017) that disfavours blended outputs and gives a higher Harmony to discrete non-blended forms. The tableaux below thus ignore blended output candidates. 3 Surface material that some models view as deriving solely from an affix input is derived here from two distinct input sources, thus blending two competing analyses: a constructivist approach that builds whole words out of morphemes and a wordbased approach that seeks relationships between whole word forms. The input representation of a lexeme that includes affixal material as an intrinsic part of the word is based upon a wholeword model; an affix as a fully independent input follows a morpheme-based model. A simi-2 A reviewer asks what advantages this model has over a \"genuine connectionist model\" such as Goldsmith and O'Brien (2006) ; Kann and Sch\u00fctze (2016) ; Malouf (2018) . In this model, the knowledge it contains is completely transparent, whereas in the last two models cited by the reviewer, it is not clear what kind of knowledge is contained in those networks. (See \u00a74 for further discussion.)", |
|
"cite_spans": [ |
|
{ |
|
"start": 197, |
|
"end": 225, |
|
"text": "(Prince and Smolensky, 1993)", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 390, |
|
"end": 407, |
|
"text": "Cho et al. (2017)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 556, |
|
"end": 557, |
|
"text": "3", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1171, |
|
"end": 1199, |
|
"text": "Goldsmith and O'Brien (2006)", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 1202, |
|
"end": 1225, |
|
"text": "Kann and Sch\u00fctze (2016)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 1228, |
|
"end": 1241, |
|
"text": "Malouf (2018)", |
|
"ref_id": "BIBREF15" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lexical base Inflectional affix", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "3 But see Zimmermann (2017b Zimmermann ( ,a, 2018 for analyses of other phenomena in which outputs are also gradient. lar approach is taken by Smolensky and Goldrick (2016) , treating French liaison as derived from both the end of a preceding word and the beginning of a following word, thus combining two competing approaches to liaison in the literature. Any matching phonological material from the two sources will combine through coalescence. When there are multiple descriptive inflectional classes, the exponent for a given inflectional combination varies, depending not only on the lexeme, but on the stem when there is stem allomorphy (Stump, 2016) . The current proposal for blended word-final input segments that mirror affixes, directly encodes the fact that lexeme/stem choice will affect affix-choice. Non-concatenative and suppletive alternations can thus be represented as well, as blends of different stem alternants occurring underlyingly.", |
|
"cite_spans": [ |
|
{ |
|
"start": 10, |
|
"end": 27, |
|
"text": "Zimmermann (2017b", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 28, |
|
"end": 49, |
|
"text": "Zimmermann ( ,a, 2018", |
|
"ref_id": "BIBREF27" |
|
}, |
|
{ |
|
"start": 143, |
|
"end": 172, |
|
"text": "Smolensky and Goldrick (2016)", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 643, |
|
"end": 656, |
|
"text": "(Stump, 2016)", |
|
"ref_id": "BIBREF23" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lexical base Inflectional affix", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "CLASS 1 2 3 4 SINGULAR NOM -o -\u2205 -a -\u2205 ACC -o -\u2205 -u -\u2205 GEN -a -a -i -i DAT -u -u -e -i LOC -e -e -e -i INST -om -om -oj -ju PLURAL NOM -a -i -i -i ACC -a -i -i -i GEN -\u2205 -ov -\u2205 -ej DAT -am -am -am -am LOC -ax -ax -ax -ax INST -am'i -am'i -am'i -am'i", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lexical base Inflectional affix", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "For ease of exposition, we first examine the relatively simple paradigms of case/number noun suffixes in Russian, with data from A&M p. 460 given in table 2. For presentation purposes, we ignore here the paradigm defectiveness of some Russian nouns (Corbett, 2007) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 249, |
|
"end": 264, |
|
"text": "(Corbett, 2007)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The exponent that surfaces for a given stem and person/number combination is the one with the highest aggregate activation that surpasses a threshold determined by MAX and DEP constraints to be defined below. When two instances of the same exponent occur in both base and inflectional input, they can coalesce together in the output, with an aggregate input activa- There are then four possible candidate segments that have a non-zero input activation: /-a/, /-u/, /-o/ and /-i/, whose input activations will be the sums of any pair of co-occurring segments that can coalesce in the output as shown in table 3.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The segment a surpasses the relevant threshold and surfaces, having the highest aggregate input activation. Table 4 shows a harmonic tableau for this form. In GSC, a MAX constraint (here weighted 5.0) contributes positive Harmony proportionate to the surfacing of underlying activation. A DEP constraint (here weighted at \u22122.0) contributes negative Harmony for the deficit between input and output activations. We assume that quantization will only allow surface activations of 0 or 1 so the tableau only considers candidates with those values. The positive contribution to Harmony from MAX for the winning candidate is is 5 times the sum of its input activations for /a/ which are 0.1 from the base input plus 0.3 from gen.sg. inflection input. The winning candidate has the highest Harmony.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 108, |
|
"end": 115, |
|
"text": "Table 4", |
|
"ref_id": "TABREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "An affixless candidate incurs no MAX reward or DEP penalty and will be optimal when all other candidates have negative Harmony (table 5.) The following simple algorithm 4 learned blended base and inflectional input values such that the optimal output candidate is the correct target of learning for all 48 forms. Following Faust and Smolensky (2017, page 2), not just an individual segment but also other structures can have an activity level.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "vin(0.1 \u2022 a, 0.02 \u2022 u, 0.16 \u2022 o) 'wine' + GEN-PL(0.28 \u2022 ov, 0.22 \u2022 ej) MAX DEP Harmony 5 \u22122 vin-a 0.5 \u22121.8 \u22121.3 vin-u 0.1 \u22121.96 \u22121.86 vin-o 0.8 \u22121.68 \u22120.88 vin-ov 1.4 \u22121.44 \u22120.04 vin-ej 1.1 \u22121.56 \u22120.46 vin-e \u22122.0 \u22122.0 vin 0.0", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "\u2022 Initialize all activations at zero.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "\u2022 Calculate the Harmony for each descriptive stem-class/number/case combination.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "\u2022 If the wrong affix is predicted, decrease its two input activations if nonzero and increase the activations of the desired affix. Stepsize 0.02.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "\u2022 Repeat until all 48 paradigm positions are correctly predicted. Tables 6 and 7 show the learned input values after 72 iterations.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 66, |
|
"end": 80, |
|
"text": "Tables 6 and 7", |
|
"ref_id": "TABREF7" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Nonzero inputs for an inflectional combination indicate its set of possible exponents. A cooccurring nonzero base input narrows down the choice of those exponents. For example, the locative singular affix is -e for all descriptive classes except class 4, where -i surfaces. The activation of 0.12 for -i for a class 4 base input allows it to surpass other competitors where it surfaces.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "In summary, representing both base and inflectional inputs with a blend of partially activated exponents allows a speaker to encode all the information in a multi-class inflectional paradigm directly on the relevant input forms. In the impossible paradigm, lines connecting the two instances of i (coloured blue) and connecting the two instances of e (coloured red) would cross. Such diagonal crossing can be shown to mathematically imply activation hierarchies that lead to a contradiction (table 9) , where \u03b9 represents the activation of i and \u03f5 of e for a given (subscripted) feature value or descriptive inflectional class.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 491, |
|
"end": 500, |
|
"text": "(table 9)", |
|
"ref_id": "TABREF10" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "\u03b9 g.sg. + \u03b9 3 > \u03f5 g.sg. + \u03f5 3", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "(1)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "\u03f5 g.sg. + \u03f5 4 > \u03b9 g.sg. + \u03b9 4 (2) \u03b9 dat.sg. + \u03b9 4 > \u03f5 dat.sg. + \u03f5 4 (3) \u03f5 dat.sg. + \u03f5 3 > \u03b9 dat.sg. + \u03b9 3 (4) (1) + (2) : \u03b9 3 + \u03f5 4 > \u03f5 3 + \u03b9 4", |
|
"eq_num": "(5)" |
|
} |
|
], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "(3) + (4) :", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "\u03b9 4 + \u03f5 3 > \u03f5 4 + \u03b9 3 (6)", |
|
"eq_num": "(6)" |
|
} |
|
], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "contradicts (5) Thus, in spite of a common misconception that \"you can do anything with numbers\", the proposed GSC model predicts that many conceivable paradigms are impossible.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning Russian noun inflection", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "We now consider verb paradigms in Chiquihuitl\u00e1n Mazatec, a language with a high degree of paradigmatic complexity. Jamieson (1982) shows that its verbal inflectional morphology has three separate dimensions in the affirmative forms of the neutral and incompletive aspects: (a) a stem formative (table 12) 1 b a 3 sae 1\u010d a 2 s\u0129 24 k w a 3 sae 1\u010d a 4 s\u0129 24 2\u010d a 2 se 2\u010d a 2 s\u0169 2\u010d a 4 se 2\u010d a 4 s\u0169 2 3 b a 3 se 2 k w a 4 se 2 possible combinations are actually reduced to 20 in Jamieson's list of paradigms. Table 10 gives the paradigm for ba 3 se 2 'remember' from A&M:450, taken from Jamieson (1982, p. 167) . This lexeme is in stem formative descriptive class 11, tone pattern class B 3-1 and final vowel class 2. Combinations of morphosyntactic features are represented by combinations of multiple exponents, but there is no apparent correspondence between individual feature values and individual exponents. The surface form for a verb expresses feature values for aspect, person and number, with a stem formative, a tone pattern and a suffix vowel.", |
|
"cite_spans": [ |
|
{ |
|
"start": 115, |
|
"end": 130, |
|
"text": "Jamieson (1982)", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 583, |
|
"end": 606, |
|
"text": "Jamieson (1982, p. 167)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 294, |
|
"end": 304, |
|
"text": "(table 12)", |
|
"ref_id": "TABREF0" |
|
}, |
|
{ |
|
"start": 505, |
|
"end": 513, |
|
"text": "Table 10", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Learning inputs in Mazatec", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Stems in many cases consist merely of a single segment. As A&M:447 observe, a stem is much more easily identified by \"its membership in an inflectional class in each of three distinct crosscutting dimensions: tone pattern, final vowel, and stem formative.\" In the proposed blended representations, a lexeme's base input carries much of the information that predicts what exponents it occurs with, in this way identifying what can descriptively be called its stem. It also directly encodes what is descriptively its inflectional class in a way that is not possible in a model in which paradigm information and stem information are separate. Table 11 shows, with activation values omitted for now, an example basic input-output structure, which contains gradient blends of consonants in the input. can be explained by our analysis. These include suffixing of verbs, verb compounding, verbs with no stem formative 5 and stem suppletion. 6 Among the 35 possible stem formatives, 11 possible final vowels and 15 possible tone patterns, a given lexeme's input form only needs a blend of a small subset of each of these exponents.", |
|
"cite_spans": [ |
|
{ |
|
"start": 934, |
|
"end": 935, |
|
"text": "6", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 640, |
|
"end": 648, |
|
"text": "Table 11", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Learning inputs in Mazatec", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Consider first the stem formatives, shown in table 12 (Data from Jamieson (1982) .) 7 The proposed representation for each base input is simply a blend of all its possible initial consonants. The vowels stay the same across a class. Which consonant surfaces depends on how this blend interacts through coalescence of identical consonants with an inflectional input blend of consonant features.", |
|
"cite_spans": [ |
|
{ |
|
"start": 65, |
|
"end": 80, |
|
"text": "Jamieson (1982)", |
|
"ref_id": "BIBREF13" |
|
}, |
|
{ |
|
"start": 84, |
|
"end": 85, |
|
"text": "7", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for stem formatives", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Tables 13 and 14 show the results of the same kind of algorithm discussed on page 3 for learning input activations of inflectional person-number combinations and base inputs respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for stem formatives", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "It took 22 iterations to learn activations that derive the correct stem formative consonants for each of 4 person-number groups in 18 classes.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for stem formatives", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Abstracting away for now from tone and final For each candidate, an inflectional input can coalesce with a base input. The segment with the highest combined activation surfaces, namely k w , with an underlying activation of 0.3+0.12 = 0.42. We assume the stem consonant and the vowel of the stem formative to both have full 1.0 underlying activations; thus no Harmony is gained by having a gradient inflectional stem formative consonant incorrectly coalesce with the stem consonant nor for a gradient inflectional final vowel to incorrectly coalesce with the stem formative vowel.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for stem formatives", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Tone patterns, which fall into 11 descriptive classes, were also learned by an algorithm that separately learned patterns for the first and second syllables. Table 15 shows tone patterns listed by class and person-number-aspect combination. (Data from Jamieson (1982) .)", |
|
"cite_spans": [ |
|
{ |
|
"start": 252, |
|
"end": 267, |
|
"text": "Jamieson (1982)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 158, |
|
"end": 166, |
|
"text": "Table 15", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "There is some phonological predictability for the tones of the second syllable, where 1 is the highest tone and 4 the lowest:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "\u2022 The initial tone on the second syllable can be no lower than the final tone of the first syllable.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "\u2022 If the initial tone of the second syllable is 3 or 4 and the same as the final tone of the first, there is a further rise of two tone levels:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "e.g. 4 can be followed by 42 but not simple 4; -3 can be followed by 31 but not 3.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "-2 can only be followed by 2 or 24. -1 can only be followed by 1.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "\u2022 We can take the level 4 tone that occurs last in Neutral aspect Class 3def 1s 1in 2s 1ex 2p A 3-1 3-1 3-31 3-1 3-14 3-1 B1-1 2-2 1-1 2-2 2-2 2-24 2-2 B3-1 3-2 3-1 2-2 2-2 2-24 2-2 C 3-24 14-3 14-42 14-3 14-34 14-3 D1-1 1-1 1-1 3-2 3-2 3-24 3-2 D3-1 3-2 3-1 3-2 3-2 3-24 3-2 Incompletive aspect A(1-7) 4-2 3-1 3-31 3-1 3-14 3-1 A(8-18) 4-2 3-1 4-31 4-1 4-14 4- 1 B1-1 (1-7) 4-2 1-1 2-2 2-2 2-24 2-2 B1-1 (8-13) 4-2 1-1 4-42 4-2 4-24 4-2 B1-1 (14-18) 4-1 1-1 4-42 4-3 4-34 4- 3 B3-1 (1-7) 4-2 3-1 2-2 2-2 2-24 2-2 B3-1 (8-13) 4-2 3-1 4-42 4-2 4-24 4-2 B3-1 (14-18) 4-1 3-1 4-42 4-3 4-34 4-3 C 3-24 14-3 14-42 14-3 14-34 14-3 D1-1 (8-13) 4-2 1-1 4-42 4-2 4-24 4-2 Tables 16 and 17 show the results of learning algorithms for first and second syllable tones. Ten iterations were required for syllable 1 and 22 iterations for syllable 2, with a stepsize of 0.05 for syllable 1 and 0.1 for syllable 2. Activations were initialized at 0.2 for each base-input and tonal person-number input for each tone pattern that actually occurs. We assume there exists a Harmonic Grammar analysis of the above constraints, in which they have higher weight than Faithfulness, which will otherwise derive the pattern with the highest aggregate activation.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 51, |
|
"end": 347, |
|
"text": "Neutral aspect Class 3def 1s 1in 2s 1ex 2p A 3-1 3-1 3-31 3-1 3-14 3-1 B1-1 2-2 1-1 2-2 2-2 2-24 2-2 B3-1 3-2 3-1 2-2 2-2 2-24 2-2 C 3-24 14-3 14-42 14-3 14-34 14-3 D1-1 1-1 1-1 3-2 3-2 3-24 3-2 D3-1 3-2 3-1 3-2 3-2 3-24 3-2 Incompletive aspect A(1-7)", |
|
"ref_id": "TABREF0" |
|
}, |
|
{ |
|
"start": 407, |
|
"end": 420, |
|
"text": "1 B1-1 (1-7)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 522, |
|
"end": 535, |
|
"text": "3 B3-1 (1-7)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 710, |
|
"end": 726, |
|
"text": "Tables 16 and 17", |
|
"ref_id": "TABREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "A two-syllable base input has a tone linked to the mora of each syllable. The inflectional input has a linearly ordered sequence of two tones on the tonal tier. 8 A strongly-weighted anchoring constraint will require the tones on the inflectional input to line up with the tones on the base input when they coalesce.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Consider the second person singular incompletive of 'remember',\u010da 4 se 2 . This is tonal class B-3-1 (8-13) with inputs shown in table 18. When the two input sources coalesce on each syllable, tone level 4 has the highest activation for syllable 1 of 0.25 + 0.1 = 0.35 and tone level 2 on syllable 2 of 0.4 + 0.0 = 0.4., both the correct tones. [\u2212hi, \u2212lo, \u2212bk, 0 nas] = e, the correct final vowel, is the output after quantization.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Learning inputs for tone patterns", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "To test how a learner could predict an unseen inflectional form for a given stem, we ran crossvaidation on Mazatec stem formatives, tone patterns and final vowels, with training set items (70% of the total stemclass\u00d7inflection combinations) picked randomly from a Zipf-Mandelbrot distribution. On ten runs for each, the average test accuracy was 83% for stem formatives, 87% on tone 1, 92% on tone 2 and 89% on the final vowel. As a baseline, we simultaneously tested prediction of unseen forms based on frequency of occurring stem formatives, for which the accuracy was 9.5%. The relative success of the model in cross-validation is due to the syncretism that occurs both within paradigms and across classes. The following tableau shows how the correct stem formative k wh is predicted for a never-encountered 3rd-definite-incompletive form of a class 7 stem 'weave' aPy 9 in the holdout set, using activation values that were learned for encountered forms. 10 The strength of this model is that it directly encodes knowledge of the tendency of an exponent to occur for a given stem and for a given morphosyntactic combination. For example, a Mazatec class 1 stem with a stem-formative blend (0.2\u2022b, 0.3\u2022k w ) encodes the relative tendencies of these two exponents to occur with this stem, with zero-valued exponents having no inclination to occur. In Kann and Sch\u00fctze (2016) Table 22 : Tableau for 'weave' class 7 3rd-def-inc.", |
|
"cite_spans": [ |
|
{ |
|
"start": 959, |
|
"end": 961, |
|
"text": "10", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1353, |
|
"end": 1376, |
|
"text": "Kann and Sch\u00fctze (2016)", |
|
"ref_id": "BIBREF14" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 1377, |
|
"end": 1385, |
|
"text": "Table 22", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Testing and comparison with other models", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "(2018), there is no indication of what kind of knowledge is represented by a given node or connection. Goldsmith and O'Brien (2006) 11 is more similar to the present model but it is not clear how weights from input to hidden-layer nodes translate into feature-values on a symbolic level. And its winner-take-all mechanism with no threshold to surpass means than zero exponents in a paradigm must be represented as such, rather than resulting from any exponent failing to surpass a threshold. This results in awkward representations when a form is expressed by multiple zero affixes. The advantage of the GSC model, (see Smolensky et al. (2014 Smolensky et al. ( , p. 1103 ) is that it combines a subsymbolic neural level with a symbolic level, i.e. 'microlevel representations and algorithms' with symbolic grammatical theory, with an interface between the two. The subsymbolic level provides a platform for optimization and allows gradient activations. When output activations are quantized to discrete values they percolate to the symbolic level which contains symbolic descriptions that are familiar from symbol-based linguistic theory.", |
|
"cite_spans": [ |
|
{ |
|
"start": 103, |
|
"end": 131, |
|
"text": "Goldsmith and O'Brien (2006)", |
|
"ref_id": "BIBREF8" |
|
}, |
|
{ |
|
"start": 620, |
|
"end": 642, |
|
"text": "Smolensky et al. (2014", |
|
"ref_id": "BIBREF22" |
|
}, |
|
{ |
|
"start": 643, |
|
"end": 671, |
|
"text": "Smolensky et al. ( , p. 1103", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Testing and comparison with other models", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Inflectional patterns in Russian and Mazatec depart from canonicity in a number of ways, as was outlined on page 1. If these systems were completely canonical, each morphosyntactic combination of feature values would have a unique, lexeme-invariant exponent. In our model, in Russian, the variance of an exponent across inflectional classes is expressed directly by a blend of segments in an underlying form, e.g. (a, i) for genitive singular. Syncretism across inflectional feature combinations, such as multiple instances of -i in several classes is expressed by its occurrence as a base input, so that it can coalesce with a matching inflectional input 12 for several feature combinations. If a different exponent occurred in each cell and there were no inflectional class divisions, the URs could simply be a single non-blended exponent for each inflectional input, with no inflectional formatives on base inputs. So the representations proposed here directly encode and capture the kinds of departures from canonicity that occur. In Mazatec, blended inputs capture not just deviation from paradigm canonicity but patterns of predicability within its paradigm structure. Out of 12 35 possible stem-formative combinations that could occur in 12 paradigm positions, Jamieson only lists 18 classes. Blended input representations, in which no more than four phonemes occur for a base input or inflectional input, limit how many stem formatives occur for a given lexeme and along with the occurrence of the same blended inflectional inputs in multiple personnumber combinations, derive the syncretism that occurs across those paradigm positions.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Summary", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "This approach can also derive subtractive morphology, suppletion, umlaut and moraic augmentation. For example, if an inflectional form is subtractive relative to a base, the subtractable part can have partial or negative-valued underlying activation in the UR and express morphosyntactic features that affect its surfacing. Umlaut (Trommer, 2017) can be derived from a blend of different vowel features where alternation occurs.", |
|
"cite_spans": [ |
|
{ |
|
"start": 331, |
|
"end": 346, |
|
"text": "(Trommer, 2017)", |
|
"ref_id": "BIBREF24" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Summary", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "Given the vast number of ways that languages can deviate from canonical inflection (Corbett, 2009 (Corbett, , 2007 Baerman and Corbett, 2010; Stump, 2016) , we hope with future research to explore in the GSC framework how it is possible to learn complex paradigmatic patterns in other languages.", |
|
"cite_spans": [ |
|
{ |
|
"start": 83, |
|
"end": 97, |
|
"text": "(Corbett, 2009", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 98, |
|
"end": 114, |
|
"text": "(Corbett, , 2007", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 115, |
|
"end": 141, |
|
"text": "Baerman and Corbett, 2010;", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 142, |
|
"end": 154, |
|
"text": "Stump, 2016)", |
|
"ref_id": "BIBREF23" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Summary", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "Thanks to Paul Smolensky, Matt Goldrick, Farrell Ackerman, three anonymous reviewers and members of the Johns Hopkins Neurosymbolic Computation Group and the Surrey Morphology Group for valuable discussion and suggestions. Research was generously funded by NSF INSPIRE grant BCS-1344269. All errors are my own.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgements", |
|
"sec_num": "6" |
|
}, |
|
{ |
|
"text": "In response to a reviewer's question about the effects and degree of simplification of the paradigms of these two languages, we contend that (a) any linguistic analysis will have some degree of simplification, (b) the paradigms are no more simplified than those analysed by A&M and in fact include more paradigm positions than are represented in their table A6 (p. 457), and (c) nothing in the analysis was excluded because the model couldn't handle it.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "A reviewer suggests that the contribution of this model is theoretical rather than computational and that the learning algorithm is 'fairly trivial.' We see no advantage in a model of human learning that is intentionally complex; the focus of this proposal is to contribute to computationally-based theories of human language rather than to test new computational techniques.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "If a stem formative is absent from the base input, an anchoring constraint will prevent an inflection-based stemformative input from surfacing.6 Stem suppletion can be accounted for by an input that includes a blend of different stem allomorphs each linked to different inflectional features.7 FollowingGolston and Kehrein (1998), we take what Jamieson transcribes as complex onsets and nuclei to be simplex segments with secondary articulations.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In the 1st person exclusive, a level 4 tone will occur further to the right of other tones.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Since the vowel a of the stem formative does not vary, for convenience we can take it here to be part of the stem.10 Because the activation values were learned just for a random training set, they will be different from those given in tables 13 and 14 above.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "See alsoGoldsmith and Rosen (2017), which has similarities with the present model but differs in crucial ways.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In terms of correspondence, this model employs the converse of whatPulleyblank (2008) proposes for his account of reduplication in which a single input has mutiple outputs. Here, pairs of identical input features or segments can each coalesce on a single output feature or segment.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Morphological Organization: the Low Conditional Entropy Conjecture", |
|
"authors": [ |
|
{ |
|
"first": "Farrell", |
|
"middle": [], |
|
"last": "Ackerman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Robert", |
|
"middle": [], |
|
"last": "Malouf", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "Language", |
|
"volume": "89", |
|
"issue": "", |
|
"pages": "429--464", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ackerman, Farrell, and Robert Malouf. 2013. Morpho- logical Organization: the Low Conditional Entropy Conjecture. Language 89:429-464.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "A typology of inflectional class interaction", |
|
"authors": [ |
|
{ |
|
"first": "Matthew", |
|
"middle": [], |
|
"last": "Baerman", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Greville", |
|
"middle": [], |
|
"last": "Corbett", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "14th International Morphology Meeting", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Baerman, Matthew, and Greville Corbett. 2010. A typology of inflectional class interaction. In 14th International Morphology Meeting. Budapest, Hun- gary.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Incremental parsing in a continuous dynamical system: sentence processing in Gradient Symbolic Computation: Supplementary Materials 3: processing", |
|
"authors": [ |
|
{ |
|
"first": "Pyeong", |
|
"middle": [], |
|
"last": "Cho", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Matthew", |
|
"middle": [], |
|
"last": "Whan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Goldrick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Smolensky", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Cho, Pyeong Whan, Matthew Goldrick, and Paul Smolensky. 2017. Incremental parsing in a continu- ous dynamical system: sentence processing in Gra- dient Symbolic Computation: Supplementary Mate- rials 3: processing. Linguistics Vanguard 3.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Canonical typology, suppletion and possible words", |
|
"authors": [ |
|
{ |
|
"first": "Greville", |
|
"middle": [], |
|
"last": "Corbett", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Language", |
|
"volume": "83", |
|
"issue": "", |
|
"pages": "8--42", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Corbett, Greville. 2007. Canonical typology, supple- tion and possible words. Language 83:8-42.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Canonical inflectional classes", |
|
"authors": [ |
|
{ |
|
"first": "Greville", |
|
"middle": [], |
|
"last": "Corbett", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Selected Proceedings of the 6th D\u00e9cembrettes: Morphology in", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--11", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Corbett, Greville. 2009. Canonical inflectional classes. In Selected Proceedings of the 6th D\u00e9cembrettes: Morphology in Bordeaux, 1-11.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "How much for that vowel? Talk, Strength in Grammar Workshop", |
|
"authors": [ |
|
{ |
|
"first": "Noam", |
|
"middle": [], |
|
"last": "Faust", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Faust, Noam. 2017. How much for that vowel? Talk, Strength in Grammar Workshop, University of Leipzig.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Activity as an alternative to autosegmental association", |
|
"authors": [ |
|
{ |
|
"first": "Noam", |
|
"middle": [], |
|
"last": "Faust", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Smolensky", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Manchester Phonology Meeting", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Faust, Noam, and Paul Smolensky. 2017. Activ- ity as an alternative to autosegmental association. In Manchester Phonology Meeting. Manchester, United Kingdom.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Coactivation in bilingual grammars: A computational account of code mixing", |
|
"authors": [ |
|
{ |
|
"first": "Matthew", |
|
"middle": [], |
|
"last": "Goldrick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Putnam", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lara", |
|
"middle": [], |
|
"last": "Schwarz", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Bilingualism: Language and Cognition", |
|
"volume": "19", |
|
"issue": "", |
|
"pages": "857--876", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Goldrick, Matthew, Michael Putnam, and Lara Schwarz. 2016. Coactivation in bilingual grammars: A computational account of code mixing. Bilingual- ism: Language and Cognition 19:857-876. ROA 1441.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Learning inflectional classes", |
|
"authors": [ |
|
{ |
|
"first": "John", |
|
"middle": [], |
|
"last": "Goldsmith", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jeremy O'", |
|
"middle": [], |
|
"last": "Brien", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Language Learning and Development", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "219--250", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Goldsmith, John, and Jeremy O'Brien. 2006. Learning inflectional classes. Language Learning and Devel- opment 2:219-250.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Geometrical morphology", |
|
"authors": [ |
|
{ |
|
"first": "John", |
|
"middle": [], |
|
"last": "Goldsmith", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Eric", |
|
"middle": [], |
|
"last": "Rosen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Goldsmith, John, and Eric Rosen. 2017. Geomet- rical morphology. CoRR abs/1703.04481. URL http://arxiv.org/abs/1703.04481.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Mazatec Onsets and Nuclei", |
|
"authors": [ |
|
{ |
|
"first": "Chris", |
|
"middle": [], |
|
"last": "Golston", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wolfgang", |
|
"middle": [], |
|
"last": "Kehrein", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1998, |
|
"venue": "International Journal of American Linguistics", |
|
"volume": "64", |
|
"issue": "", |
|
"pages": "311--337", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Golston, Chris, and Wolfgang Kehrein. 1998. Mazatec Onsets and Nuclei. International Journal of Ameri- can Linguistics 64:311-337.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Testing tolerance for lexically-specific factors in gradient symbolic computation", |
|
"authors": [ |
|
{ |
|
"first": "Janet", |
|
"middle": [ |
|
"G" |
|
], |
|
"last": "Van Hell", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Clara", |
|
"middle": [], |
|
"last": "Cohen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sarah", |
|
"middle": [], |
|
"last": "Grey", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Bilingualism: Language and Cognition", |
|
"volume": "19", |
|
"issue": "", |
|
"pages": "897--899", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "van Hell, Janet G., Clara Cohen, and Sarah Grey. 2016. Testing tolerance for lexically-specific fac- tors in gradient symbolic computation. Bilingual- ism: Language and Cognition 19:897-899.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Scalar constraints and gradient symbolic representations generate exceptional prosodification effects without exceptional prosody", |
|
"authors": [ |
|
{ |
|
"first": "Brian", |
|
"middle": [], |
|
"last": "Hsu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "West Coast Conference on Formal Linguistics", |
|
"volume": "36", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hsu, Brian. 2018. Scalar constraints and gradient sym- bolic representations generate exceptional prosodi- fication effects without exceptional prosody. Hand- out, West Coast Conference on Formal Linguistics 36.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Conflated subsystems marking person and aspect in Chiquihuitl\u00e1n Mazatec verbs", |
|
"authors": [ |
|
{ |
|
"first": "Carol", |
|
"middle": [], |
|
"last": "Jamieson", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1982, |
|
"venue": "Language", |
|
"volume": "48", |
|
"issue": "", |
|
"pages": "139--167", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jamieson, Carol. 1982. Conflated subsystems marking person and aspect in Chiquihuitl\u00e1n Mazatec verbs. Language 48:139-167.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Single-Model Encoder-Decoder with Explicit Morphological Representation for Reinflection", |
|
"authors": [ |
|
{ |
|
"first": "Katharina", |
|
"middle": [], |
|
"last": "Kann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hinrich", |
|
"middle": [], |
|
"last": "Sch\u00fctze", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "555--560", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kann, Katharina, and Hinrich Sch\u00fctze. 2016. Single- Model Encoder-Decoder with Explicit Morpholog- ical Representation for Reinflection. In Proceed- ings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Pa- pers), 555-560. Association for Computational Lin- guistics.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Generating morphological paradigms with a recurrent neural network. San Diego Linguistic Papers 6", |
|
"authors": [ |
|
{ |
|
"first": "Robert", |
|
"middle": [], |
|
"last": "Malouf", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Malouf, Robert. 2018. Generating morphological paradigms with a recurrent neural network. San Diego Linguistic Papers 6.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "Gradient symbolic representations in syntax. Handout, IGRA 02: Syntax II", |
|
"authors": [ |
|
{ |
|
"first": "Gereon", |
|
"middle": [], |
|
"last": "M\u00fcller", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "M\u00fcller, Gereon. 2017. Gradient symbolic representa- tions in syntax. Handout, IGRA 02: Syntax II.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Optimality Theory: Constraint Interaction in Generative Grammar", |
|
"authors": [ |
|
{ |
|
"first": "Alan", |
|
"middle": [], |
|
"last": "Prince", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Smolensky", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1993, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Prince, Alan, and Paul Smolensky. 1993. Optimality Theory: Constraint Interaction in Generative Gram- mar. Technical report, Rutgers University Center for Cognitive Science and Computer Science Depart- ment, University of Colorado at Boulder.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "The nature of the word: Studies in honor of paul kiparsky, chapter Patterns of Reduplication in Yoruba", |
|
"authors": [ |
|
{ |
|
"first": "Douglas", |
|
"middle": [], |
|
"last": "Pulleyblank", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Pulleyblank, Douglas. 2008. The nature of the word: Studies in honor of paul kiparsky, chapter Patterns of Reduplication in Yoruba. MIT Press.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Predicting the unpredictable: Capturing the apparent semi-regularity of rendaku voicing in japanese through gradient symbolic computation", |
|
"authors": [ |
|
{ |
|
"first": "Eric", |
|
"middle": [], |
|
"last": "Rosen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the Berkeley Linguistics Society", |
|
"volume": "42", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rosen, Eric. 2016. Predicting the unpredictable: Cap- turing the apparent semi-regularity of rendaku voic- ing in japanese through gradient symbolic compu- tation. In Proceedings of the Berkeley Linguistics Society, volume 42. ROA 1299.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Predicting semi-regular patterns in morphologically complex words", |
|
"authors": [ |
|
{ |
|
"first": "Eric", |
|
"middle": [], |
|
"last": "Rosen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rosen, Eric. 2018. Predicting semi-regular patterns in morphologically complex words. Linguistics Van- guard 4. ROA 1339.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Gradient Symbolic Representations in Grammar: The case of French Liaison. Rutgers Optimality Archive 1552", |
|
"authors": [ |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Smolensky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Matt", |
|
"middle": [], |
|
"last": "Goldrick", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Smolensky, Paul, and Matt Goldrick. 2016. Gradient Symbolic Representations in Grammar: The case of French Liaison. Rutgers Optimality Archive 1552, Rutgers University.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Optimization and Quantization in Gradient Symbol Systems: A Framework for Integrating the Continuous and the Discrete in Cognition", |
|
"authors": [ |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Smolensky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Matthew", |
|
"middle": [], |
|
"last": "Goldrick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Donald", |
|
"middle": [], |
|
"last": "Mathis", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Cognitive Science", |
|
"volume": "38", |
|
"issue": "", |
|
"pages": "1102--1138", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Smolensky, Paul, Matthew Goldrick, and Donald Mathis. 2014. Optimization and Quantization in Gradient Symbol Systems: A Framework for Inte- grating the Continuous and the Discrete in Cogni- tion. Cognitive Science 38:1102-1138.", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "Inflectional paradigms: Content and from at the syntax-morphology interface", |
|
"authors": [ |
|
{ |
|
"first": "Gregory", |
|
"middle": [], |
|
"last": "Stump", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Cambridge Studies in Linguistics", |
|
"volume": "149", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Stump, Gregory. 2016. Inflectional paradigms: Con- tent and from at the syntax-morphology interface, volume 149 of Cambridge Studies in Linguistics. Cambridge, UK: Cambridge University Press.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Scalar Cumulativity in German Umlaut. Talk, Strength in Grammar Workshop", |
|
"authors": [ |
|
{ |
|
"first": "Jochen", |
|
"middle": [], |
|
"last": "Trommer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Trommer, Jochen. 2017. Scalar Cumulativity in Ger- man Umlaut. Talk, Strength in Grammar Workshop, University of Leipzig.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "Being (slightly) stronger: Lexical stress in Moses Columbian Salish. Handout, Strength in Grammar Workshop", |
|
"authors": [ |
|
{ |
|
"first": "Eva", |
|
"middle": [], |
|
"last": "Zimmermann", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zimmermann, Eva. 2017a. Being (slightly) stronger: Lexical stress in Moses Columbian Salish. Hand- out, Strength in Grammar Workshop, University of Leipzig.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Gradient symbols and gradient markedness: A case study from Mixtec tones", |
|
"authors": [ |
|
{ |
|
"first": "Eva", |
|
"middle": [], |
|
"last": "Zimmermann", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Handout: Manchester Phonology Meeting", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zimmermann, Eva. 2017b. Gradient symbols and gra- dient markedness: A case study from Mixtec tones. Handout: Manchester Phonology Meeting.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Gradient symbolic representations and the typology of ghost segments: An argument from gradient markedness. Handout and slides", |
|
"authors": [ |
|
{ |
|
"first": "Eva", |
|
"middle": [], |
|
"last": "Zimmermann", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Annual Meeting on Phonology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zimmermann, Eva. 2018. Gradient symbolic rep- resentations and the typology of ghost segments: An argument from gradient markedness. Handout and slides, Annual Meeting on Phonology, UC San Diego.", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "Gradient Symbolic Representations in the output: A case study from Moses Columbian Salishan stress", |
|
"authors": [ |
|
{ |
|
"first": "Eva", |
|
"middle": [], |
|
"last": "Zimmermann", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Forthcoming", |
|
"suffix": "" |
|
} |
|
], |
|
"year": null, |
|
"venue": "Proceedings of NELS", |
|
"volume": "48", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zimmermann, Eva. forthcoming. Gradient Symbolic Representations in the output: A case study from Moses Columbian Salishan stress. In Proceedings of NELS, volume 48.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF0": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "" |
|
}, |
|
"TABREF1": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Russian noun paradigm" |
|
}, |
|
"TABREF3": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td colspan=\"3\">vin(0.1 \u2022 a, 0.02 \u2022 u, 0.16 \u2022 o) 'wine' + GEN-SG(0.3 \u2022 a, 0.26 \u2022 i) MAX DEP Harmony</td></tr><tr><td>5 vin-a 2.0 vin-u 0.1 vin-o 0.8 vin-i 1.3 vin-e vin</td><td>\u22122 \u22121.2 \u22121.96 \u22121.68 \u22121.48 \u22122.0</td><td>0.8 \u22121.86 \u22120.88 \u22120.18 \u22122.0 0.0</td></tr></table>", |
|
"text": "Summed activations for class 1 GEN-SG" |
|
}, |
|
"TABREF4": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td>: Tableau for class 1 GEN-SG</td></tr><tr><td>tion that is the sum of the two source activa-</td></tr><tr><td>tions. Consider again the same genitive sin-</td></tr><tr><td>gular form vina given above in table 1. Sup-</td></tr></table>", |
|
"text": "" |
|
}, |
|
"TABREF5": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "" |
|
}, |
|
"TABREF7": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td colspan=\"4\">Cell NSG ASG GSG DSG LSG ISG</td></tr><tr><td>a</td><td>0.28</td><td>0.3</td></tr><tr><td>e</td><td/><td/><td>0.26 0.3</td></tr><tr><td>u</td><td/><td>0.28</td><td>0.3</td></tr><tr><td>om</td><td/><td/><td>0.3</td></tr><tr><td>ov</td><td/><td/></tr><tr><td>ax</td><td/><td/></tr><tr><td>am</td><td/><td/></tr><tr><td>am'i</td><td/><td/></tr><tr><td>i</td><td/><td colspan=\"2\">0.26 0.22 0.24</td></tr><tr><td>oj</td><td/><td/><td>0.26</td></tr><tr><td>ju</td><td/><td/><td>0.26</td></tr><tr><td>o</td><td colspan=\"2\">0.24 0.22</td></tr><tr><td>ej</td><td/><td/></tr><tr><td colspan=\"4\">Cell NPL APL GPL DPL LPL IPL</td></tr><tr><td>a</td><td colspan=\"2\">0.24 0.24</td></tr><tr><td>e</td><td/><td/></tr><tr><td>u</td><td/><td/></tr><tr><td>om</td><td/><td/></tr><tr><td>ov</td><td/><td>0.4</td></tr><tr><td>ax</td><td/><td/><td>0.3</td></tr><tr><td>am</td><td/><td/><td>0.3</td></tr><tr><td>am'i</td><td/><td/><td>0.3</td></tr><tr><td>i</td><td>0.3</td><td>0.3</td></tr><tr><td>oj</td><td/><td/></tr><tr><td>ju</td><td/><td/></tr><tr><td>o</td><td/><td/></tr><tr><td>ej</td><td/><td>0.22</td></tr></table>", |
|
"text": "Learned nonzero input values" |
|
}, |
|
"TABREF8": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td/><td>Real</td><td/><td>Impossible</td></tr><tr><td colspan=\"2\">Class 3 Class 4</td><td colspan=\"2\">Class 3 Class 4</td></tr><tr><td>Gen. Sg. i</td><td>i</td><td>i</td><td>e</td></tr><tr><td>Dat. Sg. e</td><td>i</td><td>e</td><td>i</td></tr></table>", |
|
"text": "Learned inflectional input values2.1 Restricting possible paradigmsThe Russian noun paradigm exhibits both syncretism within descriptive inflectional classes and variance of exponents for a given position across inflectional classes. Our model predicts that there are limits to such departures from canonicity. In the dative singular, the affix is e in class 3 and i in class 4. In the genitive singular, it is i for both classes, making i syncretic across those two feature combinations for class 4. Our model predicts the impossibility of the pattern in the right-hand half of table 8." |
|
}, |
|
"TABREF9": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Crossing diagonals" |
|
}, |
|
"TABREF10": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Contradictory inequalities" |
|
}, |
|
"TABREF11": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td/><td>NEUTRAL</td><td colspan=\"2\">INCOMPLETIVE</td></tr><tr><td>SG</td><td>PL</td><td>SG</td><td>PL</td></tr><tr><td colspan=\"2\">1INCL\u010da</td><td/><td/></tr></table>", |
|
"text": ", with 35 possible exponents, (b) a final vowel(table 19), with 11 possible exponents, and (c) a tone pattern(table 15), with 15 possible exponents. For example, in the neutral 1st person singular form in table 10, ba 3 sae 1 , the stem is merely the segment s. The stem formative is ba, the final vowel is ae and the tone pattern is 3-1. Yet these 35 \u00d7 11 \u00d7 15 = 5775 hypothetically" |
|
}, |
|
"TABREF12": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Paradigm for 'remember'" |
|
}, |
|
"TABREF14": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td>Some further complexities in Jamieson's data</td></tr><tr><td>are omitted here for simplicity of exposition but</td></tr></table>", |
|
"text": "" |
|
}, |
|
"TABREF15": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Representations of stem formatives" |
|
}, |
|
"TABREF18": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td>: Tone patterns</td></tr><tr><td>the 1st exclusive to be a separate tonal input for</td></tr><tr><td>1st-exclusive.</td></tr></table>", |
|
"text": "" |
|
}, |
|
"TABREF19": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td/><td colspan=\"3\">First syllable tones</td><td/><td/></tr><tr><td>Tone class</td><td>1</td><td>2</td><td>3</td><td>4</td><td>14</td></tr><tr><td>A1-7</td><td/><td/><td colspan=\"2\">0.25 0.15</td><td/></tr><tr><td>A8-18</td><td/><td/><td>0.2</td><td>0.25</td><td/></tr><tr><td>B1-1-1-7</td><td colspan=\"2\">0.15 0.25</td><td/><td>0.05</td><td/></tr><tr><td>B1-1-18-13</td><td>0.2</td><td>0.25</td><td/><td>0.2</td><td/></tr><tr><td>B1-1-14-18</td><td colspan=\"2\">0.15 0.25</td><td/><td>0.2</td><td/></tr><tr><td>B3-1-1-7</td><td/><td>0.3</td><td>0.2</td><td>0.1</td><td/></tr><tr><td>B3-1-8-13</td><td/><td colspan=\"3\">0.25 0.15 0.25</td><td/></tr><tr><td>B3-1-14-18</td><td/><td colspan=\"2\">0.25 0.1</td><td>0.25</td><td/></tr><tr><td>C</td><td/><td/><td>0.2</td><td/><td>0.35</td></tr><tr><td>D1-1</td><td>0.2</td><td/><td colspan=\"2\">0.15 0.25</td><td/></tr><tr><td>D3-1</td><td>0.15</td><td/><td>0.2</td><td>0.25</td><td/></tr><tr><td colspan=\"2\">Person/number 1</td><td>2</td><td>3</td><td>4</td><td>14</td></tr><tr><td>3def Neu.</td><td/><td colspan=\"2\">0.05 0.25</td><td/><td/></tr><tr><td>1s Neu.</td><td>0.15</td><td/><td>0.15</td><td/><td>0.1</td></tr><tr><td>1in Neu.</td><td/><td>0.2</td><td>0.15</td><td/><td>0.05</td></tr><tr><td>2s Neu.</td><td/><td colspan=\"2\">0.15 0.15</td><td/><td>0.05</td></tr><tr><td>1ex Neu.</td><td/><td>0.2</td><td>0.15</td><td/><td>0.05</td></tr><tr><td>2p Neu.</td><td/><td>0.1</td><td>0.1</td><td/><td/></tr><tr><td>3def Inc.</td><td/><td/><td colspan=\"2\">0.15 0.35</td><td/></tr><tr><td>1s Inc.</td><td>0.25</td><td/><td>0.15</td><td/><td>0.1</td></tr><tr><td>1in Inc.</td><td/><td colspan=\"3\">0.05 0.05 0.1</td><td>0.05</td></tr><tr><td>2s Inc.</td><td/><td colspan=\"2\">0.05 0.1</td><td>0.1</td><td/></tr><tr><td>1ex Inc.</td><td/><td colspan=\"3\">0.05 0.05 0.1</td><td/></tr><tr><td>2p Inc.</td><td/><td colspan=\"3\">0.05 0.05 0.1</td><td/></tr></table>", |
|
"text": "with data fromJamieson (1982) shows the 10 descriptive classes of final vowels." |
|
}, |
|
"TABREF20": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td colspan=\"4\">Second syllable tones</td><td/><td/></tr><tr><td>Tone class</td><td>1</td><td>2</td><td>3</td><td>4</td><td>24</td></tr><tr><td>A1-7</td><td colspan=\"3\">0.2 0.2 -0.1</td><td/><td/></tr><tr><td>A8-18</td><td>0.5</td><td/><td colspan=\"2\">0.2 0.2</td><td/></tr><tr><td>B1-1-1-7</td><td colspan=\"2\">0.2 0.2</td><td/><td/><td/></tr><tr><td>B1-1-18-13</td><td/><td>0.4</td><td/><td>0.2</td><td/></tr><tr><td>B1-1-14-18</td><td>0.1</td><td/><td colspan=\"2\">0.6 0.4</td><td/></tr><tr><td>B3-1-1-7</td><td colspan=\"2\">0.2 0.2</td><td/><td/><td/></tr><tr><td>B3-1-8-13</td><td/><td>0.4</td><td/><td>0.2</td><td/></tr><tr><td>B3-1-14-18</td><td>0.5</td><td/><td colspan=\"2\">0.4 0.2</td><td/></tr><tr><td>C</td><td/><td/><td colspan=\"3\">0.3 0.3 0.2</td></tr><tr><td>D1-1</td><td/><td>0.5</td><td/><td>0.2</td><td/></tr><tr><td>D3-1</td><td colspan=\"2\">0.1 0.5</td><td/><td>0.2</td><td/></tr><tr><td>Person/number</td><td/><td>2</td><td>3</td><td>4</td><td>24</td></tr><tr><td>3def Neu.</td><td colspan=\"2\">0.2 0.4</td><td/><td/><td>0.2</td></tr><tr><td>1s Neu.</td><td>0.5</td><td/><td>0.2</td><td/><td/></tr><tr><td>1in Neu.</td><td/><td colspan=\"3\">0.1 0.5 0.8</td><td/></tr><tr><td>2s Neu.</td><td colspan=\"3\">0.2 0.1 0.2</td><td/><td/></tr><tr><td>1ex Neu.</td><td colspan=\"3\">0.2 0.1 0.2</td><td/><td/></tr><tr><td>2p Neu.</td><td colspan=\"3\">0.2 0.1 0.2</td><td/><td/></tr><tr><td>3def Inc.</td><td colspan=\"3\">0.2 0.3 0.1</td><td/><td>0.2</td></tr><tr><td>1s Inc.</td><td>0.5</td><td/><td>0.3</td><td/><td/></tr><tr><td>1in Inc.</td><td/><td/><td colspan=\"2\">0.3 0.7</td><td/></tr><tr><td>2s Inc.</td><td>0.1</td><td/><td>0.3</td><td/><td/></tr><tr><td>1ex Inc.</td><td>0.1</td><td/><td>0.2</td><td/><td/></tr><tr><td>2p Inc.</td><td>0.1</td><td/><td>0.2</td><td/><td/></tr></table>", |
|
"text": "Input activations for \u03c3 1 tones" |
|
}, |
|
"TABREF21": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td/><td/><td colspan=\"2\">Syllable 1</td><td/></tr><tr><td>Tone \u2192 Tonal class B-31 (8-13)</td><td>1</td><td colspan=\"3\">2 0.25 0.15 0.25 3 4</td></tr><tr><td>2nd sg. incomp.</td><td/><td colspan=\"2\">0.05 0.1</td><td>0.1</td></tr><tr><td>Total</td><td/><td colspan=\"3\">0.3 Syllable 2 0.25 0.35</td></tr><tr><td>Tone \u2192 Tonal class B-31 (8-13)</td><td>1</td><td>2 0.4</td><td>3</td><td>4 0.2</td></tr><tr><td>2nd sg. incomp.</td><td>0.1</td><td/><td>0.3</td><td/></tr><tr><td>Total</td><td colspan=\"2\">0.1 0.4</td><td>0.3</td><td>0.2</td></tr></table>", |
|
"text": "Input activations for \u03c3 2 tonesThe algorithm learns input values, shown in table 20, for \u00b1hi, \u00b1lo, \u00b1bk and nas features, with coalescence of values from the two input sources. A negative value in the table represents a positive value for the negative binary feature: e.g. \u22120.1" |
|
}, |
|
"TABREF22": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td/><td colspan=\"3\">: Aggregate activations</td></tr><tr><td>Class</td><td colspan=\"3\">3def 1s 1in 2s 1ex 2p</td></tr><tr><td/><td>(basic)</td><td/><td/></tr><tr><td>1</td><td colspan=\"2\">i ae\u1ebd</td><td>i\u0129\u0169</td></tr><tr><td>2</td><td colspan=\"2\">e ae\u1ebd</td><td>e\u0129\u0169</td></tr><tr><td>3</td><td colspan=\"2\">ae ae\u1ebd</td><td>e\u0129\u0169</td></tr><tr><td>4</td><td colspan=\"2\">u u\u0169</td><td>i\u0129\u0169</td></tr><tr><td>5</td><td colspan=\"2\">o o\u00f5</td><td>e\u0129\u0169</td></tr><tr><td>6</td><td>a</td><td>a\u00e3</td><td>e\u0129\u0169</td></tr><tr><td colspan=\"4\">7\u0129\u1ebd\u1ebd\u0129\u0129\u0169</td></tr><tr><td colspan=\"4\">8\u1ebd\u1ebd\u1ebd\u0129\u0129\u0169</td></tr><tr><td colspan=\"4\">9\u0169\u0169\u0169\u0129\u0129\u0169</td></tr><tr><td colspan=\"4\">10\u00e3\u00e3\u00e3\u0129\u0129\u0169</td></tr></table>", |
|
"text": "" |
|
}, |
|
"TABREF23": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td/><td>hi</td><td>lo</td><td>bk nas</td></tr><tr><td>3def</td><td>-0.1</td><td/></tr><tr><td>1s</td><td colspan=\"2\">-0.2 0.2</td></tr><tr><td>1in</td><td>-0.2</td><td/><td>0.2</td></tr><tr><td>2s</td><td colspan=\"3\">0.1 -0.3 -0.2</td></tr><tr><td>1ex</td><td colspan=\"3\">0.4 -0.2 -0.2 0.2</td></tr><tr><td>2p</td><td colspan=\"3\">0.4 -0.2 0.3 0.2</td></tr><tr><td/><td>hi</td><td>lo</td><td>bk nas</td></tr><tr><td>Class 1</td><td>0.2</td><td/><td>-0.1</td></tr><tr><td>Class 2</td><td colspan=\"3\">-0.2 -0.1 -0.1</td></tr><tr><td>Class 3</td><td colspan=\"3\">-0.2 0.2 -0.1</td></tr><tr><td>Class 4</td><td colspan=\"3\">0.4 -0.2 0.1</td></tr><tr><td>Class 5</td><td colspan=\"3\">-0.2 -0.3 0.1</td></tr><tr><td>Class 6</td><td colspan=\"3\">-0.2 0.1 0.1</td></tr><tr><td>Class 7</td><td colspan=\"3\">0.2 -0.1 -0.2 0.2</td></tr><tr><td>Class 8</td><td/><td/><td>-0.1 0.2</td></tr><tr><td>Class 9</td><td colspan=\"3\">0.4 -0.2 0.1 0.1</td></tr><tr><td>Class 10</td><td/><td/><td>0.1 0.1</td></tr></table>", |
|
"text": "Classes of final vowels hi means 0.1 \u2212hi. We assume that \u2200i both +f i and \u2212f i do not occur in the same source.Table 21shows input vowel features for\u010da 4 se 2 'remember', final-vowel class 2 in the 2nd sg. incomp. (W v denotes base input.)" |
|
}, |
|
"TABREF24": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td colspan=\"2\">Feature 2nd-sg. W v</td><td>Winner</td></tr><tr><td>hi lo bk nas</td><td colspan=\"2\">0.1 \u22120.2 \u2212hi \u22120.3 \u22120.1 \u2212lo \u22120.2 \u22120.1 \u2212bk 0 nas</td></tr></table>", |
|
"text": "Input activations for vowel featuresWe adopt a second, strongly-weighted quantiza-tion constraint fromCho et al. (2017) that requires the sum of activations from a binary feature group such as {+hi, \u2212hi} to be 1. Thus \u2212hi competes with +hi and \u2212hi surfaces because its higher aggregate activation results in greater Harmony." |
|
}, |
|
"TABREF25": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "Values for final-vowel class 2, 2nd-sg." |
|
} |
|
} |
|
} |
|
} |