|
{ |
|
"paper_id": "C82-1045", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T13:12:57.755258Z" |
|
}, |
|
"title": "A HEURISTIC APPROACH TO ENGLISH-INTO-JAPANESE MACHINE TRANSLATION", |
|
"authors": [ |
|
{ |
|
"first": "Yoshihiko", |
|
"middle": [], |
|
"last": "Nitta", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Hitachi Ltd. Kawasaki", |
|
"location": { |
|
"country": "Kanagawa Japan" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Atsushi", |
|
"middle": [], |
|
"last": "Okajima", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Hitachi Ltd. Kawasaki", |
|
"location": { |
|
"country": "Kanagawa Japan" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Fumiyuki", |
|
"middle": [], |
|
"last": "Yamano", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Hitachi Ltd. Kawasaki", |
|
"location": { |
|
"country": "Kanagawa Japan" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Koichiro", |
|
"middle": [], |
|
"last": "Ishihara", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Hitachi Ltd. Kawasaki", |
|
"location": { |
|
"country": "Kanagawa Japan" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "", |
|
"pdf_parse": { |
|
"paper_id": "C82-1045", |
|
"_pdf_hash": "", |
|
"abstract": [], |
|
"body_text": [ |
|
{ |
|
"text": "Is it true that the recipe to realize a successful machine translation is in precise and rigid language parsing? So far many studies have been done on rigid and detailed natural language parsing, some of which are so powerful as to detect some ungrammatical sentences If, 2, 3, 4] . Notwithstanding it seems that the detailed parsing is not always connected with practically satisfying machine translations. On the other hand actual human, even foreign language learners, can translate fairly difficult English sentences without going into details of parsing. They only use an elementary grammatical knowledge and dictionaries.", |
|
"cite_spans": [ |
|
{ |
|
"start": 275, |
|
"end": 277, |
|
"text": "3,", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 278, |
|
"end": 280, |
|
"text": "4]", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "]. INTRODUCTION", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Thu. we have paid attention on the heuristic methods of language-learners and have dew~ed a rather non-standard linguistic model named HPM (= Heuristic Parsing Model). Here, \"non-standard\" implies that sentential constituents in HPM are different from those in widely accepted modern English grammars [5] or in phrase structure grammars [6] . In order to prove the reasonability of HPM, we have developed an English-into-Japanese translation system named ATHENE (= Automatic T_ranslation of Hitachi from E_nglish into Nihongo with Editing Support)~f. Fig. I ).", |
|
"cite_spans": [ |
|
{ |
|
"start": 301, |
|
"end": 304, |
|
"text": "[5]", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 337, |
|
"end": 340, |
|
"text": "[6]", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 551, |
|
"end": 557, |
|
"text": "Fig. I", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "]. INTRODUCTION", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "The essential features of heuristic translation are summarized as in following three points. (I) To segment an input sentence into new elements named Phrasal Elements (PE) and Clausal Elements (CE),", |
|
"cite_spans": [ |
|
{ |
|
"start": 93, |
|
"end": 96, |
|
"text": "(I)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 167, |
|
"end": 171, |
|
"text": "(PE)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "]. INTRODUCTION", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "To assign syntactic roles to PE's and CE's, and restructure the segmented elements into tree-forms by inclusive relation and into list-forms by modifying relation.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "]. INTRODUCTION", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "(3) To permute the segmented elements, and to assiqn appropriate Japanese equivalents with necessary case suffixes and postpos~tions.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "]. INTRODUCTION", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "The next section presents an overview of HPM, which is followed in Sec. 3 by a rough explication of machine translation process in ATHENE. Sec. 4 2.1 Parsed Tree: A parsed sentence is represented in a \"tree\" or \"list\" of nodes linked by pointers. Each node corresponds to a certain \"constituent of sentence\". \"Tree (~/)\" is for inclusive relation, and \"list ('~)\" is for modifying relation.", |
|
"cite_spans": [ |
|
{ |
|
"start": 144, |
|
"end": 145, |
|
"text": "4", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "]. INTRODUCTION", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "2. 2.9 Syntactic Role (SR): SR is important to represent parsing results and to generate Japanese sentences. For example, the sequencWof SR such as \"SUBJ + GOV + OBJ\" will readily imply the Japanese sentence such as \"SUBJ + {ga/wa/ no} + OBJ + {wo/ni} + GOV\". This implication may be quite natural for language-learners.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "]. INTRODUCTION", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "From the viewpoint of simplicity and maintenability, it might be desirable to describe all the Grammatical Data (GD) in static pattern form. But unfortunately, the pattern form description is lacking in the flexibility to change control structures. Thus we have adopted a combination of \"program\" and \"pattern\" to describe GD.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TRANSLATION PROCESS", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "In the followings, we will describe the translation process along with the examples of grara~atical data (GD) to be referred. The essential point of the translation process is \"to replace some specified node pattern sequences with others, under the appropriate control with grammatical data\". This replacement process is composed of following twelve steps:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TRANSLATION PROCESS", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "(I) Text Input: To produce upper-most node sequence in the parsed tree.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TRANSLATION PROCESS", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "(2) Morphological Resolution: To reduce the inflected word to root form.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TRANSLATION PROCESS", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "(3) Lexicon Retrieval and Attribute Assignment: To assign all possible attributes to \"WE's\". \"tree-transformation\" is reduced to only a simple repetition of permu~tion, which has a strong resemblance ~ language learners' translation methods (Fig. B) . ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 241, |
|
"end": 249, |
|
"text": "(Fig. B)", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "TRANSLATION PROCESS", |
|
"sec_num": "3." |
|
}, |
|
{ |
|
"text": "A pro~pe machine translation system from English in~ Japanese named ATHENE, as is sketched in Fig. l , has been implemented. The lexicons contain nearly.ten thousand words, not counting idioms and other multi-word groups, which are mainly composed of educational basic words (up to senior-high-school-level in Japan) and of about a thousand computer terminologies. Our system has translated a series of t t passages extracted randomly from English readers of senior high school and c~ outer system journals.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 94, |
|
"end": 100, |
|
"text": "Fig. l", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "EXPERIi 4ENTAL RESULTS", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "The results of the tests are encouraging on the whole. The system can translate fairly complicated sen~nces when equiped with the adequate gram~tical data and idiomatic phrases. Output sen~nces, even though far from eloquent style, are worth post-editing, and can be considerably improved with multiple meaning correction through interactive work. Some interesting technical findings are the following:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "EXPERIi 4ENTAL RESULTS", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "(1) The ~llowing items are sometimes syntactically ambiguous to the system. (i) ING + N (ambiguity among ADJ + SUBJ/OBJ, GOV + OBJ, and ~e like). (ii) To-infinitives (ambigui~ be~een adjective and adverbial). (iii) Linking scope ambigui~ w.r.t. \"and\", \"or\", \"of\" (A and B of C for D).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "EXPERIi 4ENTAL RESULTS", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "(iv) Embedded appositional phrases.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "EXPERIi 4ENTAL RESULTS", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "(2) Very long PE's (Phrasal Elements) appear occasionally. (eg. the PE node numbered 52 in Fig. 2 and Fig. 3 ).", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 91, |
|
"end": 108, |
|
"text": "Fig. 2 and Fig. 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "EXPERIi 4ENTAL RESULTS", |
|
"sec_num": "4." |
|
}, |
|
{ |
|
"text": "In this paper we t~ to contend that machine translation should be studied from more heuristic side, or from actual language-learner's methodology side rather than from purely rigid liguistical analysis side. Researchers of ve~ \"high level\" linguistic analysis side, as is poin~d out by Boitet [8] , \"seem too often to concentrate on in,resting high level phenomena as anaphoric re~rence, discourse structure, causality and reasoning and to forget at the same time persisting and very frequent lower-level difficulties .... \" This \"frequent lower-level difficulty\" is the very problem to be solved in practical machine translation, and is actually solved easily by naive foreign language learners only with the help of elementary grammatical knowledge. You had better recall that language-learners must solve the whole even though it is incomplete, on the other hand, pure linguists must solve completely even though it is very limited.", |
|
"cite_spans": [ |
|
{ |
|
"start": 293, |
|
"end": 296, |
|
"text": "[8]", |
|
"ref_id": "BIBREF7" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CONCLUDING REMARKS", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "In the light of this contention, we have devised a heuristic parsing model named HPM to accommodate the machine translation to the actual human translation methodologies, and at the same time, on HPM we have constructed a machine translation system named ATHENE. Experimental translation by ATHNE shows the following advantageous points of our heuristic approach.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CONCLUDING REMARKS", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "(1) Contribution to the flexibility, simplicity and maintenability in grammatical description.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CONCLUDING REMARKS", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "(2) Contribution to the simplicity and transparency in transforming phase and generating phase.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CONCLUDING REMARKS", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "One of further problems is to extend the grammatical data heuristically, so as to intensify our machine translation system from learner's level to expert's level.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CONCLUDING REMARKS", |
|
"sec_num": "5." |
|
}, |
|
{ |
|
"text": "Though our system can translate fairly complex sentences, it still commits learner's level errors when encountering difficulties such as ambiguity of prepositional group modification or of word linking scope for conjunction. Heuristic aspects of semantics are also our current interests of research. Especially the case-grammetical idea [9] seems to be useful to refine our syntactic-role assignment process so as to improve the quality of generated Japanese sentences. A kind of semantic code system (or thesaurus) will also be required to be introduced in our lexicons. Space limitation of this proceeding does not allow us to describe Our linguistic model: HPM in detail. We are planning to present the more detailed version of HPM together with later improvement in some appropriate journals.", |
|
"cite_spans": [ |
|
{ |
|
"start": 337, |
|
"end": 340, |
|
"text": "[9]", |
|
"ref_id": "BIBREF8" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CONCLUDING REMARKS", |
|
"sec_num": "5." |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "We would like to thank Prof. Nagao of Kyoto University for his kind and stimulative discussion on various aspects of machine translation. Thanks are also due to Dr. Miura, Dr. Kawasaki, Dr. Mitsumaki and Dr. Mitsumori of SDL Hitachi Ltd. for their constant encouragement to this ~ork.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "ACKNOWLEDGMENTS", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Mathematical Linguistics and Automatic Translation", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Kuno", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1962, |
|
"venue": "", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kuno, S. et. al., Mathematical Linguistics and Automatic Translation (Harvard Univ. Report, NSF-8 vol. l&2, 1962 & NSF-9 vol. 1&2, 1963).", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "A Theory of Syntactic Recognition for Natural Language", |
|
"authors": [ |
|
{ |
|
"first": "M", |
|
"middle": [ |
|
"P" |
|
], |
|
"last": "Marcus", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1980, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Marcus, M.P., A Theory of Syntactic Recognition for Natural Language (MIT Press, Cambridge, MA, 1980).", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Natural Language Information Processing", |
|
"authors": [ |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Sager", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1981, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sager, N., Natural Language Information Processing (Addison Wesley, Reading, MA, 1981).", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "DIAGRAM: A Grammar for Dialogues", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Robinson", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1982, |
|
"venue": "Comm. ACM", |
|
"volume": "25", |
|
"issue": "", |
|
"pages": "27--47", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Robinson, J.J., DIAGRAM: A Grammar for Dialogues, Comm. ACM 25, 1 (1982) 27-47.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "A Grammar of Contemporary English", |
|
"authors": [ |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Quirk", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1972, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Quirk et. al., A Grammar of Contemporary English (Longman, London; Seminar Press, New York, 1972).", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Aspects of the Theory of Syntax", |
|
"authors": [ |
|
{ |
|
"first": "N", |
|
"middle": [], |
|
"last": "Chomsky", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1965, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chomsky, N., Aspects of the Theory of Syntax (MIT Press, Cambridge, MA, 1965).", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Guide to Patterns and Usage in English", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Hornby", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1975, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hornby, A.S., Guide to Patterns and Usage in English, second edition (Oxford University Press, London, 1975).", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Present and Future Paradigms in the Automatized Translation of Natural Languages", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Boitet", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1980, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "430--436", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Boitet, C. et. al., Present and Future Paradigms in the Automatized Transla- tion of Natural Languages, COLING-80, Tokyo (1980) 430-436.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "The Case for Case", |
|
"authors": [ |
|
{ |
|
"first": "C", |
|
"middle": [ |
|
"J" |
|
], |
|
"last": "Fillmore", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1968, |
|
"venue": "Universals in Linguistic Theory", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--90", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Fillmore, C.J., The Case for Case, in: Bach and Harms {eds.), Universals in Linguistic Theory.(Holt, Rinehart and Winston, New York, 1968) 1-90.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"text": "Ambiguity Resolution in Attributes: To select most likely one from among many possibi)ities. (5) Segmentation into \"PE's\" and Attribute Assignment: To make a PE from matched WE group and give attribute(s). (6) Re-retrieval of Lexicon: To find again possible WE or PE, especially for \u2022 II \"the separated PE\" such as \"take,into consideratlon . Syntactic Role Assignment to PE's: To determine Syntactic Role of PE's by referring a pattern GD as in Fig. 4. l Attr. or Synt. Role Pattern I--'; .... Pattern to -~-Newly Assigne_d S.y.nt. Role Patter, i. ! ----,-SUBJ, \u00a2 ,',NAPP :, GOV SUBJ ,!ADJV , GOV GOV , NTHAT i , SUBJ, GOV Tari(et \"PE\") Assign \"Syntactic Role\" to PE Segmentation into \"CE's\" and Synt. Role Assignment: To make a CE from matched PE group and give a Synt. Role by referring patterns as in Fig. 5. \"(g) Determination of Modifying Relationships: To determine the appropriate element which the modifier PE should modify. (I0) Construction of Sentence Node (SENT): To complete the whole tree with the root node, SENT. (ll) Tree Transfor~tion: -To permute ~e PE's in each CE. Note that in our HPM,", |
|
"uris": null, |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"FIGREF1": { |
|
"text": "43 ~L.~,e,u~m-~,,~f__~ ~ ~_~ \"i: passive, as ~ possibl Transfo~ation and Sentence Generation (12) Assignment of case suffixes, postpositions and Japanese equivalents.", |
|
"uris": null, |
|
"type_str": "figure", |
|
"num": null |
|
}, |
|
"TABREF1": { |
|
"text": "Clausal Element (CE): CE is composed of one or more PE('s) which carries a part of sentential meaning in a nexus-like form. CE is nearly corresponding to a Japanese simple sentence such as:\"'v{wa/ga/wo/no}~,{suru/dearu} [koto]~' CE's allow mutual intersection. Typical examples are the underlined parts in the following: \"It is important for you to do so.\" 2.L Sentence (SE): SE is composed of one or more CE('s) and is located at the bottom of a parsed tree.", |
|
"type_str": "table", |
|
"content": "<table><tr><td colspan=\"5\">N)\", \"would not have been doen (V)\", and \"for everyday production use (PNAL)\".</td></tr><tr><td colspan=\"5\">2.5 2.7 Dependency Pattern of Verb: Verb-dependency-type code is deten~ined by</td></tr><tr><td colspan=\"3\">simplifying Hornby's classification[ 7], as i</td><td colspan=\"2\">Table 1.</td></tr><tr><td colspan=\"2\">Sub-Attr. of V I</td><td>Dependency Pattern</td><td colspan=\"2\">Sub-Att~. of N I</td><td>Examples</td></tr><tr><td>Vl</td><td colspan=\"2\">iBe + ...</td><td>Nl</td><td>Place</td></tr><tr><td>V6</td><td colspan=\"2\">Vi + To-infinitive</td><td>N2</td><td>Person, Organization</td></tr><tr><td>V7</td><td colspan=\"2\">Vt + Object</td><td>N3</td><td>Time</td></tr><tr><td>V8</td><td colspan=\"2\">Vt + that + ..-</td><td>N6</td><td>Abstract Concept</td></tr><tr><td>Vl4</td><td colspan=\"2\">Vt + Object [+not] + To-infinitive</td><td>N8</td><td>Means, Method</td></tr><tr><td colspan=\"3\">Table I. Sub-Attr. and Dependency</td><td colspan=\"2\">Table 2. Sub-Attr. of Noun</td></tr><tr><td/><td colspan=\"2\">Pattern of Verb</td><td/></tr><tr><td colspan=\"5\">2.8 Sub-Attribute of Noun: Noun is classified from somewhat semantical viewpoints</td></tr><tr><td colspan=\"2\">(cf. Table 2).</td><td/><td/></tr></table>", |
|
"num": null, |
|
"html": null |
|
} |
|
} |
|
} |
|
} |