|
{ |
|
"paper_id": "S07-1004", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:22:49.988045Z" |
|
}, |
|
"title": "SemEval-2007 Task 5: Multilingual Chinese-English Lexical Sample", |
|
"authors": [ |
|
{ |
|
"first": "Peng", |
|
"middle": [], |
|
"last": "Jin", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Peking University", |
|
"location": { |
|
"settlement": "Beijing", |
|
"country": "China" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Yunfang", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Peking University", |
|
"location": { |
|
"settlement": "Beijing", |
|
"country": "China" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Shiwen", |
|
"middle": [], |
|
"last": "Yu", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Peking University", |
|
"location": { |
|
"settlement": "Beijing", |
|
"country": "China" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "The Multilingual Chinese-English lexical sample task at SemEval-2007 provides a framework to evaluate Chinese word sense disambiguation and to promote research. This paper reports on the task preparation and the results of six participants.", |
|
"pdf_parse": { |
|
"paper_id": "S07-1004", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "The Multilingual Chinese-English lexical sample task at SemEval-2007 provides a framework to evaluate Chinese word sense disambiguation and to promote research. This paper reports on the task preparation and the results of six participants.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "The Multilingual Chinese-English lexical sample task is designed following the leading ideas of the Senseval-3 Multilingual English-Hindi lexical sample task . The \"sense tags\" for the ambiguous Chinese target words are given in the form of their English translations.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The data preparation is introduced in the second section. And then the participating systems are briefly described and their scores are listed.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In the conclusions we bring forward some suggestion for the next campaign.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "All the training and test data come from the People's Daily in January, February and March of 2000. The People's Daily is the most popular newspaper in China and is open domain. Before manually sense annotating, the texts have been word-segmented and part of speech (PoS) tagged according to the PoS tagging scheme of Institute of Computational Linguistics in Peking University (ICL/PKU). The corpus had been used as one of the gold-standard data set for the second international Chinese word segmentation bakeoff in 2005. 1", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Chinese Word Sense Annotated Corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The sense annotated corpus is manually constructed with the help of a word sense annotating interface developed in Java. Three native annotators, two major in Chinese linguistics and one major in computer science took part in the construction of the sense-annotated corpus. A text generally is first annotated by one annotator and then verified by two checkers. Checking is of course a necessary procedure to keep the consistency. Inspired by the observation that checking all the instances of a word in a specific time frame will greatly improve the precision and accelerate the speed, a software tool is designed in Java to gather all the occurrences of a word in the corpus into a checking file with the sense KWIC (Key Word in Context) format in sense tags order. The interannotator agreement gets to 84.8% according to Wu. et al. (2006) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 824, |
|
"end": 841, |
|
"text": "Wu. et al. (2006)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Manual Annotation", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "The sense entries are specified in the Chinese Semantic Dictionary (CSD) developed by ICL/PKU. The sense distinctions are made mainly according to the Contemporary Chinese Dictionary, the most widely used dictionary in mandarin Chinese, with necessary adjustment and improvement is implemented according to words usage in real texts. Word senses are described using the featurebased formalism. The features, which appear in the form \"Attribute =Value\", can incorporate extensive distributional information about a word sense. The feature set constitutes the representation of a sense, while the verbal definitions of meaning serve only as references for human use. The English translation is assigned to each sense in the attribute \"English translation\" in CSD.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Manual Annotation", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Based on the sense-annotated corpus, a sense is replaced by its English translation, which might group different senses together under the same English word.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Manual Annotation", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "In this task together 40 Chinese ambiguous words: 19 nouns and 21 verbs are selected for the evaluation. Each sense of one word is provided at least 15 instances and at most 40 instances, in which around 2/3 is used as the training data and 1/3 as the test data. In order to escape from the sense-skewed distribution that really exists in the corpus of People's Daily, many instances of some senses have been removed from the sense annotated corpus. So the sense distribution of the ambiguous words in this task does not reflect the usages in real texts.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Instances selection", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "In order to facilitate participators to select the features, we gave a specification for the PoS-tag set. Both word-segmented and un-segmented context are provided.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Two kinds of precisions are evaluated. One is micro-average:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2211 \u2211 = = = N i i N i i mir n m P 1 1 /", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "N is the number of all target word-types. is the number of labeled correctly to one specific tar-get word-type and is the number of all test instances for this word-type.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "i m i n", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "The other is macro-average:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2211 = = N i i mar N p P 1 / , i i i n m p / =", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "All teams attempted all test instances. So the recall is the same with the precision. The precision baseline is obtained by the most frequent sense. Because the corpus is not reflected the real usage, the precision is very low.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Six teams participated in this word sense disambiguation task. Four of them used supervised learning algorithms and two used un-supervised method. For each team two kinds of precision are given as in \u2022 All the verbs and nouns in the context, that is, the words with tags \"n, nr, ns, nt, nz, v, vd, vn\"", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 PoS of the left word and the right word \u2022 noun phrase, verb phrase, adjective phrase, time phrase, place phrase and quantity phrase.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "These phrases are considered as constituents of context, as well as words and punctuations which do not belong to any phrase.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022the type of these phrases which are around the target phrases", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 word category information comes from Chinese thesaurus I2R system used a semi-supervised classification algorithm (label propagation algorithm) (Niu, et al., 2005) . They used three types of features: PoS of neighboring words with position information, unordered single words in topical context, and local collocations.", |
|
"cite_spans": [ |
|
{ |
|
"start": 146, |
|
"end": 165, |
|
"text": "(Niu, et al., 2005)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In the label propagation algorithm (LP) (Zhu and Ghahramani, 2002) , label information of any vertex in a graph is propagated to nearby vertices through weighted edges until a global stable stage is achieved. Larger edge weights allow labels to travel through easier. Thus the closer the examples, the more likely they have similar labels (the global consistency assumption). In label propagation process, the soft label of each initial labeled example is clamped in each iteration to replenish label sources from these labeled data. Thus the labeled data act like sources to push out labels through unlabeled data. With this push from labeled examples, the class boundaries will be pushed through edges with large weights and settle in gaps along edges with small weights. If the data structure fits the classification goal, then LP algorithm can use these unlabeled data to help learning classification plane. CITYU-HIF system was a fully supervised one based on a Na\u00efve Bayes classifier with simple feature selection for each target word. The features used are as follows:", |
|
"cite_spans": [ |
|
{ |
|
"start": 40, |
|
"end": 66, |
|
"text": "(Zhu and Ghahramani, 2002)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "\u2022 Local features at specified positions:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "PoS of word at w -2 , w -1 , w 1 , w 2 Word at w -2 , w -1 , w 1 , w 2 \u2022 Topical features within a given window:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Content words appearing within w -10 to w 10 \u2022 Syntactic features:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "PoS bi-gram at w -2 w 0 , w -1 w 0 , w 0 w 1 , w 0 w 2 PoS tri-gram at w -2 w -1 w 0 and w 0 w 1 w 2", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "One characteristic of this system is the incorporation of the intrinsic nature of each target word in disambiguation. It is assumed that WSD is highly lexically sensitive and each word is best characterized by different lexical information. Human judged to consider for each target word the type of disambiguation information if they found useful. During disambiguation, they run two Na\u00efve Bayes classifiers, one on all features above, and the other only on the type of information deemed useful by the human judges. When the probability of the best guess from the former is under a certain threshold, the best guess from the latter was used instead.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "SWAT system uses a weighted vote from three different classifiers to make the prediction. The three systems are: a Na\u00efve Bayes classifier that compares similarities based on Bayes' Rule, a classifier that creates a decision list of context features, and a classifier that compares the angles between vectors of the features found most commonly with each sense. The features include bigrams, and trigrams, and unigrams are weighted by distance from the ambiguous word.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "TorMd used an unsupervised naive Bayes classifier. They combine Chinese text and an English thesaurus to create a `Chinese word'--`English category' co-occurrence matrix. This system generated the prior-probabilities and likelihoods of a Na\u00efve Bayes word sense classifier not from senseannotated (in this case English translation annotated) data, but from this word--category cooccurrence matrix. They used the Macquarie Thesaurus as very coarse sense inventory.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "They asked a native speaker of Chinese to map the English translations of the target words to appropriate thesaurus categories. Once the Na\u00efve Bayes classifier identifies a particular category as the intended sense, the mapping file is used to label the target word with the corresponding English translation. They rely simply on the bag of words that co-occur with the target word (window size of 5 words on either side).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "HIT is a fully unsupervised WSD system, which puts bag of words of Chinese sentences and the English translations of target ambiguous word to search engine (Google and Baidu). Then they could get all kinds of statistic data. The correct translation was found through comparing their cross entropy.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Participating Systems", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "The goal of this task is to create a framework to evaluate Chinese word sense disambiguation and to promote research. Table 3 : Performance on verbs. Micro / macro average precisions are spitted by \"/\" at the last row.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 118, |
|
"end": 125, |
|
"text": "Table 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Together six teams participate in this WSD task, four of them adopt supervised learning methods and two of them used unsupervised algorithms. All of the four supervised learning systems exceed obviously the baseline obtained by the most frequent sense. It is noted that the performances of the first three systems are very close. Two unsupervised methods' scores are below the baseline. More unlabeled data maybe improve their performance.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Scores", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Although the SRCB-WSD system got the highest scores among the six participants, it does not perform always better than other system from table 2 and table 3. But to each word, the four supervised systems always predict correctly more instances than the two un-supervised systems.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Scores", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Besides the corpus, we provide a specification of the PoS tag set. Only SRCB-WSD system utilized this knowledge in feature selection. We will provide more instances in the next campaign. Table 4 : Performance on nouns. Micro / macro average precisions are spitted by \"/\" at the last row.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 187, |
|
"end": 194, |
|
"text": "Table 4", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Scores", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "http://sighan.cs.uchicago.edu/bakeoff2005/", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "http:// maxent.sourceforge.net/", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "We would like to thank Tao Guo and Yulai Pei for their hard work to guarantee the quality of the corpus. Huiming Duan provides us the corpus which has been word-segmented and PoS-tagged and gives some suggestions during the manual annotation.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgements", |
|
"sec_num": "5" |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "The Senseval-3 English lexical sample task", |
|
"authors": [ |
|
{ |
|
"first": "Rada", |
|
"middle": [], |
|
"last": "Mihalcea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Timothy", |
|
"middle": [], |
|
"last": "Chklovski", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Adam", |
|
"middle": [], |
|
"last": "Kilgarriff", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of SENSEVAL-3", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "25--28", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rada Mihalcea, Timothy Chklovski and Adam Kilgar- riff. 2004. The Senseval-3 English lexical sample task. Proceedings of SENSEVAL-3. 25-28.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "The Senseval-3 Multilingual English-Hindi lexical sample task", |
|
"authors": [ |
|
{ |
|
"first": "Timothy", |
|
"middle": [], |
|
"last": "Chklovski", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rada", |
|
"middle": [], |
|
"last": "Mihalcea", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of SENSEVAL-3", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "5--8", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Timothy Chklovski, Rada Mihalcea, Ted Pedersen and Amruta Purandare. 2004. The Senseval-3 Multilin- gual English-Hindi lexical sample task. Proceedings of SENSEVAL-3. 5-8.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Learning from Labeled and Unlabeled Data with Label Propagation", |
|
"authors": [ |
|
{ |
|
"first": "Xiaojin", |
|
"middle": [], |
|
"last": "Zhu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Zoubin", |
|
"middle": [], |
|
"last": "Ghahramani", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "CMU CALD tech report CMU", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Xiaojin Zhu, Zoubin Ghahramani. 2002. Learning from Labeled and Unlabeled Data with Label Propagation. CMU CALD tech report CMU-CALD-02-107.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "A Chinese Corpus with Word Sense Annotation", |
|
"authors": [ |
|
{ |
|
"first": "Yunfang", |
|
"middle": [], |
|
"last": "Wu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Peng", |
|
"middle": [], |
|
"last": "Jin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yangsen", |
|
"middle": [], |
|
"last": "Zhang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Shiwen", |
|
"middle": [], |
|
"last": "Yu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of ICCPOL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "414--421", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yunfang Wu, Peng Jin, Yangsen Zhang, and Shiwen Yu. 2006. A Chinese Corpus with Word Sense Annota- tion. Proceedings of ICCPOL, Singapore, 414-421.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Word Sense Disambiguation Using Label Propagation Based Semi Supervised Learning", |
|
"authors": [ |
|
{ |
|
"first": "Zhen-Yu", |
|
"middle": [], |
|
"last": "Niu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dong-Hong", |
|
"middle": [], |
|
"last": "Ji", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chew-Lim", |
|
"middle": [], |
|
"last": "Tan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "395--402", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Zhen-Yu Niu, Dong-Hong Ji and Chew-Lim Tan. 2005. Word Sense Disambiguation Using Label Propaga- tion Based Semi Supervised Learning. Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics.395-402", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF0": { |
|
"type_str": "table", |
|
"num": null, |
|
"text": "", |
|
"content": "<table><tr><td/><td colspan=\"3\">presents the number of words</td></tr><tr><td colspan=\"4\">under each part of speech, the average number of</td></tr><tr><td colspan=\"4\">senses for each PoS and the number of instances</td></tr><tr><td colspan=\"4\">respectively in the training and test set.</td></tr><tr><td/><td># Average</td><td># training</td><td># test</td></tr><tr><td/><td>senses</td><td>instances</td><td>instances</td></tr><tr><td>19</td><td>2.58</td><td>1019</td><td>364</td></tr><tr><td>nouns</td><td/><td/><td/></tr><tr><td>21</td><td>3.57</td><td>1667</td><td>571</td></tr><tr><td>verbs</td><td/><td/><td/></tr><tr><td colspan=\"4\">Table 1: Summary of the sense inventory and</td></tr><tr><td colspan=\"3\">number of training data and test set</td><td/></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF1": { |
|
"type_str": "table", |
|
"num": null, |
|
"text": "", |
|
"content": "<table><tr><td>Team</td><td colspan=\"2\">Micro-average Macro-average</td></tr><tr><td>SRCB-WSD</td><td>0.716578</td><td>0.749236</td></tr><tr><td>I2R</td><td>0.712299</td><td>0.746824</td></tr><tr><td>CITYU-HIF</td><td>0.710160</td><td>0.748761</td></tr><tr><td>SWAT</td><td>0.657754</td><td>0.692487</td></tr><tr><td>TorMd</td><td>0.375401</td><td>0.431243</td></tr><tr><td>HIT</td><td>0.336898</td><td>0.395993</td></tr><tr><td>baseline</td><td>0.4053</td><td>0.4618</td></tr></table>", |
|
"html": null |
|
}, |
|
"TABREF2": { |
|
"type_str": "table", |
|
"num": null, |
|
"text": "The scores of all participating systems As follow the participating systems are briefly introduced. SRCB-WSD system exploited maximum entropy model as the classifier from OpenNLP 2 The following features are used in this WSD system:", |
|
"content": "<table/>", |
|
"html": null |
|
} |
|
} |
|
} |
|
} |