|
{ |
|
"paper_id": "2020", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T10:24:09.237322Z" |
|
}, |
|
"title": "FINSIM20 at the FinSim Task: Making Sense of Text in Financial Domain", |
|
"authors": [ |
|
{ |
|
"first": "Vivek", |
|
"middle": [], |
|
"last": "Anand", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Yash", |
|
"middle": [], |
|
"last": "Agrawal", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Aarti", |
|
"middle": [], |
|
"last": "Pol", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "VIT", |
|
"institution": "", |
|
"location": { |
|
"settlement": "Pune" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Vasudeva", |
|
"middle": [], |
|
"last": "Varma", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "Semantics play an important role when it comes to automated systems using text or language and it is different for different domains. In this paper, we tackle the FinSim 2020 shared task at IJCAI-PRICAI 2020. The task deals with designing a semantic model which can automatically classify short phrases/terms from financial domain into the most relevant hypernym (or top-level) concept in an external ontology. We perform several experiments using different kinds of word and phrase level embeddings to solve the problem in an unsupervised manner. We also explore the use of a supplementary financial domain data; either to learn better concept representation or generate more training samples. We discuss both the positive and negative results that we observed while applying these approaches.", |
|
"pdf_parse": { |
|
"paper_id": "2020", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "Semantics play an important role when it comes to automated systems using text or language and it is different for different domains. In this paper, we tackle the FinSim 2020 shared task at IJCAI-PRICAI 2020. The task deals with designing a semantic model which can automatically classify short phrases/terms from financial domain into the most relevant hypernym (or top-level) concept in an external ontology. We perform several experiments using different kinds of word and phrase level embeddings to solve the problem in an unsupervised manner. We also explore the use of a supplementary financial domain data; either to learn better concept representation or generate more training samples. We discuss both the positive and negative results that we observed while applying these approaches.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Semantics has been a tough area in NLP research. This also comes in the disguise of getting to know the taxonomy or hypernymy relations of terms. There have been tasks in NLP for this purpose [Bordea et al., 2016; Camacho-Collados et al., 2018] . These tasks get tougher when applied for a specific domain; for example, the word \"stocks\" can have several meaning in general sense but while in financial domain; the meaning can be narrowed down. Thus making the semantics a bit more clear.", |
|
"cite_spans": [ |
|
{ |
|
"start": 192, |
|
"end": 213, |
|
"text": "[Bordea et al., 2016;", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 214, |
|
"end": 244, |
|
"text": "Camacho-Collados et al., 2018]", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The purpose of FinSim 2020 Shared Task is to automatically map financial phrases/terms to a more general financial concept. Alternatively, it means to map a hyponymy to its hypernymy. This kind of task in financial domain has been introduced for the first time.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The task provides us with a training data that maps some of the financial concepts to its hypernymy; for example, \"Alternative Debenture\" is mapped to \"Bonds\". The given set of hypernymy labels has cardinality of 8. This set includes Bonds, Forward, Funds, Future, MMIs, Option, Stocks and Swaps. The pre-mapped data (training data) contains very * Authors contributed equally. low number of labelled examples so we explore unsupervised techniques.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "We explore the use of pre-trained word embeddings [Pennington et al., 2014] . These pre-trained word embeddings are trained on general corpus and not domain specific. We make use of the given training data to explore if the pre-trained word embeddings can be used for this financial domain task. We perform experiments with several kinds of unsupervised algorithms based on word embeddings using cosine similarity and variants of KNN algorithm as well as using some deep learning based methods.", |
|
"cite_spans": [ |
|
{ |
|
"start": 50, |
|
"end": 75, |
|
"text": "[Pennington et al., 2014]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Further the task also provides us with corpus of financial domain text. This text can be used to learn representations or patterns useful for the task. We explore the use of Hearst Patterns [Hearst, 1992; Seitner et al., 2016] on this text to automatically mine hypernymy-hyponymy relations from the given text which is useful in extending the training dataset. We perform similar experiments on this extended dataset and report the results.", |
|
"cite_spans": [ |
|
{ |
|
"start": 190, |
|
"end": 204, |
|
"text": "[Hearst, 1992;", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 205, |
|
"end": 226, |
|
"text": "Seitner et al., 2016]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The remainder of paper is organized as follows. Section 2 describes the approaches we have tried for the task. Section 3 describes experimental setting and some details regarding the approaches. Section 4 describes the results achieved from several methods and out ranking in the task. We then conclude in Section 5.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "We explore the use of pre-trained word and sentence embeddings in this approach. For basleine, we consider GloVe [Pennington et al., 2014] word embeddings and it's finetuned versioned on given financial documents. Since, GloVe embedding are word based, to get the embedding of financial concept; we take pre-trained word embedding of each word in the input financial concept and average it. For each of the input financial concept we try and map it to its hypernymy using the averaged word of both the input financial concept and the hypernymy label itself. We find the cosine similarity with each of the average embedding of hypernymy labels to get a ranked list of labels for a given financial concept.", |
|
"cite_spans": [ |
|
{ |
|
"start": 113, |
|
"end": 138, |
|
"text": "[Pennington et al., 2014]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Cosine Similarity Based", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "We also experiment with mapping financial concept to averaged embedding of the description of hypernymy as per The Financial Industry Business Ontology (FIBO). Due to the av- eraging, description based approach performed poorly compared to just using embedding of hypernymy. For sentence embeddings, we use Universal Sentence Encoder (USE) [Cer et al., 2018] pre-trained embeddings. USE gives 512 dimensional embedding of the given input phrase. A t-SNE representation of these embedding is for each financial concept is shown in Figure 3 . Again, we find the cosine similarity of given financial concept with each of the hypernymy labels to get a ranked list of labels for that concept. In case of USE, description of labels performs better than GloVe embeddings but is still worse than using hypernymy labels only.", |
|
"cite_spans": [ |
|
{ |
|
"start": 340, |
|
"end": 358, |
|
"text": "[Cer et al., 2018]", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 530, |
|
"end": 538, |
|
"text": "Figure 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Cosine Similarity Based", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "We have very few labeltavled examples to train a supervised model and the class distribution isn't consistent. For example, label \"bonds\" and \"swap\" have close to half training samples and other half is distributed among six labels. In order to handle class imbalance, we use weighted cross-entropy loss function. We experiment with CNN for text [Kim, 2014] , LSTM [Hochreiter and Schmidhuber, 1997 ] and a transformer based RoBERTa model [Liu et al., 2019] . We find that these supervised model are able to learn the task to a reasonably good extent and all gives almost similar results. However, even with a very large number of parameters, RoBERTa model gives only comparable scores to both CNN and LSTM based models. This might be due to the fact that RoBERTa model need larger number of labelled samples for the finetuning.", |
|
"cite_spans": [ |
|
{ |
|
"start": 346, |
|
"end": 357, |
|
"text": "[Kim, 2014]", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 365, |
|
"end": 398, |
|
"text": "[Hochreiter and Schmidhuber, 1997", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 439, |
|
"end": 457, |
|
"text": "[Liu et al., 2019]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Deep Learning Based", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "In this approach, the main idea is to map the given input financial concept with one of the financial concept in the training set. We use average pre-trained embedding to get representation of the given financial concept. We get the cosine similarity score with average pre-trained embedding of all the financial concepts present in the training set. We only consider the top k most similar financial concepts to the input financial concept. We finally consider the label of these k financial concepts and output the most frequent label. This intuition behind this approach is that the input financial concept will be most similar to all the other financial concepts which come under the same category. Alternatively, it can be said as find the most similar sibling and concluding they have same parent.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Naive KNN Based", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "This approach is similar to the Naive KNN based approach but we introduce the external financial domain documents in this case. We consider all the documents and run Hearst Patterns [Hearst, 1992; Seitner et al., 2016] on it. We get a database of automatically extracted hyponymy-hypernymy relations in financial domain form this. We make use of these extracted relations to infer relations for concepts during test time. We hypothesize that the input financial concept whose hypernymy is to be predicted is present in the automatically extracted database. However the exact term match would be crude way to do so. So we use word embedding based similarity to get a perfect match. even if there is an exact match, word embedding based similarity would give the highest score in that case.", |
|
"cite_spans": [ |
|
{ |
|
"start": 182, |
|
"end": 196, |
|
"text": "[Hearst, 1992;", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 197, |
|
"end": 218, |
|
"text": "Seitner et al., 2016]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Extended KNN based", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "For a given test financial concept, we take its average pretrained embedding. We also consider average pre-trained embedding of all the hyponymys present in automatically extracted database. We compare both these embeddings using cosine similarity. We find the most representative hyponymy from automatically extracted database. This can be thought of as KNN with k = 1. We then take this representative hyponymy and compare its hypernymy with our set of labels. This is again done by taking average pre-trained embeddings and taking cosine similarity. This gives us the most similar label from the set.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Extended KNN based", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "We again make use of the automatically extracted hyponymyhypernymy relations from external financial domain text. This database can contain different type of entities which may not be the exact match with concepts of our interest. There can be several hops in relations before we finally reach the parent hypernymy. For example \"Equity Linked Bond\" is a \"Variable Coupon Bond\" which in turn is a \"Bond\". Therefore we have to traverse the relations completely in order to get the broader picture. For this effect we turn to a graph based approach. We build a graph with entities as nodes and relations as edges. These entities come form the automatically extracted hypernymy-hyponymy database done using Hearst Patterns [Hearst, 1992; Seitner et al., 2016] . For each relations we add an undirected edge.", |
|
"cite_spans": [ |
|
{ |
|
"start": 719, |
|
"end": 733, |
|
"text": "[Hearst, 1992;", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 734, |
|
"end": 755, |
|
"text": "Seitner et al., 2016]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Graph Based", |
|
"sec_num": "2.5" |
|
}, |
|
{ |
|
"text": "We leverage the connections in the graph to predict the hypernymy of the financial concept. For the input financial concept we find a representative node using cosine similarity among the average pre-trained embeddings. Once we get the representative node we consider the connected component of the graph containing that representative node. The intuition is that the hypernymy label should be present in one of the nodes in this connected component. This is because of the whole taxonomic structure and relations among entities. So we only consider the connected component containing the representative node and find the hypernymy label node in it. For each possible label we compute similarity scores again using average pre-trained embedding. We consider the maximum similarity score that we get when comparing each node with the label and assign it to that particular label. This way we will have scores for each of the label. Label with the maximum score is given as the prediction. Figure 2 gives the overview of the overall graph based approach.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 988, |
|
"end": 996, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Graph Based", |
|
"sec_num": "2.5" |
|
}, |
|
{ |
|
"text": "For pre-trained embeddings, we use 100 dimensional GloVe word embeddings and 512 dimensional USE sentence embeddings. We trained all deep learing architechure using Adam optimizer [Kingma and Ba, 2014] with 0.001 learning rate. For Naive KNN based we use the K = 10 as it gave best results for this method. We use hearstPatterns python library for the implementation of hearst patterns. It was used the extended mode to mine additional patterns. NetworkX python library was used for implementation of graph based algorithms. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 180, |
|
"end": 201, |
|
"text": "[Kingma and Ba, 2014]", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experiment", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "This paper mainly discusses how we tackle the FinSim 2020 shared task. The task is to automatically map financial concepts with its hypernymy. For this purpose we we explore the different ways in which we can learn the semantics in financial document for automatically predicting the hypernymy relations of financial concepts. We explore how pre-trained word and sentence embeddings can be used for this task. We experiment with both traditional and current deep learning architectures. We further explore how external financial documents can be useful. Our best method accomplishes good results for the task and puts us in one of the top positions among other participants of the task.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "Proceedings of the Second Workshop on Financial Technology and Natural Language Processing", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Semeval-2016 task 13: Taxonomy extraction evaluation (texeval-2)", |
|
"authors": [ |
|
{ |
|
"first": "[", |
|
"middle": [], |
|
"last": "References", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Bordea", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the 10th International Workshop on Semantic Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "References [Bordea et al., 2016] Georgeta Bordea, Els Lefever, and Paul Buitelaar. Semeval-2016 task 13: Taxonomy extraction evaluation (texeval-2). In Proceedings of the 10th Interna- tional Workshop on Semantic Evaluation. Association for Computational Linguistics, 2016.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Vered Shwartz, Roberto Navigli, and Horacio Saggion. Semeval-2018 task 9: Hypernym discovery", |
|
"authors": [ |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Camacho-Collados", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 12th International Workshop on Semantic Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Camacho-Collados et al., 2018] Jose Camacho-Collados, Claudio Delli Bovi, Luis Espinosa-Anke, Sergio Oramas, Tommaso Pasini, Enrico Santus, Vered Shwartz, Roberto Navigli, and Horacio Saggion. Semeval-2018 task 9: Hypernym discovery. In Proceedings of the 12th Interna- tional Workshop on Semantic Evaluation (SemEval-2018);", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "ACL (Association for Computational Linguistics)", |
|
"authors": [], |
|
"year": 2018, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "712--736", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "2018. p. 712-24. ACL (Association for Computational Linguistics), 2018.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Chris Tar, Yun-Hsuan Sung, Brian Strope, and Ray Kurzweil. Universal sentence encoder", |
|
"authors": [ |
|
{ |
|
"first": "[", |
|
"middle": [], |
|
"last": "Cer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "[Cer et al., 2018] Daniel Cer, Yinfei Yang, Sheng yi Kong, Nan Hua, Nicole Limtiaco, Rhomni St. John, Noah Con- stant, Mario Guajardo-Cespedes, Steve Yuan, Chris Tar, Yun-Hsuan Sung, Brian Strope, and Ray Kurzweil. Uni- versal sentence encoder, 2018.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Automatic acquisition of hyponyms from large text corpora", |
|
"authors": [ |
|
{ |
|
"first": ";", |
|
"middle": [], |
|
"last": "Hearst", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Marti", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Hearst", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1992, |
|
"venue": "Proceedings of the 14th conference on Computational linguistics", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "539--545", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Hearst, 1992] Marti A Hearst. Automatic acquisition of hy- ponyms from large text corpora. In Proceedings of the 14th conference on Computational linguistics-Volume 2, pages 539-545. Association for Computational Linguis- tics, 1992.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Long short-term memory", |
|
"authors": [ |
|
{ |
|
"first": "Sepp", |
|
"middle": [], |
|
"last": "Hochreiter", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J\u00fcrgen", |
|
"middle": [], |
|
"last": "Schmidhuber", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1997, |
|
"venue": "Neural Comput", |
|
"volume": "9", |
|
"issue": "8", |
|
"pages": "1735--1780", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "[Hochreiter and Schmidhuber, 1997] Sepp Hochreiter and J\u00fcrgen Schmidhuber. Long short-term memory. Neural Comput., 9(8):1735-1780, November 1997.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "A large database of hypernymy relations extracted from the web", |
|
"authors": [ |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Kim ; Yoon Kim ; Diederik", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jimmy", |
|
"middle": [], |
|
"last": "Kingma", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": ";", |
|
"middle": [], |
|
"last": "Ba", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Liu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC'16)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "360--367", |
|
"other_ids": { |
|
"arXiv": [ |
|
"arXiv:1412.6980" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kim, 2014] Yoon Kim. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Con- ference on Empirical Methods in Natural Language Pro- cessing (EMNLP), pages 1746-1751, Doha, Qatar, Octo- ber 2014. Association for Computational Linguistics. [Kingma and Ba, 2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014. [Liu et al., 2019] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019. [Pennington et al., 2014] Jeffrey Pennington, Richard Socher, and Christopher D. Manning. Glove: Global vectors for word representation. In Empirical Meth- ods in Natural Language Processing (EMNLP), pages 1532-1543, 2014. [Seitner et al., 2016] Julian Seitner, Christian Bizer, Kai Eckert, Stefano Faralli, Robert Meusel, Heiko Paulheim, and Simone Paolo Ponzetto. A large database of hyper- nymy relations extracted from the web. In Proceedings of the Tenth International Conference on Language Re- sources and Evaluation (LREC'16), pages 360-367, 2016.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "Example of financial concept labeling" |
|
}, |
|
"FIGREF1": { |
|
"num": null, |
|
"type_str": "figure", |
|
"uris": null, |
|
"text": "Overview of Graph Based Approach Figure 3: Illustration of financial concept USE embeddings using t-SNE" |
|
}, |
|
"TABREF0": { |
|
"type_str": "table", |
|
"text": "summarizes the results for various methods described above. USE embedding based cosine similarity gave the best results in both metrics -mean rank and accuracy. Deep learning based architectures also gave similar good scores while graph based methods didn't perform well. ResultsTable.Glove and USE embedding methods are Cosine Similarity based.", |
|
"content": "<table><tr><td>Method</td><td colspan=\"2\">Mean Rank Accuracy</td></tr><tr><td>GloVe</td><td>1.84</td><td>0.63</td></tr><tr><td>GloVe (fine-tuned)</td><td>1.79</td><td>0.67</td></tr><tr><td colspan=\"2\">GloVe (FIBO Description) 2.07</td><td>0.43</td></tr><tr><td>USE</td><td>1.43</td><td>0.79</td></tr><tr><td>USE (FIBO Description)</td><td>2.08</td><td>0.46</td></tr><tr><td>CNN</td><td>1.44</td><td>0.77</td></tr><tr><td>LSTM</td><td>1.44</td><td>0.78</td></tr><tr><td>RoBERTa</td><td>1.45</td><td>0.78</td></tr><tr><td>Naive KNN Based</td><td>1.75</td><td>0.61</td></tr><tr><td>Extended KNN based</td><td>3.80</td><td>0.08</td></tr><tr><td>Graph Based</td><td>2.68</td><td>0.19</td></tr></table>", |
|
"html": null, |
|
"num": null |
|
} |
|
} |
|
} |
|
} |