|
{ |
|
"paper_id": "2020", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T10:23:59.702598Z" |
|
}, |
|
"title": "Knowledge Graph and Deep Neural Network for Extractive Text Summarization by Utilizing Triples", |
|
"authors": [ |
|
{ |
|
"first": "Amit", |
|
"middle": [], |
|
"last": "Vhatkar", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Pushpak", |
|
"middle": [], |
|
"last": "Bhattacharyya", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Kavi", |
|
"middle": [], |
|
"last": "Arya", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "In our research work, we represent the content of the sentence in graphical form after extracting triples from the sentences. In this paper, we will discuss novel methods to generate an extractive summary by scoring the triples. Our work has also touched upon sequence-to-sequence encoding of the content of the sentence, to classify it as a summary or a non-summary sentence. Our findings help to decide the nature of the sentences forming the summary and the length of the system generated summary as compared to the length of the reference summary.", |
|
"pdf_parse": { |
|
"paper_id": "2020", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "In our research work, we represent the content of the sentence in graphical form after extracting triples from the sentences. In this paper, we will discuss novel methods to generate an extractive summary by scoring the triples. Our work has also touched upon sequence-to-sequence encoding of the content of the sentence, to classify it as a summary or a non-summary sentence. Our findings help to decide the nature of the sentences forming the summary and the length of the system generated summary as compared to the length of the reference summary.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Extractive summaries contain the most informative sentences from the input text. The ordered pair of Subject(S), Verb(V) and Object(O) i.e. 1 < S, V, O> represent the content of the sentence. We form a knowledge-graph(KG) by considering words in the triples. Our novel methods choose the informative sentences based on the count of frequencies calculated using generated KG. We also have implemented machine Learning(ML) and Deep Neural Network(DNN) based models. These models make use of the KG based features which try to capture information available. We are making use of dataset made available for FNS-2020 shared task by El-Haj et al. (2020) . 2 We have used SpaCy library for extracting triples. We have used python implementation of Rouge package made available by 3 PyPI, which implements ROUGE described by Lin (2004) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 627, |
|
"end": 647, |
|
"text": "El-Haj et al. (2020)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 650, |
|
"end": 651, |
|
"text": "2", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 817, |
|
"end": 827, |
|
"text": "Lin (2004)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In general, we pose extractive summarization as a sentence classification and a triple classification task. We perform this classification using algorithms like SVM, SVR, Neural Network(NN) and Long Short-Term Memory(LSTM/DNN). This section describes our implemented approaches in details.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Implemented Approaches", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Summary Sentences Non Summary Sentences Total Sentences Train Set 0.3M 2.6M 2.9M Validation Set 0.051M 0.663M 0.714M Table 1 : Distribution of Summary, Non-Summary Sentences Extracted from Training and Validation Set FNS-2020 Shared task training and validation dataset comes with up to seven gold summaries. All the sentences present in the gold summary are extractive in nature. All available gold summaries of the specific document are used for labelling the sentences in the given text.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 117, |
|
"end": 124, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Labelling and Feature Extraction", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "This work is licensed under a Creative Commons Attribution 4.0 International License. License details: http:// creativecommons.org/licenses/by/4.0/.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labelling and Feature Extraction", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "1 Referred to as triple 2 https://spacy.io/ 3 https://pypi.org/project/rouge/", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labelling and Feature Extraction", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "We have considered features as, Position of the sentenceeach sentence is marked according to its position (i.e. 1-Introductory, 2-Concluding and 3-Explanatory), Lengthnumber of words present in the sentence, Thematic Wordsnumber of most 4 frequent words present in the sentence, Indicator Words-Count of words present in the available 5 synset(i.e. the group of synonymous words) of words 'conclusion' and 'summary', Uppercasenumber of uppercase words present in the 6 sentence, Important Word Feature-It represents quotient of the available triple in the sentence to the total triples in the text file, and KG File Feature-It represents quotient of the available triples in the sentence represented in terms of 7 lookup frequencies to the number of total triples in the text file. We also have used pre-trained 100-dimensional GloVe (Pennington et al., 2014) embedding for DNN based approaches.", |
|
"cite_spans": [ |
|
{ |
|
"start": 834, |
|
"end": 859, |
|
"text": "(Pennington et al., 2014)", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Labelling and Feature Extraction", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Subject, verb and object(<S, V, O> i.e. triple) are main content words available in any sentence. Individual count of S, V and O present in the document fails to represent content available the sentence. To generate a content-aware extractive summary, TFM chooses sentences containing the highest scoring triples. Based on score computation, we have defined three different models.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Triple Frequency-based Models(TFM)", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "This is the simplistic approach to generate extractive summaries by making use of Triple Frequencies. This method fails to identify important sentences in the case of equal distribution of triples.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Plain Frequency Model(PKG)", |
|
"sec_num": "2.2.1" |
|
}, |
|
{ |
|
"text": "1. Extract all triples available in the text document and maintain it's count 2. Generate the score of the triple by considering its count 3. Generate extractive summary by selecting the sentences containing top-K triples", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Plain Frequency Model(PKG)", |
|
"sec_num": "2.2.1" |
|
}, |
|
{ |
|
"text": "PFM fails when the majority of the extracted triples gets an equal score. We tried to remove the equal scoring by considering frequencies of the subject and object present in the triples for scoring the triples.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Updated Frequency Model(UKG)", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "1. Extract all triples available in the text document 2. Generate the score of the triple using the following formula, triple score = F requency of triple + f requency of subject + f requency of object", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Updated Frequency Model(UKG)", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "(1)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Updated Frequency Model(UKG)", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "3. Generate extractive summary by selecting the sentences containing top-K triples", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Updated Frequency Model(UKG)", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "The formula 1, helps to break the uniformity of the scores occurring in the PFM by giving importance to the subjects and the objects available in the sentence.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Updated Frequency Model(UKG)", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "This approach considers two disjoint sets of documents to generate a score of the triples, 1-Train fold: used to extract and score the triple, 2-Test fold: used to generate the summaries and check the performance.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Five Fold Cross Validation Model(FKG)", |
|
"sec_num": "2.2.3" |
|
}, |
|
{ |
|
"text": "1. Extract and pre-compute frequencies of triples based on all documents present in the training folds and extract triples from the document present in test fold 2. Generate the score of the triple extracted from test document by considering its count which is computed (after considering all documents in remaining folds) in Step-1. 4 Based on occurrence in the document 5 We have used wordnet made available in nltk library for getting synset 6 Excluding word 'I'(most commonly occurring uppercase word) 7 Entire dataset is divided in five disjoint folds of which four-fold forms training set i.e. lookup 132 3. Generate extractive summary by selecting the sentences containing top-K triples This method helps us to gain an insight over the presence of general sentences related to the topic or domain and the presence of the sentences specific to the document. Table 2 represents the extraction statistics of FNS-2020 training set. We have considered 2580 training documents for extraction. ", |
|
"cite_spans": [ |
|
{ |
|
"start": 334, |
|
"end": 335, |
|
"text": "4", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 372, |
|
"end": 373, |
|
"text": "5", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 864, |
|
"end": 871, |
|
"text": "Table 2", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Five Fold Cross Validation Model(FKG)", |
|
"sec_num": "2.2.3" |
|
}, |
|
{ |
|
"text": "We have implemented two machine learning-based approaches, 1-SVM, 2-SVR. Implementation of SVM for extractive summarization roughly follows the method used by Chali et al. (2009) ; where the task of extractive summarization is posed as a binary classification task. SVR for extractive summarization is based on the discussion by Li et al. (2007) . SVR for extractive summarization assigns the score to the sentence. Summary sentences are scored as 1 and non-summary as 0. SVR approach predicts the score for sentences in the document and we select top-k scoring sentences as a summary sentences. SVM and SVR make use to the features mentioned in section 2.1. Both of the models are trained on 0.6M sentences with an equal mix of the classes.", |
|
"cite_spans": [ |
|
{ |
|
"start": 159, |
|
"end": 178, |
|
"text": "Chali et al. (2009)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 329, |
|
"end": 345, |
|
"text": "Li et al. (2007)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Machine Learning based Summarization", |
|
"sec_num": "2.3" |
|
}, |
|
{ |
|
"text": "Along with KG and ML-based approaches, we have implement DNN based approaches to generate extractive summaries by performing binary class classification.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Deep Neural Network based Approaches", |
|
"sec_num": "2.4" |
|
}, |
|
{ |
|
"text": "We have trained a feed-forward neural network to classify a given sentence. Input layer consumes the features mentioned in the section 2.1. The model's architecture is the input layer followed by a dense layer with eight neurons followed by an output layer. ReLU and Sigmoid activation functions were used with Binary Cross Entropy as a loss function and Adam as an optimizer. Model Performs best when we set batch size as 32, Train-Validation split as 70-30% and train for 150 epochs.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Neural Network Model", |
|
"sec_num": "2.4.1" |
|
}, |
|
{ |
|
"text": "We have trained LSTM models to classify the sentence as a summary and non-summary sentence. The encoder uses the entire sequence of the words present in the sentence to capture content information, for that we have used pre-trained 100-dimensional GloVe embedding which does not evolve during the training phase. The architecture is made up of an embedding layer followed by LSTM layer (with 2% dropout) followed by a softmax layer. Categorical cross-entropy(a generalized form of binary-cross entropy) with Adam optimization technique is used for training. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "S-LSTM for Extractive Summarization", |
|
"sec_num": "2.4.2" |
|
}, |
|
{ |
|
"text": "In this approach, words present in the triples were passed to the LSTM encoder, unlike S-LSTM where all the words in the sentence get passed. Based on the presence in the summary sentences, triples are marked as a summary or a non-summary triples. We have trained this model on positional embedding. All words in all triples are used to generate positional embedding. Embedding of the words presented in the triples are concatenated and is passed as input to the encoder of LSTM. The total number of words in the triples are less than the total number of words in all of the documents. Positional embedding tries to get an exact representation of the content of the sentence represented by the triple. Architecture details of this model remain the same as S-LSTM. We have implemented eight different approaches while considering TextRank (Mihalcea and Tarau, 2004) as the baseline approach. In this section, we discuss the performance of the implemented models on the validation set followed by a performance on the test set.", |
|
"cite_spans": [ |
|
{ |
|
"start": 838, |
|
"end": 864, |
|
"text": "(Mihalcea and Tarau, 2004)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "T-LSTM for Extractive Summarization", |
|
"sec_num": "2.4.3" |
|
}, |
|
{ |
|
"text": "The validation set of FNS-2020 Shared Task consists of 363 documents each having up to seven gold summaries. We are comparing results obtained over single gold summary of the specific document. System generated summaries were constrained to have 1000 words and our approaches select first 1000 words because, After segmenting given text in three parts each containing equal portions of the text we have found that in the training set 96%(i.e. 0.28M out of 0.29M) and in the validation set 95%(i.e. 49K out of 51K) of the summary sentences of the gold summaries are present in the first part of the text. Table 5 represents ROUGE-1 score of all implemented models when reference gold summary is allowed to contain all of its text(i.e. Full Length) and when it is allowed to contain first 1000 words(i.e. Limited Length) of its text. The ROUGE score values in the table are averaged over the averaged(over 3 runs) ROUGE score of all documents in the validation set. In the result table, overall highest score are bold-faced while highest score among specific category is italicized.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 604, |
|
"end": 611, |
|
"text": "Table 5", |
|
"ref_id": "TABREF7" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Validation Set", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Considering the full-length gold summary S-LSTM performs the best amongst all approaches. In the setting of limited length gold summary T-LSTM approach performs best. All approaches perform best in the limited length gold summary setting. Therefore, to obtain better results, length of system generated summary should be equal to the length of the reference summary.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model Name", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Even after being rule-based models, TFM models have performed comparably well. Nature of text causes UKG to give an equal score to the triples affecting its performance. In the FKG model, generic triples get a higher score as the effect of considering all documents in training fold. This leads to a summary containing generic sentences. However, as PKG and UKG performance better than FKG, the summary should contain sentences specific to the document", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model Name", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We have generated summaries of the 500 documents present in the test set using our 11 NN, S-LSTM and an SVM approach. Gold summaries of the documents in the test set are used by the organizers as a reference summaries to compute the results. The ROUGE values in the table 6 are published by the organizers of the shared task. Organizers also had computed the ROUGE scores of their baseline approaches(as mentioned in When compared on ROUGE-1, our NN based approach is among top-5 while SVM and S-LSTM approaches have secured 9th and 10th position respectively. SVM, NN and S-LSTM ranked 10th, 11th and 12th respectively on ROUGE-2 metric. Our approaches perform quite well as compare to the baseline approaches. No one approach outperformed the others in all ROUGE metrics.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Test Set", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "We have successfully generated extractive summaries using our novel methods of triple scoring which are based on KG generated by the words in the triples. We have also proposed novel DNN based approaches for extractive summarization, where summarization is carried by performing binary classification after sequence-to-sequence encoding(either sentence or triples) content present in the input text. From our discussion in section 3.1, we can conclude that the summary should contain sentences specific to the document. We have seen that, in order to get better results, length of system generated summary should be equal to the length of the reference summary. We also have seen that KG-based Triple Frequency models perform comparably well than baseline models and possess scope of the improvement.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "Our Baseline Model", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "11 Only up to 3 summaries per document are allowed by organizers of FNS-Shared Task 12 Different systems performed well on different ROUGE metric", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "acknowledgement", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We have created web service for end-users to summarize their text either by uploading a file or by type-in text functionality. Figure 1 depicts the landing page of web service where users can see the summarized text along with the 13 KG that gets generated and used for summarizing the input text. The web portal also mentions the percentage reduction in terms of the number of words and ROUGE score calculated using TextRank as a reference summary. Web service displays PKG, UKG summaries along with TextRank summary. It also has provision to host a dataset and it currently hosts part of CNN-Daily mail dataset. Currently, the web service is running in a private domain. ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 127, |
|
"end": 135, |
|
"text": "Figure 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "A Web Service", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "A svm-based ensemble approach to multi-document summarization", |
|
"authors": [ |
|
{ |
|
"first": "Yllias", |
|
"middle": [], |
|
"last": "Chali", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Sadid", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Shafiq R Joty", |
|
"middle": [], |
|
"last": "Hasan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Canadian Conference on Artificial Intelligence", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "199--202", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Yllias Chali, Sadid A Hasan, and Shafiq R Joty. 2009. A svm-based ensemble approach to multi-document summarization. In Canadian Conference on Artificial Intelligence, pages 199-202. Springer.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "The Financial Narrative Summarisation Shared Task (FNS 2020)", |
|
"authors": [], |
|
"year": 2020, |
|
"venue": "The 1st Joint Workshop on Financial Narrative Processing and MultiLing Financial Summarisation (FNP-FNS 2020", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mahmoud El-Haj, Ahmed AbuRa'ed, Nikiforos Pittaras, and George Giannakopoulos. 2020. The Financial Nar- rative Summarisation Shared Task (FNS 2020). In The 1st Joint Workshop on Financial Narrative Processing and MultiLing Financial Summarisation (FNP-FNS 2020, Barcelona, Spain.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Multi-document summarization using support vector regression", |
|
"authors": [ |
|
{ |
|
"first": "Sujian", |
|
"middle": [], |
|
"last": "Li", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "You", |
|
"middle": [], |
|
"last": "Ouyang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Wei", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bin", |
|
"middle": [], |
|
"last": "Sun", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of DUC. Citeseer", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Sujian Li, You Ouyang, Wei Wang, and Bin Sun. 2007. Multi-document summarization using support vector regression. In Proceedings of DUC. Citeseer.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "ROUGE: A package for automatic evaluation of summaries", |
|
"authors": [ |
|
{ |
|
"first": "Chin-Yew", |
|
"middle": [], |
|
"last": "Lin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Text Summarization Branches Out", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "74--81", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74-81, Barcelona, Spain, July. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Textrank: Bringing order into text", |
|
"authors": [ |
|
{ |
|
"first": "Rada", |
|
"middle": [], |
|
"last": "Mihalcea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Tarau", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the 2004 conference on empirical methods in natural language processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "404--411", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rada Mihalcea and Paul Tarau. 2004. Textrank: Bringing order into text. In Proceedings of the 2004 conference on empirical methods in natural language processing, pages 404-411.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Glove: Global vectors for word representation", |
|
"authors": [ |
|
{ |
|
"first": "Jeffrey", |
|
"middle": [], |
|
"last": "Pennington", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Richard", |
|
"middle": [], |
|
"last": "Socher", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Christopher D", |
|
"middle": [], |
|
"last": "Manning", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1532--1543", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014. Glove: Global vectors for word represen- tation. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pages 1532-1543.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF1": { |
|
"num": null, |
|
"content": "<table/>", |
|
"html": null, |
|
"type_str": "table", |
|
"text": "Fold wise Extraction Statistics of FNS-2020 Shared Task Training Set" |
|
}, |
|
"TABREF3": { |
|
"num": null, |
|
"content": "<table/>", |
|
"html": null, |
|
"type_str": "table", |
|
"text": "" |
|
}, |
|
"TABREF4": { |
|
"num": null, |
|
"content": "<table><tr><td>represents parameters used for S-LSTM and T-LSTM</td></tr></table>", |
|
"html": null, |
|
"type_str": "table", |
|
"text": "" |
|
}, |
|
"TABREF5": { |
|
"num": null, |
|
"content": "<table><tr><td>3 Results and Analysis</td></tr></table>", |
|
"html": null, |
|
"type_str": "table", |
|
"text": "Extraction Statistics of Triples from Sentences from FNS-2020 Shared Task Dataset" |
|
}, |
|
"TABREF7": { |
|
"num": null, |
|
"content": "<table/>", |
|
"html": null, |
|
"type_str": "table", |
|
"text": "ROUGE-1 Score Comparison of All Implemented Models concerning Single Full Length Gold Summary and Single Limited Length Gold Summary" |
|
}, |
|
"TABREF8": { |
|
"num": null, |
|
"content": "<table><tr><td colspan=\"5\">) employing SUMM-TL-MUSE, LEXRANK-SUMMARY, SUMM-</td></tr><tr><td>BL-POLY, TEXTRANK-SUMMARY.</td><td/><td/><td/><td/></tr><tr><td>Model Name</td><td colspan=\"4\">ROUGE-1 F ROUGE-2 F ROUGE-L F ROUGE-SU F</td></tr><tr><td>12 Best Performing</td><td>0.466</td><td>0.306</td><td>0.456</td><td>0.318</td></tr><tr><td>NN</td><td>0.445</td><td>0.246</td><td>0.318</td><td>0.242</td></tr><tr><td>S-LSTM</td><td>0.438</td><td>0.243</td><td>0.317</td><td>0.245</td></tr><tr><td>SVM</td><td>0.438</td><td>0.247</td><td>0.312</td><td>0.248</td></tr><tr><td>SUMM-TL-MUSE</td><td>0.433</td><td>0.234</td><td>0.407</td><td>0.253</td></tr><tr><td>LEXRANK-SUMMARY</td><td>0.264</td><td>0.12</td><td>0.218</td><td>0.14</td></tr><tr><td>SUMM-BL-POLY</td><td>0.274</td><td>0.105</td><td>0.205</td><td>0.135</td></tr><tr><td>TEXTRANK-SUMMARY</td><td>0.172</td><td>0.07</td><td>0.206</td><td>0.079</td></tr></table>", |
|
"html": null, |
|
"type_str": "table", |
|
"text": "" |
|
}, |
|
"TABREF9": { |
|
"num": null, |
|
"content": "<table/>", |
|
"html": null, |
|
"type_str": "table", |
|
"text": "ROUGE Score Comparison of KG-based Approaches with Top Scoring Approaches and Baseline Approaches, on Test Set, Computed by the Organizers using the Gold Summary" |
|
} |
|
} |
|
} |
|
} |