|
{ |
|
"paper_id": "R19-1038", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:01:33.443851Z" |
|
}, |
|
"title": "Song Lyrics Summarization Inspired by Audio Thumbnailing", |
|
"authors": [ |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Fell", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "CNRS", |
|
"location": { |
|
"postCode": "I3S", |
|
"region": "Inria", |
|
"country": "France" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Elena", |
|
"middle": [], |
|
"last": "Cabrio", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "CNRS", |
|
"location": { |
|
"postCode": "I3S", |
|
"region": "Inria", |
|
"country": "France" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Fabien", |
|
"middle": [], |
|
"last": "Gandon", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "CNRS", |
|
"location": { |
|
"postCode": "I3S", |
|
"region": "Inria", |
|
"country": "France" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Alain", |
|
"middle": [], |
|
"last": "Giboin", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "CNRS", |
|
"location": { |
|
"postCode": "I3S", |
|
"region": "Inria", |
|
"country": "France" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "Given the peculiar structure of songs, applying generic text summarization methods to lyrics can lead to the generation of highly redundant and incoherent text. In this paper, we propose to enhance state-of-the-art text summarization approaches with a method inspired by audio thumbnailing. Instead of searching for the thumbnail clues in the audio of the song, we identify equivalent clues in the lyrics. We then show how these summaries that take into account the audio nature of the lyrics outperform the generic methods according to both an automatic evaluation and human judgments.", |
|
"pdf_parse": { |
|
"paper_id": "R19-1038", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "Given the peculiar structure of songs, applying generic text summarization methods to lyrics can lead to the generation of highly redundant and incoherent text. In this paper, we propose to enhance state-of-the-art text summarization approaches with a method inspired by audio thumbnailing. Instead of searching for the thumbnail clues in the audio of the song, we identify equivalent clues in the lyrics. We then show how these summaries that take into account the audio nature of the lyrics outperform the generic methods according to both an automatic evaluation and human judgments.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Automatic text summarization is the task of producing a concise and fluent summary while preserving key information content and overall meaning of a text (Allahyari et al., 2017) . Numerous approaches have been developed to address this task and applied widely in various domains including news articles (Cheng and Lapata, 2016) , scientific papers (Mei and Zhai, 2008) , web content as blogs (Hu et al., 2007) , customer reviews (Pecar, 2018) and social media messages (He and Duan, 2018) . Just as we may need to summarize a story, we may also need to summarize song lyrics, for instance to produce adequate snippets for a search engine dedicated to an online song collection or for music digital libraries. From a linguistic point of view however, lyrics are a very peculiar genre of document and generic summarization methods may not be appropriate when the input for summarization comes from a specific domain or type of genre as songs are (Nenkova et al., 2011) . Compared to news documents, for instance, lyrics have a very different structure. Given the repeating forms, peculiar structure (e.g. the segmentation into verse, chorus, etc.) and other unique characteristics of song lyrics, we need the summariza-tion algorithms to take advantage of these additional elements to more accurately identify relevant information in song lyrics. But just as such characteristics enable the exploration of new approaches, other characteristics make the application of summarization algorithms very challenging, as the presence of repeated lines, the discourse structure that strongly depends on the interrelation of music and words in the melody composition, the heterogeneity of musical genres each featuring peculiar styles and wording (Brackett, 1995) , and simply the fact that not all songs tell a story.", |
|
"cite_spans": [ |
|
{ |
|
"start": 154, |
|
"end": 178, |
|
"text": "(Allahyari et al., 2017)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 304, |
|
"end": 328, |
|
"text": "(Cheng and Lapata, 2016)", |
|
"ref_id": "BIBREF7" |
|
}, |
|
{ |
|
"start": 349, |
|
"end": 369, |
|
"text": "(Mei and Zhai, 2008)", |
|
"ref_id": "BIBREF24" |
|
}, |
|
{ |
|
"start": 393, |
|
"end": 410, |
|
"text": "(Hu et al., 2007)", |
|
"ref_id": "BIBREF16" |
|
}, |
|
{ |
|
"start": 430, |
|
"end": 443, |
|
"text": "(Pecar, 2018)", |
|
"ref_id": "BIBREF33" |
|
}, |
|
{ |
|
"start": 470, |
|
"end": 489, |
|
"text": "(He and Duan, 2018)", |
|
"ref_id": "BIBREF14" |
|
}, |
|
{ |
|
"start": 945, |
|
"end": 967, |
|
"text": "(Nenkova et al., 2011)", |
|
"ref_id": "BIBREF28" |
|
}, |
|
{ |
|
"start": 1737, |
|
"end": 1753, |
|
"text": "(Brackett, 1995)", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this direction, this paper focuses on the following research questions: What is the impact of the context in summarizing song lyrics?. This question is broken down into two sub questions: 1) How do generic text summarization methods perform over lyrics? and 2) Can such peculiar context be leveraged to identify relevant sentences to improve song text summarization? To answer our research questions, we experiment with generic unsupervised state-of-the-art text summarization methods (i.e. TextRank, and a topic distribution based method) to perform lyrics summarization, and show that adding contextual information helps such models to produce better summaries. Specifically, we enhance text summarization approaches with a method inspired by audio thumbnailing techniques, that leverages the repetitive structure of song texts to improve summaries. We show how summaries that take into account the audio nature of the lyrics outperform the generic methods according to both an automatic evaluation over 50k lyrics, and judgments of 26 human subjects.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In the following, Section 2 reports on related work. Section 3 presents the lyrics summarization task and the proposed methods. Sections 4 and 5 report on the experiments and on the evaluation, respectively. Section 6 concludes the paper.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "This section reports on the related work on both text and audio summarization methods.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Summarization Methods", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "In the literature, there are two different families of approaches for automatic text summarization: extraction and abstraction (Allahyari et al., 2017) . Extractive summarization methods identify important elements of the text and generate them verbatim (they depend only on extraction of sentences or words from the original text). In contrast, abstractive summarization methods interpret and examine the text to generate a new shorter text that conveys the most critical information from the original text. Even though summaries created by humans are usually not extractive, most of the summarization research has focused on extractive methods. Purely extractive summaries often give better results (Nallapati et al., 2016) , due to the fact that latter methods cope with more complex problems such as semantic representation, inference and natural language generation. Existing abstractive summarizers often rely on an extractive pre-processing component to produce the abstract of the text (Berg-Kirkpatrick et al., 2011; Knight and Marcu, 2000) . Consequently, in this paper we focus on extractive summarization methods, also given the fact that lyrics i) strongly use figurative language which makes abstractive summarization even more challenging; and ii) the choice of the words by the composer may also have an importance for capturing the style of the song.", |
|
"cite_spans": [ |
|
{ |
|
"start": 127, |
|
"end": 151, |
|
"text": "(Allahyari et al., 2017)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 701, |
|
"end": 725, |
|
"text": "(Nallapati et al., 2016)", |
|
"ref_id": "BIBREF27" |
|
}, |
|
{ |
|
"start": 994, |
|
"end": 1025, |
|
"text": "(Berg-Kirkpatrick et al., 2011;", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 1026, |
|
"end": 1049, |
|
"text": "Knight and Marcu, 2000)", |
|
"ref_id": "BIBREF19" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Text Summarization", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "In the following, we focus on unsupervised methods for text summarization, the ones targeted in our study (no available gold-standard of humanproduced summaries of song texts exists). Most methods have in common the process for summary generation: given a text, the importance of each sentence of that text is determined. Then, the sentences with highest importance are selected to form a summary. The ways different summarizers determine the importance of each sentence may differ: Statistics-based summarizers extract indicator features from each sentence, e.g. (Fattah and Ren, 2009) use among others the sentence position and length and named entities as features. Topicbased summarizers aim to represent each sentence by its underlying topics. For instance, (Hennig, 2009) apply Probabilistic Latent Semantic Analysis, while Latent Dirichlet Allocation is used in (Arora and Ravindran, 2008) to model each sentence's distribution over latent topics. Another type of summarization methods is graph-based summarizers. Three of the most popular graphbased summarizers are TextRank (Mihalcea and Tarau, 2004) , LexRank (Erkan and Radev, 2004) , and . These methods work by constructing a graph whose nodes are sentences and whose graph edge weights are sentence similarities. Then, the sentences that are central to the graph are found by computing the PageRank (Page et al., 1999) . Contrarily to all previously described methods, systems using supervised machine learning form another type of summarizers. For instance, (Fattah, 2014) treats extractive summarization as a binary classification task, where they extract indicator features from sentences of gold summaries and learn to detect the sentences that should be included in a summary.", |
|
"cite_spans": [ |
|
{ |
|
"start": 564, |
|
"end": 586, |
|
"text": "(Fattah and Ren, 2009)", |
|
"ref_id": "BIBREF12" |
|
}, |
|
{ |
|
"start": 763, |
|
"end": 777, |
|
"text": "(Hennig, 2009)", |
|
"ref_id": "BIBREF15" |
|
}, |
|
{ |
|
"start": 869, |
|
"end": 896, |
|
"text": "(Arora and Ravindran, 2008)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 1083, |
|
"end": 1109, |
|
"text": "(Mihalcea and Tarau, 2004)", |
|
"ref_id": "BIBREF26" |
|
}, |
|
{ |
|
"start": 1120, |
|
"end": 1143, |
|
"text": "(Erkan and Radev, 2004)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 1363, |
|
"end": 1382, |
|
"text": "(Page et al., 1999)", |
|
"ref_id": "BIBREF30" |
|
}, |
|
{ |
|
"start": 1523, |
|
"end": 1537, |
|
"text": "(Fattah, 2014)", |
|
"ref_id": "BIBREF11" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Text Summarization", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Context-Specific Summarization. If specific knowledge about the application scenario or the domain of the summarized text is available, generic summarization methods can be adapted to take into account the prior information. In querybased summarization (Otterbacher et al., 2005; Wang et al., 2016) , the user's query is taken into account when generating a summary. Summarization of a scientific paper can be improved by considering the citations of it, as in (Delort et al., 2003) . However, to the best of our knowledge no summarization methods have been proposed for the domain of song texts. In this paper we present a summarization method that uses prior knowledge about the text it summarizes to help generic summarizers generate better summaries.", |
|
"cite_spans": [ |
|
{ |
|
"start": 253, |
|
"end": 279, |
|
"text": "(Otterbacher et al., 2005;", |
|
"ref_id": "BIBREF29" |
|
}, |
|
{ |
|
"start": 280, |
|
"end": 298, |
|
"text": "Wang et al., 2016)", |
|
"ref_id": "BIBREF39" |
|
}, |
|
{ |
|
"start": 461, |
|
"end": 482, |
|
"text": "(Delort et al., 2003)", |
|
"ref_id": "BIBREF8" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Text Summarization", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Evaluation Criteria and Methods. Summaries should i) contain the most important information from input documents, ii) not contain redundant information, iii) be readable, hence they should be grammatical and coherent . While a multitude of methods to identify important sentences has been described above, several approaches aim to make summaries less redundant and more coherent. The simplest way to evaluate summaries is to let humans assess the quality, but this is extremely expensive. The factors that humans must consider when giving scores to each candidate summary are grammaticality, non redundancy, integration of most important pieces of information, structure and coherence (Saggion and Poibeau, 2013) . The more common way is to let humans generate possibly multiple summaries for a text and then automatically assess how close a machine-made summary is to the human gold summaries computing ROUGE scores (Lin, 2004) , which boils down to measuring ngram overlaps between gold summaries and automatic summary. More recently there have been attempts to rate summaries automatically without the need for gold summaries (Nenkova et al., 2011) . The key idea is that a summary should be similar to the original text in regard to characteristic criteria as the word distribution. (Mackie et al., 2014) find that topic words are a suitable metric to automatically evaluate micro blog summaries.", |
|
"cite_spans": [ |
|
{ |
|
"start": 686, |
|
"end": 713, |
|
"text": "(Saggion and Poibeau, 2013)", |
|
"ref_id": "BIBREF36" |
|
}, |
|
{ |
|
"start": 918, |
|
"end": 929, |
|
"text": "(Lin, 2004)", |
|
"ref_id": "BIBREF21" |
|
}, |
|
{ |
|
"start": 1130, |
|
"end": 1152, |
|
"text": "(Nenkova et al., 2011)", |
|
"ref_id": "BIBREF28" |
|
}, |
|
{ |
|
"start": 1288, |
|
"end": 1309, |
|
"text": "(Mackie et al., 2014)", |
|
"ref_id": "BIBREF23" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Text Summarization", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "Lyrics are texts that accompany music. Therefore, it is worthwhile to see if methods in audio summarization can be transferred to lyrics summarization. In audio summarization the goal is to find the most representative parts in a song, in Pop songs those are usually the chorus and the bridge, in instrumental music the main theme. The task of creating short audio summaries is also known as audio thumbnailing (Bartsch and Wakefield, 2005; Chai and Vercoe, 2003; Levy et al., 2006) , as the goal is to produce a short representation of the music that fits onto a thumbnail, but still covers the most representative parts of it. In a recent approach of audio thumbnailing (Jiang and M\u00fcller, 2015) , the authors generate a Double Thumbnail from a musical piece by finding the two most representative parts in it. For this, they search for candidate musical segments in an a priori unsegmented song. Candidate musical segments are defined as sequences of music that more or less exactly repeat themselves. The representativeness of each candidate segment to the whole piece is then estimated by their fitness metric. They define the fitness of a segment as a trade-off between how exactly a part is repeated and how much of the whole piece is covered by all repetitions of that segment. Then, the audio segments along with their fitness allow them to create an audio double thumbnail consisting of the two fittest audio segments.", |
|
"cite_spans": [ |
|
{ |
|
"start": 411, |
|
"end": 440, |
|
"text": "(Bartsch and Wakefield, 2005;", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 441, |
|
"end": 463, |
|
"text": "Chai and Vercoe, 2003;", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 464, |
|
"end": 482, |
|
"text": "Levy et al., 2006)", |
|
"ref_id": "BIBREF20" |
|
}, |
|
{ |
|
"start": 672, |
|
"end": 696, |
|
"text": "(Jiang and M\u00fcller, 2015)", |
|
"ref_id": "BIBREF17" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Audio Summarization", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "Song texts are arranged in segments and lines. For instance the song text depicted in Figure 1 consists of 8 segments and 38 lines. Given a song text S consisting of n lines of text, S = (x 1 , ..., x n ), we define the task of extractive lyrics summarization as the task of producing a concise summary sum of the song text, consisting of a subset of the original text lines: sum(S) \u2286 S, where usually |sum(S)| << |S|. We define the goal of a summary as to preserve key information and the overall meaning of a song text. To address this task, we apply the following methods from the literature: the popular graph-based summarizer TextRank; an adaptation of a topic-based method (TopSum). Moreover, we introduce a method inspired by audio thumbnailing (which we dub Lyrics Thumbnail) which aims at creating a summary from the most representative parts of the original song text. While for TextRank we rely on the off-the-shelf implementation of (Barrios et al., 2016) , in the following we describe the other two methods.", |
|
"cite_spans": [ |
|
{ |
|
"start": 945, |
|
"end": 967, |
|
"text": "(Barrios et al., 2016)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 86, |
|
"end": 94, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Lyrics Summarization", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We implement a simple topic-based summarization model that aims to construct a summary whose topic distribution is as similar as possible to that of the original text. Following (Kleedorfer et al., 2008) , we train a topic model by factorizing a tf-idf-weighted term-document matrix of a song text corpus (see Section 4.2) using non-negative matrix factorization into a term-topic and a topicdocument matrix. Given the learnt term-topic matrix, we compute a topic vector t for each new document (song text). In order to treat t as a (pseudo-) probability distribution over latent topics t i , we normalize t by applying \u03bbt.t/ t i \u2208t t i to it. Given the distributions over latent topics for each song text, we then incrementally construct a summary by greedily adding one line from the original text at a time (same mechanism as in KLSum algorithm in (Haghighi and Vanderwende, 2009) ); that line x * of the original text that minimizes the distance between the topic distribution t S of the original text S and the topic distribution of the incremental summary sum(S):", |
|
"cite_spans": [ |
|
{ |
|
"start": 178, |
|
"end": 203, |
|
"text": "(Kleedorfer et al., 2008)", |
|
"ref_id": "BIBREF18" |
|
}, |
|
{ |
|
"start": 851, |
|
"end": 883, |
|
"text": "(Haghighi and Vanderwende, 2009)", |
|
"ref_id": "BIBREF13" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TopSum", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "x * = argmin x\u2208(S\\sum(S)) {W (t S , t sum(S)+x )}", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TopSum", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "W is the Wasserstein distance (Villani, 2008) and is used to measure the distance between two probability distributions (an alternative to Jensen-Shannon divergence (Louis and Nenkova, 2013) ). ", |
|
"cite_spans": [ |
|
{ |
|
"start": 30, |
|
"end": 45, |
|
"text": "(Villani, 2008)", |
|
"ref_id": "BIBREF38" |
|
}, |
|
{ |
|
"start": 165, |
|
"end": 190, |
|
"text": "(Louis and Nenkova, 2013)", |
|
"ref_id": "BIBREF22" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "TopSum", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "Inspired by (Jiang and M\u00fcller, 2015) , we transfer their fitness measure for audio segments to compute the fitness of lyrics segments. Analog to an audio thumbnail, we define a Lyrics Thumbnail as the most representative and repetitive part of the song text. Consequently, it usually consists of (a part of) the chorus. In our corpus the segments are annotated (as double line breaks in the lyrics), so unlike in audio thumbnailing, we do not have to induce segments, but rather measure their fitness. In the following, we describe the fitness measure for lyrics segments and how we use this to produce a summary of the lyrics.", |
|
"cite_spans": [ |
|
{ |
|
"start": 12, |
|
"end": 36, |
|
"text": "(Jiang and M\u00fcller, 2015)", |
|
"ref_id": "BIBREF17" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Lyrics Fitness Given a segmented song text S = (S 1 , ..., S m ) consisting of text segments S i , where each S i consists of |S i | text lines, we cluster the S i into partitions of similar segments. For instance, the lyrics in Figure 1 consists of 8 segments and 38 lines and the cluster of chorus consists of {S 5 , S 6 , S 7 }. The fitness F it of the segment cluster C \u2286 S is defined through the precision pr of the cluster and the coverage co of the cluster. pr describes how similar the segments in C are to each other while co is the relative amount of lyrics lines covered by C:", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 229, |
|
"end": 237, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "pr(C) = ( S i ,S j \u2208C i<j 1) \u22121 \u2022 S i ,S j \u2208C i<j sim(S i , S j ) co(C) = ( S i \u2208S |S i |) \u22121 \u2022 S i \u2208C |S i |", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "where sim is a normalized similarity measure between text segments. F it is the harmonic mean between pr and co. The fitness of a segment S i is defined as the fitness of the cluster to which S i belongs:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "\u2200S i \u2208 C : F it(S i ) = F it(C) = 2 pr(C) \u2022 co(C) pr(C) + co(C)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "For lyrics segments without repetition the fitness is defined as zero. Based on the fitness F it for segments, we define a fitness measure for a text line x. This allows us to compute the fitness of arbitrary summaries (with no or unknown segmentation). If the text line x occurs f i (x) times in text segment S i , then its line fitness f it is defined as:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "f it(x) = ( S i \u2208S f i (x)) \u22121 \u2022 S i \u2208S f i (x) \u2022 F it(S i )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Fitness-Based Summary Analog to (Jiang and M\u00fcller, 2015) 's audio thumbnails, we create fitness-based summaries for a song text. A Lyrics Double Thumbnail consists of two segments: one from the fittest segment cluster (usually the chorus), and one from the second fittest segment cluster (usually the bridge). 1 If the second fittest cluster has a fitness of 0, we generate a Lyrics Single Thumbnail solely from the fittest cluster (usually the chorus). If the thumbnail generated has a length of k lines and we want to produce a summary of p < k lines, we select the p lines in the middle of the thumbnail following (Chai and Vercoe, 2003) 's \"Section-transition Strategy\" that they find to capture the \"hook\" of the music more likely. 2", |
|
"cite_spans": [ |
|
{ |
|
"start": 32, |
|
"end": 56, |
|
"text": "(Jiang and M\u00fcller, 2015)", |
|
"ref_id": "BIBREF17" |
|
}, |
|
{ |
|
"start": 617, |
|
"end": 640, |
|
"text": "(Chai and Vercoe, 2003)", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Lyrics Thumbnail", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "We now describe the WASABI dataset of song lyrics (Section 4.1), and the tested configurations of the summarization methods (Section 4.2).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Setting", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "From the WASABI corpus (Meseguer-Brocal et al., 2017) we select a subset of 190k unique song texts with available genre information. As the corpus has spurious genres (416 different ones), we focus on the 10 most frequent ones in order to evaluate our methods dependent on the genre. We add 2 additional genres from the underrepresented Rap field (Southern Hip Hop and Gangsta Rap). The dataset contains 95k song lyrics.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Dataset", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "To define the length of sum(S) (see Section 3), we rely on (Bartsch and Wakefield, 2005) that recommend to create audio thumbnails of the median length of the chorus on the whole corpus. We therefore estimate the median chorus length on our corpus by computing a Lyrics Single Thumbnail on each text, and we find the median chorus length to be 4 lines. Hence, we decide to generate summaries of such length for all lyrics and all summarization models to exclude the length bias in the methods comparison 3 . As the length of the lyrics thumbnail is lower-bounded by the length of the chorus in the song text, we keep only those lyrics with an estimated chorus length of at least 4. The final corpus of 12 genres consists of 50k lyrics with the following genre distribution: Rock: 8.4k, Country: 8.3k, Alternative Rock: 6.6k, Pop: 6.9k, R&B: 5.2k, Indie Rock: 4.4k, Hip Hop: 4.2k, Hard Rock: 2.4k, Punk Rock: 2k, Folk: 1.7k, Southern Hip Hop: 281, Gangsta Rap: 185.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Dataset", |
|
"sec_num": "4.1" |
|
}, |
|
{ |
|
"text": "We create summaries using the three summarization methods described in Section 3, i.e. a graphbased (TextRank), a topic-based (TopSum), and fitness-based (Lyrics Thumbnail) method, plus two additional combined models (described below). While the Lyrics Thumbnail is generated from the full segment structure of the lyrics including its duplicate lines, all other models are fed with unique text lines as input (i.e. rendundant lines are deleted). This is done to produce less redundant summaries, given that for instance, Tex-tRank scores each duplicate line the same, hence it may create summaries with all identical lines. TopSum can suffer from a similar shortcoming: if there is a duplicate line close to the ideal topic distribution, adding that line again will let the incremental summary under construction stay close to the ideal topic distribution. All models were instructed to produce summaries of 4 lines, as this is the estimated median chorus length in our corpus (see Section 4.1). The summary lines were arranged in the same order they appear in the original text. 4 We use the TextRank implementation 5 of (Barrios et al., 2016) without removing stop words (lyrics lines in input can be quite short, therefore we avoid losing all content of the line if removing stop words). The topic model for Top-Sum is built using non-negative matrix factorization with scikit-learn 6 (Pedregosa et al., 2011) for 30 topics on the full corpus of 190k lyrics. 7 For the topical distance, we only consider the distance between the 3 most relevant topics in the original text, following the intuition that one song text usually covers only a small amount of topics. The Lyrics Thumbnail is computed using String-based distance between text segments to facilitate clustering. This similarity has been shown in (Watanabe et al., 2016) to indicate segment borders successfully. In our implementation, segments are clustered using the DBSCAN (Ester et al., 1996) algorithm. 8 We also produce two summaries by combining TextRank + TopSum and TextRank + TopSum + Lyrics Thumbnail, to test if summaries can benefit from the complementary perspectives the three different summarization methods take.", |
|
"cite_spans": [ |
|
{ |
|
"start": 1081, |
|
"end": 1082, |
|
"text": "4", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1123, |
|
"end": 1145, |
|
"text": "(Barrios et al., 2016)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 1389, |
|
"end": 1413, |
|
"text": "(Pedregosa et al., 2011)", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1463, |
|
"end": 1464, |
|
"text": "7", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 1810, |
|
"end": 1833, |
|
"text": "(Watanabe et al., 2016)", |
|
"ref_id": "BIBREF40" |
|
}, |
|
{ |
|
"start": 1939, |
|
"end": 1959, |
|
"text": "(Ester et al., 1996)", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 1971, |
|
"end": 1972, |
|
"text": "8", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Models and Configurations", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Model Combination For any lyrics line, we can obtain a score from each of the applied methods. TextRank provides a score for each line, TopSum provides a distance between the topic distributions of an incremental summary and the original text, and f it provides the fitness of each line. We treat our summarization methods as blackboxes and use a simple method to combine the scores the different methods provide for each line. Given the original text separated into lines S = (x 1 , ..., x n ), a summary is constructed by greedily adding one line x * at a time to the incremental summary sum(S) \u2286 S such that the sum of normalized ranks of all scores is minimal:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Models and Configurations", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "x * = argmin x { A R A (x)}", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Models and Configurations", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Here x \u2208 (S \\ sum(S)) and A \u2208 {TextRank, TopSum, fit}.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Models and Configurations", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "The normalized rank R A (x) of the score that method A assigns to line x is computed as follows: first, the highest scores 9 are assigned rank 0, the second highest scores get rank 1, and so forth. Then the ranks are linearly scaled to the [0,1] interval, so each sum of ranks A R A (x) is in [0, 3] .", |
|
"cite_spans": [ |
|
{ |
|
"start": 293, |
|
"end": 296, |
|
"text": "[0,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 297, |
|
"end": 299, |
|
"text": "3]", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Models and Configurations", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "Model Nomenclature For abbreviation, we call the TextRank model henceforth M r , the Top-Sum model M s , the fitness-based summarizer M f , model combinations M rs and M rsf , respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Models and Configurations", |
|
"sec_num": "4.2" |
|
}, |
|
{ |
|
"text": "We evaluate the quality of the produced lyrics summary both soliciting human judgments on the goodness and utility of a given summary (Section 5.1), and through an automatic evaluation of the summarization methods (Section 5.2) to provide a comprehensive evaluation.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Evaluation", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "We performed human evaluation of the different summarization methods introduced before by asking participants to rate the different summaries presented to them by specifying their agreement / disagreement according to the following standard criteria :", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Human Evaluation", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "Informativeness: The summary contains the main points of the original song text.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Human Evaluation", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "Non-redundancy: The summary does not contain duplicate or redundant information.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Human Evaluation", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "Coherence: The summary is fluent to read and grammatically correct. Plus one additional criterion coming from our definition of the lyrics summarization task:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Human Evaluation", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "Meaning: The summary preserves the meaning of the original song text.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Human Evaluation", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "An experimental psychologist expert in Human Computer Interaction advised us in defining the questionnaire and setting up the experiment. 26 participants -12 nationalities, 18 men, 8 women, aged from 21 to 59 -were taking a questionnaire (Google Forms), consisting of rating 30 items with respect to the criteria defined before on a Likert scale from 1 (low) to 5 (high). Each participant was presented with 5 different summaries -each produced by one of the previously described summarization models -for 6 different song texts. Participants were given example ratings for the different criteria in order to familiarize them with the procedure. Then, for each song text, the original song text along with its 5 summaries were presented in random order and had to be rated according to the above criteria. For the criterion of Meaning, we asked participants to give a short explanation in free text for their score. The selected 6 song texts 10 have a minimum and a median chorus length of 4 lines and are from different genres, i.e. Pop/Rock (4), Folk (1) and Rap (1), similar to our corpus genre distribution. Song texts were selected from different lengths (18-63 lines), genders of singer (3 male, 3 female), topics (family, life, drugs, relationship, depression), and mood (depressive, angry, hopeful, optimistic, energetic). The artist name and song title were not shown to the participants.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Human Evaluation", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "Results Figure 2 shows the ratings obtained for each criterion. We examine the significant differences between the models performances by performing a paired two-tailed t-test. The significance levels are: 0.05 * , 0.01 * * , 0.001 * * * , and n.s. First, Informativeness and Meaning are rated higher * * for the combined model M rs compared to the single models M r and M s . Combining all three models improves the summaries further: both for Informativeness and Meaning the model M rsf is rated higher * * * than M rs . Further, summaries created by M rsf are rated higher * * * in Coherence than summaries from any other modelexcept from M f (n.s. difference). Summaries are rated on the same level (n.s. differences) for Non-redundancy in all but the M r and M f summaries, which are perceived as lower * * * in Nonredundancy than all others. Note, how the model M rsf is more stable than all others by exhibiting lower standard deviations in all criteria except Overall, leveraging the Lyrics Fitness in a song text summary improves summary quality. Especially with respect to the criteria that, we believe, indicate the summary quality the most -Informativeness and Meaning -the M rsf method is significantly better performing and more consistent. Figure 1 shows an example song text and example summaries from the experiment. Summary 1 is generated by M f and consists of the chorus. Summary 2 is made by the method M rsf and has relevant parts of the verses and the chorus, and was rated much higher in Informativeness and Meaning. We analyzed the free text written by the participants to comment on the Meaning criterion, but no relevant additional information was provided (the participants mainly summarized their ratings).", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 8, |
|
"end": 16, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
}, |
|
{ |
|
"start": 1255, |
|
"end": 1263, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Human Evaluation", |
|
"sec_num": "5.1" |
|
}, |
|
{ |
|
"text": "We computed four different indicators of summary quality on the dataset of 50k songs described in Section 4.1. Three of the criteria use the similarity between probability distributions P, Q, which means we compute the Wasserstein distance between P and Q (cf. Section 3.1) and apply \u03bbx. x \u22121 to it. 11 The criteria are:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "Distributional Semantics: similarity between the word distributions of original and summary, cf. (Louis and Nenkova, 2013) . We give results relative to the similarity of the best performing model (=100%).", |
|
"cite_spans": [ |
|
{ |
|
"start": 97, |
|
"end": 122, |
|
"text": "(Louis and Nenkova, 2013)", |
|
"ref_id": "BIBREF22" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "Topical: similarity between the topic distributions of original and summary. Restricted to the 3 most relevant topics of the original song text. We give results relative to the similarity of the best performing model (=100%).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "Coherence: average similarity between word distributions in consecutive sentences of the summary, cf. (ShafieiBavani et al., 2018) . We give results relative to the coherence of the original song text (=100%).", |
|
"cite_spans": [ |
|
{ |
|
"start": 102, |
|
"end": 130, |
|
"text": "(ShafieiBavani et al., 2018)", |
|
"ref_id": "BIBREF37" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "Lyrics fitness: average line-based fitness f it (cf. Section 3) of the lines in the summary. We give results relative to the Lyrics fitness of the original song text (=100%).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "Results When evaluating each of the 12 genres, we found two clusters of genres to behave very similarly. Therefore, we report the results for these two groups: the Rap genre cluster contains Hip Hop, Southern Hip Hop, and Gangsta Rap. The Rock / Pop cluster contains the 9 other genres. Results of the different automatic evaluation metrics are shown in Table 1 . Distributional Semantics metrics have previously been shown (Louis and Nenkova, 2013; ShafieiBavani et al., 2018) to highly correlate with user responsiveness judgments. We would expect correlations of this metric with Informativeness or Meaning criteria therefore, as those criteria are closest to responsiveness, but we have found no large differences between the different models for this criterion. The summaries of the M s model have the highest similarity to the original text and the M f have the lowest similarity of 90%. The difference between the highest and lowest values are low.", |
|
"cite_spans": [ |
|
{ |
|
"start": 424, |
|
"end": 449, |
|
"text": "(Louis and Nenkova, 2013;", |
|
"ref_id": "BIBREF22" |
|
}, |
|
{ |
|
"start": 450, |
|
"end": 477, |
|
"text": "ShafieiBavani et al., 2018)", |
|
"ref_id": "BIBREF37" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 354, |
|
"end": 361, |
|
"text": "Table 1", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "For the Topical similarity, the results are mostly in the same order as the Distributional Semantics ones, but with much larger differences. While the M s model reaches the highest similarity, this is a self-fulfilling prophecy, as summaries of M s were generated with the objective of maximizing topical similarity. The other two models that incorporate M s (M rs and M rsf ), show a much higher topical similarity to the original text than M r and M f .", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "Coherence is rated best in M r with 110%. All other models show a coherence close to that of the original text -between 97% and 101%. We believe that the increased coherence of M r is not linguistically founded, but merely algorithmic. M r produces summaries of the most central sentences in a text. The centrality is using the concept of sentence similarity. Therefore, M r implicitly optimizes for the automatic evaluation metric of coherence, based on similar consecutive sentences. Sentence similarity seems to be insufficient to predict human judgments of coherence in this case.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "As might be expected, methods explicitly incorporating the Lyrics fitness produce summaries with a fitness much higher than the original text -214% for the M f and 191% for the M rsf model. The methods not incorporating fitness produce summaries with much lower fitness than the original -M r 62%, M s 47%, and M rs 55%. In the Rap genre this fitness is even zero, i.e. summaries (in median) contain no part of the chorus.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "Overall, no single automatic evaluation criterion was able to explain the judgments of our human participants. However, considering Topical similarity and fitness together gives us a hint. The model M f has high fitness (214%), but low Topical similarity (42%). The M s model has the highest Topical similarity (100%), but low fitness (47%). M rsf might be preferred by humans as it strikes a balance between Topical similarity (64%) and fitness (191%). Hence, M rsf succeeds in capturing lines from the most relevant parts of the lyrics, such as the chorus, while jointly representing the important topics of the song text.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Automatic Evaluation", |
|
"sec_num": "5.2" |
|
}, |
|
{ |
|
"text": "In this paper we have defined and addressed the task of lyrics summarization. We have applied both generic unsupervised text summarization methods (TextRank and a topic-based method we called TopSum), and a method inspired by audio thumbnailing on 50k lyrics from the WASABI corpus. We have carried out an automatic evaluation on the produced summaries computing standard metrics in text summarization, and a human evaluation with 26 participants, showing that using a fitness measure transferred from the musicology literature, we can amend generic text summarization algorithms and produce better summaries.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "6" |
|
}, |
|
{ |
|
"text": "In future work, we will model the importance of a line given the segment to avoid cutting off important parts of the chorus, as we sometimes observed. Moreover, we plan to address the challenging task of abstractive summarization over song lyrics, with the goal of creating a summary of song texts in prose-style -more similar to what humans would do, using their own words.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "6" |
|
}, |
|
{ |
|
"text": "We pick the first occurring representative of the segment cluster. Which segment to pick from the cluster is a potential question for future work.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "They also experiment with other methods to create a thumbnail, such as section initial or section ending.3 We leave the study of other measures to estimate the summary length to future work.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In case of repeated parts, the first position of each line was used as original position.5 https://github.com/summanlp/textrank 6 https://scikit-learn.org 7 loss='kullback-leibler' 8 eps=0.3, min samples=2", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In the case of topical distance, a \"higher score\" means a lower value.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "\"Pills N Potions\" by Nicki Minaj, \"Hurt\" by Nine Inch Nails, \"Real to me\" by Brian McFadden, \"Somebody That I Used To Know\" by Gotye, \"Receive\" by Alanis Morissette, \"Let's Start A Band\" by Amy MacDonald", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "This works as we always deal with distances > 0.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "This work is partly funded by the French Research National Agency (ANR) under the WASABI project (contract ANR-16-CE23-0017-01).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgement", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Text summarization techniques: A brief survey", |
|
"authors": [ |
|
{ |
|
"first": "Mehdi", |
|
"middle": [], |
|
"last": "Allahyari", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mehdi", |
|
"middle": [], |
|
"last": "Seyed Amin Pouriyeh", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Saeid", |
|
"middle": [], |
|
"last": "Assefi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Elizabeth", |
|
"middle": [ |
|
"D" |
|
], |
|
"last": "Safaei", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Trippe", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Juan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mehdi Allahyari, Seyed Amin Pouriyeh, Mehdi Assefi, Saeid Safaei, Elizabeth D. Trippe, Juan B. Gutier- rez, and Krys Kochut. 2017. Text summarization techniques: A brief survey. CoRR, abs/1707.02268.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Latent dirichlet allocation based multi-document summarization", |
|
"authors": [ |
|
{ |
|
"first": "Rachit", |
|
"middle": [], |
|
"last": "Arora", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Balaraman", |
|
"middle": [], |
|
"last": "Ravindran", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Proceedings of the second workshop on Analytics for noisy unstructured text data", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "91--97", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rachit Arora and Balaraman Ravindran. 2008. Latent dirichlet allocation based multi-document summa- rization. In Proceedings of the second workshop on Analytics for noisy unstructured text data, pages 91- 97. ACM.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Variations of the similarity function of textrank for automated summarization", |
|
"authors": [ |
|
{ |
|
"first": "Federico", |
|
"middle": [], |
|
"last": "Barrios", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Federico", |
|
"middle": [], |
|
"last": "L\u00f3pez", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luis", |
|
"middle": [], |
|
"last": "Argerich", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rosa", |
|
"middle": [], |
|
"last": "Wachenchauzer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Federico Barrios, Federico L\u00f3pez, Luis Argerich, and Rosa Wachenchauzer. 2016. Variations of the simi- larity function of textrank for automated summariza- tion. CoRR, abs/1602.03606.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Audio thumbnailing of popular music using chromabased representations", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Mark", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Gregory", |
|
"middle": [ |
|
"H" |
|
], |
|
"last": "Bartsch", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Wakefield", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Trans. Multi", |
|
"volume": "7", |
|
"issue": "1", |
|
"pages": "96--104", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1109/TMM.2004.840597" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mark A. Bartsch and Gregory H. Wakefield. 2005. Au- dio thumbnailing of popular music using chroma- based representations. Trans. Multi., 7(1):96-104.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Jointly learning to extract and compress", |
|
"authors": [ |
|
{ |
|
"first": "Taylor", |
|
"middle": [], |
|
"last": "Berg-Kirkpatrick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dan", |
|
"middle": [], |
|
"last": "Gillick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dan", |
|
"middle": [], |
|
"last": "Klein", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "481--490", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Taylor Berg-Kirkpatrick, Dan Gillick, and Dan Klein. 2011. Jointly learning to extract and compress. In Proceedings of the 49th Annual Meeting of the Asso- ciation for Computational Linguistics: Human Lan- guage Technologies -Volume 1, HLT '11, pages 481-490, Stroudsburg, PA, USA. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Interpreting Popular Music", |
|
"authors": [ |
|
{ |
|
"first": "David", |
|
"middle": [], |
|
"last": "Brackett", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1995, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "David Brackett. 1995. Interpreting Popular Music. Cambridge University Press.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Music thumbnailing via structural analysis", |
|
"authors": [ |
|
{ |
|
"first": "Wei", |
|
"middle": [], |
|
"last": "Chai", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Barry", |
|
"middle": [], |
|
"last": "Vercoe", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the eleventh ACM international conference on Multimedia", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "223--226", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1145/957013.957057" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wei Chai and Barry Vercoe. 2003. Music thumbnail- ing via structural analysis. In Proceedings of the eleventh ACM international conference on Multime- dia, pages 223-226.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Neural summarization by extracting sentences and words", |
|
"authors": [ |
|
{ |
|
"first": "Jianpeng", |
|
"middle": [], |
|
"last": "Cheng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mirella", |
|
"middle": [], |
|
"last": "Lapata", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "484--494", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/P16-1046" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jianpeng Cheng and Mirella Lapata. 2016. Neural summarization by extracting sentences and words. In Proceedings of the 54th Annual Meeting of the As- sociation for Computational Linguistics (Volume 1: Long Papers), pages 484-494. Association for Com- putational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Enhanced web document summarization using hyperlinks", |
|
"authors": [ |
|
{ |
|
"first": "Yves", |
|
"middle": [], |
|
"last": "Jean", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bernadette", |
|
"middle": [], |
|
"last": "Delort", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Maria", |
|
"middle": [], |
|
"last": "Bouchon-Meunier", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Rifqi", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "Proceedings of the Fourteenth ACM Conference on Hypertext and Hypermedia, HYPERTEXT '03", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "208--215", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1145/900051.900097" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jean Yves Delort, Bernadette Bouchon-Meunier, and Maria Rifqi. 2003. Enhanced web document sum- marization using hyperlinks. In Proceedings of the Fourteenth ACM Conference on Hypertext and Hy- permedia, HYPERTEXT '03, pages 208-215, New York, NY, USA. ACM.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Lexrank: Graph-based lexical centrality as salience in text summarization", |
|
"authors": [ |
|
{ |
|
"first": "G\u00fcnes", |
|
"middle": [], |
|
"last": "Erkan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Dragomir R Radev", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Journal of artificial intelligence research", |
|
"volume": "22", |
|
"issue": "", |
|
"pages": "457--479", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "G\u00fcnes Erkan and Dragomir R Radev. 2004. Lexrank: Graph-based lexical centrality as salience in text summarization. Journal of artificial intelligence re- search, 22:457-479.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "A density-based algorithm for discovering clusters in large spatial databases with noise", |
|
"authors": [ |
|
{ |
|
"first": "Martin", |
|
"middle": [], |
|
"last": "Ester", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hans-Peter", |
|
"middle": [], |
|
"last": "Kriegel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J\u00f6rg", |
|
"middle": [], |
|
"last": "Sander", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaowei", |
|
"middle": [], |
|
"last": "Xu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1996, |
|
"venue": "Kdd", |
|
"volume": "96", |
|
"issue": "", |
|
"pages": "226--231", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Martin Ester, Hans-Peter Kriegel, J\u00f6rg Sander, Xiaowei Xu, et al. 1996. A density-based algorithm for discovering clusters in large spatial databases with noise. In Kdd, volume 96, pages 226-231.", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "A hybrid machine learning model for multi-document summarization", |
|
"authors": [ |
|
{ |
|
"first": "Mohamed", |
|
"middle": [ |
|
"Abdel" |
|
], |
|
"last": "Fattah", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Applied intelligence", |
|
"volume": "40", |
|
"issue": "4", |
|
"pages": "592--600", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mohamed Abdel Fattah. 2014. A hybrid machine learning model for multi-document summarization. Applied intelligence, 40(4):592-600.", |
|
"links": null |
|
}, |
|
"BIBREF12": { |
|
"ref_id": "b12", |
|
"title": "Ga, mr, ffnn, pnn and gmm based models for automatic text summarization", |
|
"authors": [ |
|
{ |
|
"first": "Mohamed", |
|
"middle": [], |
|
"last": "", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Abdel", |
|
"middle": [], |
|
"last": "Fattah", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Fuji", |
|
"middle": [], |
|
"last": "Ren", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Comput. Speech Lang", |
|
"volume": "23", |
|
"issue": "1", |
|
"pages": "126--144", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1016/j.csl.2008.04.002" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mohamed Abdel Fattah and Fuji Ren. 2009. Ga, mr, ffnn, pnn and gmm based models for automatic text summarization. Comput. Speech Lang., 23(1):126- 144.", |
|
"links": null |
|
}, |
|
"BIBREF13": { |
|
"ref_id": "b13", |
|
"title": "Exploring content models for multi-document summarization", |
|
"authors": [ |
|
{ |
|
"first": "Aria", |
|
"middle": [], |
|
"last": "Haghighi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lucy", |
|
"middle": [], |
|
"last": "Vanderwende", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "362--370", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Aria Haghighi and Lucy Vanderwende. 2009. Explor- ing content models for multi-document summariza- tion. In Proceedings of Human Language Tech- nologies: The 2009 Annual Conference of the North American Chapter of the Association for Computa- tional Linguistics, pages 362-370. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF14": { |
|
"ref_id": "b14", |
|
"title": "Twitter summarization based on social network and sparse reconstruction", |
|
"authors": [ |
|
{ |
|
"first": "Ruifang", |
|
"middle": [], |
|
"last": "He", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xingyi", |
|
"middle": [], |
|
"last": "Duan", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "AAAI", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ruifang He and Xingyi Duan. 2018. Twitter summa- rization based on social network and sparse recon- struction. In AAAI.", |
|
"links": null |
|
}, |
|
"BIBREF15": { |
|
"ref_id": "b15", |
|
"title": "Topic-based multi-document summarization with probabilistic latent semantic analysis", |
|
"authors": [ |
|
{ |
|
"first": "Leonhard", |
|
"middle": [], |
|
"last": "Hennig", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2009, |
|
"venue": "Proceedings of the International Conference RANLP-2009", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "144--149", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Leonhard Hennig. 2009. Topic-based multi-document summarization with probabilistic latent semantic analysis. In Proceedings of the International Con- ference RANLP-2009, pages 144-149.", |
|
"links": null |
|
}, |
|
"BIBREF16": { |
|
"ref_id": "b16", |
|
"title": "Comments-oriented blog summarization by sentence extraction", |
|
"authors": [ |
|
{ |
|
"first": "Meishan", |
|
"middle": [], |
|
"last": "Hu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Aixin", |
|
"middle": [], |
|
"last": "Sun", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ee-Peng", |
|
"middle": [], |
|
"last": "Lim", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ee-Peng", |
|
"middle": [], |
|
"last": "Lim", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of the Sixteenth ACM Conference on Conference on Information and Knowledge Management, CIKM '07", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "901--904", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1145/1321440.1321571" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Meishan Hu, Aixin Sun, Ee-Peng Lim, and Ee-Peng Lim. 2007. Comments-oriented blog summarization by sentence extraction. In Proceedings of the Six- teenth ACM Conference on Conference on Informa- tion and Knowledge Management, CIKM '07, pages 901-904, New York, NY, USA. ACM.", |
|
"links": null |
|
}, |
|
"BIBREF17": { |
|
"ref_id": "b17", |
|
"title": "Estimating double thumbnails for music recordings", |
|
"authors": [ |
|
{ |
|
"first": "Nanzhu", |
|
"middle": [], |
|
"last": "Jiang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Meinard", |
|
"middle": [], |
|
"last": "M\u00fcller", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "146--150", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1109/ICASSP.2015.7177949" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Nanzhu Jiang and Meinard M\u00fcller. 2015. Estimat- ing double thumbnails for music recordings. In 2015 IEEE International Conference on Acous- tics, Speech and Signal Processing (ICASSP), pages 146-150.", |
|
"links": null |
|
}, |
|
"BIBREF18": { |
|
"ref_id": "b18", |
|
"title": "Oh oh oh whoah! towards automatic topic detection in song lyrics", |
|
"authors": [ |
|
{ |
|
"first": "Florian", |
|
"middle": [], |
|
"last": "Kleedorfer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Peter", |
|
"middle": [], |
|
"last": "Knees", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tim", |
|
"middle": [], |
|
"last": "Pohle", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "Ismir", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "287--292", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Florian Kleedorfer, Peter Knees, and Tim Pohle. 2008. Oh oh oh whoah! towards automatic topic detection in song lyrics. In Ismir, pages 287-292.", |
|
"links": null |
|
}, |
|
"BIBREF19": { |
|
"ref_id": "b19", |
|
"title": "Statisticsbased summarization -step one: Sentence compression", |
|
"authors": [ |
|
{ |
|
"first": "Kevin", |
|
"middle": [], |
|
"last": "Knight", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Marcu", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Proceedings of the Seventeenth National Conference on Artificial Intelligence and Twelfth Conference on Innovative Applications of Artificial Intelligence", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "703--710", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kevin Knight and Daniel Marcu. 2000. Statistics- based summarization -step one: Sentence compres- sion. In Proceedings of the Seventeenth National Conference on Artificial Intelligence and Twelfth Conference on Innovative Applications of Artificial Intelligence, pages 703-710. AAAI Press.", |
|
"links": null |
|
}, |
|
"BIBREF20": { |
|
"ref_id": "b20", |
|
"title": "Extraction of high-level musical structure from audio data and its application to thumbnail generation", |
|
"authors": [ |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Levy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mark", |
|
"middle": [], |
|
"last": "Sandler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Casey", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings", |
|
"volume": "5", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mark Levy, Mark Sandler, and Michael Casey. 2006. Extraction of high-level musical structure from au- dio data and its application to thumbnail generation. In 2006 IEEE International Conference on Acous- tics Speech and Signal Processing Proceedings, vol- ume 5, pages V-V. IEEE.", |
|
"links": null |
|
}, |
|
"BIBREF21": { |
|
"ref_id": "b21", |
|
"title": "Rouge: A package for automatic evaluation of summaries", |
|
"authors": [ |
|
{ |
|
"first": "Chin-Yew", |
|
"middle": [], |
|
"last": "Lin", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Text Summarization Branches Out", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text Summarization Branches Out.", |
|
"links": null |
|
}, |
|
"BIBREF22": { |
|
"ref_id": "b22", |
|
"title": "Automatically assessing machine summary content without a gold standard", |
|
"authors": [ |
|
{ |
|
"first": "Annie", |
|
"middle": [], |
|
"last": "Louis", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ani", |
|
"middle": [], |
|
"last": "Nenkova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "Computational Linguistics", |
|
"volume": "", |
|
"issue": "2", |
|
"pages": "", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1162/COLI_a_00123" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Annie Louis and Ani Nenkova. 2013. Automatically assessing machine summary content without a gold standard. Computational Linguistics, 39(2).", |
|
"links": null |
|
}, |
|
"BIBREF23": { |
|
"ref_id": "b23", |
|
"title": "On choosing an effective automatic evaluation metric for microblog summarisation", |
|
"authors": [ |
|
{ |
|
"first": "Stuart", |
|
"middle": [], |
|
"last": "Mackie", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Richard", |
|
"middle": [], |
|
"last": "Mccreadie", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Craig", |
|
"middle": [], |
|
"last": "Macdonald", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Iadh", |
|
"middle": [], |
|
"last": "Ounis", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the 5th Information Interaction in Context Symposium", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "115--124", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Stuart Mackie, Richard McCreadie, Craig Macdonald, and Iadh Ounis. 2014. On choosing an effective au- tomatic evaluation metric for microblog summarisa- tion. In Proceedings of the 5th Information Interac- tion in Context Symposium, pages 115-124. ACM.", |
|
"links": null |
|
}, |
|
"BIBREF24": { |
|
"ref_id": "b24", |
|
"title": "Generating impact-based summaries for scientific literature", |
|
"authors": [ |
|
{ |
|
"first": "Qiaozhu", |
|
"middle": [], |
|
"last": "Mei", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chengxiang", |
|
"middle": [], |
|
"last": "Zhai", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "ACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Qiaozhu Mei and ChengXiang Zhai. 2008. Generating impact-based summaries for scientific literature. In ACL.", |
|
"links": null |
|
}, |
|
"BIBREF25": { |
|
"ref_id": "b25", |
|
"title": "WASABI: a Two Million Song Database Project with Audio and Cultural Metadata plus WebAudio enhanced Client Applications", |
|
"authors": [ |
|
{ |
|
"first": "Gabriel", |
|
"middle": [], |
|
"last": "Meseguer-Brocal", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Geoffroy", |
|
"middle": [], |
|
"last": "Peeters", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Guillaume", |
|
"middle": [], |
|
"last": "Pellerin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michel", |
|
"middle": [], |
|
"last": "Buffa", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Elena", |
|
"middle": [], |
|
"last": "Cabrio", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Catherine", |
|
"middle": [ |
|
"Faron" |
|
], |
|
"last": "Zucker", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Alain", |
|
"middle": [], |
|
"last": "Giboin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Isabelle", |
|
"middle": [], |
|
"last": "Mirbel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Romain", |
|
"middle": [], |
|
"last": "Hennequin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Manuel", |
|
"middle": [], |
|
"last": "Moussallam", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Francesco", |
|
"middle": [], |
|
"last": "Piccoli", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Thomas", |
|
"middle": [], |
|
"last": "Fillon", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Web Audio Conference 2017 -Collaborative Audio #WAC2017", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Gabriel Meseguer-Brocal, Geoffroy Peeters, Guil- laume Pellerin, Michel Buffa, Elena Cabrio, Cather- ine Faron Zucker, Alain Giboin, Isabelle Mirbel, Ro- main Hennequin, Manuel Moussallam, Francesco Piccoli, and Thomas Fillon. 2017. WASABI: a Two Million Song Database Project with Audio and Cul- tural Metadata plus WebAudio enhanced Client Ap- plications. In Web Audio Conference 2017 -Col- laborative Audio #WAC2017, London, United King- dom. Queen Mary University of London.", |
|
"links": null |
|
}, |
|
"BIBREF26": { |
|
"ref_id": "b26", |
|
"title": "Textrank: Bringing order into text", |
|
"authors": [ |
|
{ |
|
"first": "Rada", |
|
"middle": [], |
|
"last": "Mihalcea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Tarau", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Rada Mihalcea and Paul Tarau. 2004. Textrank: Bring- ing order into text. In Proceedings of the 2004 Con- ference on Empirical Methods in Natural Language Processing.", |
|
"links": null |
|
}, |
|
"BIBREF27": { |
|
"ref_id": "b27", |
|
"title": "Summarunner: A recurrent neural network based sequence model for extractive summarization of documents", |
|
"authors": [ |
|
{ |
|
"first": "Ramesh", |
|
"middle": [], |
|
"last": "Nallapati", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Feifei", |
|
"middle": [], |
|
"last": "Zhai", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bowen", |
|
"middle": [], |
|
"last": "Zhou", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "AAAI", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ramesh Nallapati, Feifei Zhai, and Bowen Zhou. 2016. Summarunner: A recurrent neural network based se- quence model for extractive summarization of docu- ments. In AAAI.", |
|
"links": null |
|
}, |
|
"BIBREF28": { |
|
"ref_id": "b28", |
|
"title": "Automatic summarization", |
|
"authors": [ |
|
{ |
|
"first": "Ani", |
|
"middle": [], |
|
"last": "Nenkova", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kathleen", |
|
"middle": [], |
|
"last": "Mckeown", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "Foundations and Trends R in Information Retrieval", |
|
"volume": "5", |
|
"issue": "2-3", |
|
"pages": "103--233", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ani Nenkova, Kathleen McKeown, et al. 2011. Auto- matic summarization. Foundations and Trends R in Information Retrieval, 5(2-3):103-233.", |
|
"links": null |
|
}, |
|
"BIBREF29": { |
|
"ref_id": "b29", |
|
"title": "Using random walks for questionfocused sentence retrieval", |
|
"authors": [ |
|
{ |
|
"first": "Jahna", |
|
"middle": [], |
|
"last": "Otterbacher", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G\u00fcne\u015f", |
|
"middle": [], |
|
"last": "Erkan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Dragomir R Radev", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "915--922", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jahna Otterbacher, G\u00fcne\u015f Erkan, and Dragomir R Radev. 2005. Using random walks for question- focused sentence retrieval. In Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Process- ing, pages 915-922. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF30": { |
|
"ref_id": "b30", |
|
"title": "The pagerank citation ranking: Bringing order to the web", |
|
"authors": [ |
|
{ |
|
"first": "Lawrence", |
|
"middle": [], |
|
"last": "Page", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sergey", |
|
"middle": [], |
|
"last": "Brin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rajeev", |
|
"middle": [], |
|
"last": "Motwani", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Terry", |
|
"middle": [], |
|
"last": "Winograd", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1999, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lawrence Page, Sergey Brin, Rajeev Motwani, and Terry Winograd. 1999. The pagerank citation rank- ing: Bringing order to the web. Technical report, Stanford InfoLab.", |
|
"links": null |
|
}, |
|
"BIBREF31": { |
|
"ref_id": "b31", |
|
"title": "Topical coherence for graph-based extractive summarization", |
|
"authors": [ |
|
{ |
|
"first": "Daraksha", |
|
"middle": [], |
|
"last": "Parveen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Hans-Martin Ramsl", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Strube", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1949--1954", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Daraksha Parveen, Hans-Martin Ramsl, and Michael Strube. 2015. Topical coherence for graph-based ex- tractive summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Lan- guage Processing, pages 1949-1954.", |
|
"links": null |
|
}, |
|
"BIBREF32": { |
|
"ref_id": "b32", |
|
"title": "Integrating importance, non-redundancy and coherence in graph-based extractive summarization", |
|
"authors": [ |
|
{ |
|
"first": "Daraksha", |
|
"middle": [], |
|
"last": "Parveen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Strube", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the 24th International Conference on Artificial Intelligence, IJCAI'15", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1298--1304", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Daraksha Parveen and Michael Strube. 2015. Inte- grating importance, non-redundancy and coherence in graph-based extractive summarization. In Pro- ceedings of the 24th International Conference on Artificial Intelligence, IJCAI'15, pages 1298-1304. AAAI Press.", |
|
"links": null |
|
}, |
|
"BIBREF33": { |
|
"ref_id": "b33", |
|
"title": "Towards opinion summarization of customer reviews", |
|
"authors": [ |
|
{ |
|
"first": "Samuel", |
|
"middle": [], |
|
"last": "Pecar", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of ACL 2018, Student Research Workshop", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1--8", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Samuel Pecar. 2018. Towards opinion summarization of customer reviews. In Proceedings of ACL 2018, Student Research Workshop, pages 1-8. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF35": { |
|
"ref_id": "b35", |
|
"title": "Scikit-learn: Machine learning in Python", |
|
"authors": [], |
|
"year": null, |
|
"venue": "Journal of Machine Learning Research", |
|
"volume": "12", |
|
"issue": "", |
|
"pages": "2825--2830", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825-2830.", |
|
"links": null |
|
}, |
|
"BIBREF36": { |
|
"ref_id": "b36", |
|
"title": "Automatic Text Summarization: Past, Present and Future", |
|
"authors": [ |
|
{ |
|
"first": "Horacio", |
|
"middle": [], |
|
"last": "Saggion", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Thierry", |
|
"middle": [], |
|
"last": "Poibeau", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2013, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "3--21", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1007/978-3-642-28569-1_1" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Horacio Saggion and Thierry Poibeau. 2013. Auto- matic Text Summarization: Past, Present and Fu- ture, pages 3-21. Springer, Berlin, Heidelberg.", |
|
"links": null |
|
}, |
|
"BIBREF37": { |
|
"ref_id": "b37", |
|
"title": "Summarization evaluation in the absence of human model summaries using the compositionality of word embeddings", |
|
"authors": [ |
|
{ |
|
"first": "Elaheh", |
|
"middle": [], |
|
"last": "Shafieibavani", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mohammad", |
|
"middle": [], |
|
"last": "Ebrahimi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Raymond", |
|
"middle": [], |
|
"last": "Wong", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Fang", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 27th International Conference on Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "905--914", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Elaheh ShafieiBavani, Mohammad Ebrahimi, Ray- mond Wong, and Fang Chen. 2018. Summariza- tion evaluation in the absence of human model sum- maries using the compositionality of word embed- dings. In Proceedings of the 27th International Con- ference on Computational Linguistics, pages 905- 914. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF38": { |
|
"ref_id": "b38", |
|
"title": "Optimal transport: old and new", |
|
"authors": [ |
|
{ |
|
"first": "C\u00e9dric", |
|
"middle": [], |
|
"last": "Villani", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2008, |
|
"venue": "", |
|
"volume": "338", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "C\u00e9dric Villani. 2008. Optimal transport: old and new, volume 338. Springer Science & Business Media.", |
|
"links": null |
|
}, |
|
"BIBREF39": { |
|
"ref_id": "b39", |
|
"title": "A sentence compression based framework to query-focused multi-document summarization", |
|
"authors": [ |
|
{ |
|
"first": "Lu", |
|
"middle": [], |
|
"last": "Wang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hema", |
|
"middle": [], |
|
"last": "Raghavan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Vittorio", |
|
"middle": [], |
|
"last": "Castelli", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Radu", |
|
"middle": [], |
|
"last": "Florian", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Claire", |
|
"middle": [], |
|
"last": "Cardie", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"arXiv": [ |
|
"arXiv:1606.07548" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Lu Wang, Hema Raghavan, Vittorio Castelli, Radu Florian, and Claire Cardie. 2016. A sentence compression based framework to query-focused multi-document summarization. arXiv preprint arXiv:1606.07548.", |
|
"links": null |
|
}, |
|
"BIBREF40": { |
|
"ref_id": "b40", |
|
"title": "Modeling discourse segments in lyrics using repeated patterns", |
|
"authors": [ |
|
{ |
|
"first": "Kento", |
|
"middle": [], |
|
"last": "Watanabe", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yuichiroh", |
|
"middle": [], |
|
"last": "Matsubayashi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Naho", |
|
"middle": [], |
|
"last": "Orita", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Naoaki", |
|
"middle": [], |
|
"last": "Okazaki", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kentaro", |
|
"middle": [], |
|
"last": "Inui", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Satoru", |
|
"middle": [], |
|
"last": "Fukayama", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tomoyasu", |
|
"middle": [], |
|
"last": "Nakano", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jordan", |
|
"middle": [], |
|
"last": "Smith", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Masataka", |
|
"middle": [], |
|
"last": "Goto", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "1959--1969", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kento Watanabe, Yuichiroh Matsubayashi, Naho Orita, Naoaki Okazaki, Kentaro Inui, Satoru Fukayama, Tomoyasu Nakano, Jordan Smith, and Masataka Goto. 2016. Modeling discourse segments in lyrics using repeated patterns. In Proceedings of COLING 2016, the 26th International Conference on Compu- tational Linguistics: Technical Papers, pages 1959- 1969.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Song text of \"Let's start a band\" by Amy MacDonald along with two example summaries.", |
|
"uris": null |
|
}, |
|
"FIGREF1": { |
|
"type_str": "figure", |
|
"num": null, |
|
"text": "Human ratings per summarization model in terms of average and standard deviation.Non-redundancy. The criteria Informativeness and Meaning are highly correlated (Pearson correlation coefficient 0.84). Correlations between other criteria range between 0.29 and 0.51.", |
|
"uris": null |
|
}, |
|
"TABREF1": { |
|
"text": "Automatic evaluation results for the 5 summarization models and 2 genre clusters. Distributional Semantics and Topical are relative to the best model (=100%), Coherence and Fitness to the original text (=100%).", |
|
"html": null, |
|
"num": null, |
|
"type_str": "table", |
|
"content": "<table/>" |
|
} |
|
} |
|
} |
|
} |