File size: 6,719 Bytes
2ea1065 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 |
# TweetSumm
A dataset focused on summarization of dialogs, which represents the rich domain of Twitter customer care conversations
Tweetsumm comprises of 1,100 dialogs reconstructed from Tweets that appear in the [Kaggle Customer Support On Twitter](http://www.kaggle.com/thoughtvector/customer-support-on-twitter) dataset each accompanied by 3 extractive and 3 abstractive summaries generated by human annotators.
If you use this dataset in your work, please cite our paper
```
@inproceedings{feigenblat-etal-2021-tweetsumm-dialog,
title = "{TWEETSUMM} - A Dialog Summarization Dataset for Customer Service",
author = "Feigenblat, Guy and
Gunasekara, Chulaka and
Sznajder, Benjamin and
Joshi, Sachindra and
Konopnicki, David and
Aharonov, Ranit",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
month = nov,
year = "2021",
address = "Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.findings-emnlp.24",
pages = "245--260",
}
```
## TweetSumm format
TweetSumm is released with a partition in Train/Test/Valid files.
Each file is in [JSON Lines format](https://jsonlines.org/).
A TweetSum entry (line) has the following format:
- **conversation_id** : a unique identifier of the dialog
- **tweet_ids_sentence_offset** : a list of
- **tweet_id** : corresponding to Twitter Id in [Kaggle Customer Support On Twitter](http://www.kaggle.com/thoughtvector/customer-support-on-twitter) dataset
- **sentence offsets** : the offsets of the sentences splitting we used.
- **annotations** : a list of summaries generated by the human annotators - each entry contains :
- **extractive** : a list of sentences selected from the initial dialog. The sentences are in the format **tweet_id**, **sentence offsets**
- **abstractive** : a list of one or two sentences.
## TweetSumm processing
The script `tweet_summ_processor` allows to transform the TweetSumm entries to readable strings for further processing.
The `TweetSumProcessor` class requires one parameter : the path to the `twcs file` of [Kaggle Customer Support On Twitter](http://www.kaggle.com/thoughtvector/customer-support-on-twitter). This file can be download [here](https://www.kaggle.com/thoughtvector/customer-support-on-twitter/download).
The `TweetSumProcessor` has one method : `get_dialog_with_summaries` which gets a list of TweetSumm entries (lines in the TweetSumm files) and returns a list of corresponding `DialogWithSummaries` objects. These objects allow to access to the readable conversation associated with their readable human generated summaries.
**Code sample:**
```
processor = TweetSumProcessor(TWCS_FILE_PATH)
with open(TWEET_SUMM_FILE_PATH) as f:
dialog_with_summaries = processor.get_dialog_with_summaries(f.readlines())
for dialog_with_summary in dialog_with_summaries:
json_format = dialog_with_summary.get_json()
string_format = str(dialog_with_summary)
...
```
## TweetSumm entry sample
```
{
"conversation_id": "bbde6d8ec7c39c4551da1ff6024f997b",
"tweet_ids_sentence_offset": [
{
"tweet_id": 2263653,
"sentence_offsets": [
"[0, 80]",
"[82, 95]"
]
},
{
"tweet_id": 2263651,
"sentence_offsets": [
"[0, 43]",
"[44, 68]",
"[69, 134]"
]
},
{
"tweet_id": 2263652,
"sentence_offsets": [
"[0, 57]"
]
},
{
"tweet_id": 2263654,
"sentence_offsets": [
"[0, 14]",
"[15, 114]"
]
},
{
"tweet_id": 2263655,
"sentence_offsets": [
"[0, 24]",
"[25, 76]"
]
},
{
"tweet_id": 2263656,
"sentence_offsets": [
"[0, 42]",
"[43, 132]"
]
},
{
"tweet_id": 2263657,
"sentence_offsets": [
"[0, 67]",
"[68, 118]",
"[119, 163]",
"[164, 177]"
]
},
{
"tweet_id": 2263658,
"sentence_offsets": [
"[0, 16]",
"[17, 45]",
"[46, 108]",
"[109, 110]"
]
}
],
"annotations": [
{
"extractive": [
{
"tweet_id": 2263653,
"sentence_offset": "[0, 80]"
},
{
"tweet_id": 2263654,
"sentence_offset": "[15, 114]"
},
{
"tweet_id": 2263656,
"sentence_offset": "[43, 132]"
},
{
"tweet_id": 2263658,
"sentence_offset": "[46, 108]"
}
],
"abstractive": [
"Customer is complaining that the watchlist is not updated with new episodes from past two days.",
"Agent informed that the team is working hard to investigate to show new episodes on page."
]
},
{
"extractive": [
{
"tweet_id": 2263653,
"sentence_offset": "[0, 80]"
},
{
"tweet_id": 2263651,
"sentence_offset": "[69, 134]"
},
{
"tweet_id": 2263654,
"sentence_offset": "[15, 114]"
},
{
"tweet_id": 2263655,
"sentence_offset": "[25, 76]"
},
{
"tweet_id": 2263656,
"sentence_offset": "[43, 132]"
}
],
"abstractive": [
"Customer is complaining that my watch list is not updating with the new episodes.",
"Agent updated to recommend checking the show page for these shows as the new eps will be there."
]
},
{
"extractive": [
{
"tweet_id": 2263653,
"sentence_offset": "[0, 80]"
},
{
"tweet_id": 2263651,
"sentence_offset": "[69, 134]"
},
{
"tweet_id": 2263654,
"sentence_offset": "[15, 114]"
},
{
"tweet_id": 2263656,
"sentence_offset": "[43, 132]"
},
{
"tweet_id": 2263657,
"sentence_offset": "[0, 67]"
}
],
"abstractive": [
"Customer is complaining that he is not getting the updated watch list with new episodes for some shows.",
"Agent assures that their team is working on it to resolve the issue and requests to navigate manually in the meantime."
]
}
]
}
```
## License
Dataset released under the CDLA-Sharing license https://cdla.io/sharing-1-0/
## Disclaimer
IBM is not responsible for the content of the data, nor for any claim related to the data (including claims related to alleged intellectual property or privacy breach).
|