ACL-OCL / Base_JSON /prefixC /json /cl /2020.cl-2.8.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "2020",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T13:18:08.326710Z"
},
"title": "The Limitations of Stylometry for Detecting Machine-Generated Fake News",
"authors": [
{
"first": "Tal",
"middle": [],
"last": "Schuster",
"suffix": "",
"affiliation": {
"laboratory": "Artificial Intelligence Laboratory",
"institution": "Massachusetts Institute of Technology",
"location": {}
},
"email": "[email protected]"
},
{
"first": "Roei",
"middle": [],
"last": "Schuster",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Darsh",
"middle": [
"J"
],
"last": "Shah",
"suffix": "",
"affiliation": {
"laboratory": "Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology",
"institution": "",
"location": {}
},
"email": ""
},
{
"first": "Regina",
"middle": [],
"last": "Barzilay",
"suffix": "",
"affiliation": {
"laboratory": "Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology",
"institution": "",
"location": {}
},
"email": "[email protected]"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Recent developments in neural language models (LMs) have raised concerns about their potential misuse for automatically spreading misinformation. In light of these concerns, several studies have proposed to detect machine-generated fake news by capturing their stylistic differences from human-written text. These approaches, broadly termed stylometry, have found success in source attribution and misinformation detection in human-written texts. However, in this work, we show that stylometry is limited against machine-generated misinformation. Whereas humans speak differently when trying to deceive, LMs generate stylistically consistent text, regardless of underlying motive. Thus, though stylometry can successfully prevent impersonation by identifying text provenance, it fails to distinguish legitimate LM applications from those that introduce false information. We create two benchmarks demonstrating the stylistic similarity between malicious and legitimate uses of LMs, utilized in auto-completion and editing-assistance settings. 1 Our findings highlight the need for non-stylometry approaches in detecting machinegenerated misinformation, and open up the discussion on the desired evaluation benchmarks.",
"pdf_parse": {
"paper_id": "2020",
"_pdf_hash": "",
"abstract": [
{
"text": "Recent developments in neural language models (LMs) have raised concerns about their potential misuse for automatically spreading misinformation. In light of these concerns, several studies have proposed to detect machine-generated fake news by capturing their stylistic differences from human-written text. These approaches, broadly termed stylometry, have found success in source attribution and misinformation detection in human-written texts. However, in this work, we show that stylometry is limited against machine-generated misinformation. Whereas humans speak differently when trying to deceive, LMs generate stylistically consistent text, regardless of underlying motive. Thus, though stylometry can successfully prevent impersonation by identifying text provenance, it fails to distinguish legitimate LM applications from those that introduce false information. We create two benchmarks demonstrating the stylistic similarity between malicious and legitimate uses of LMs, utilized in auto-completion and editing-assistance settings. 1 Our findings highlight the need for non-stylometry approaches in detecting machinegenerated misinformation, and open up the discussion on the desired evaluation benchmarks.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Many previous studies on stylometry-the extraction of stylistic features from written text-showed promising results on text classification. Two of stylometry's common applications are: (1) Detecting the provenance of text (i.e., identifying the author) in order to prevent impersonations (Tweedie, Singh, and Holmes 1996; Brennan, Afroz, and Greenstadt 2012; Afroz et al. 2014; Caliskan-Islam et al. 2015; Neal et al. 2017; Sari, Vlachos, and Stevenson 2017) ; and (2) Detecting misinformation in text due to deception (Enos et al. 2007; Mihalcea and Strapparava 2009; Ott et al. 2011; Feng, Banerjee, and Choi 2012; Afroz, Brennan, and Greenstadt 2012) , fake news (Rashkin et al. 2017; P\u00e9rez-Rosas et al. 2018) , or other false or illegal content (Choshen et al. 2019) . In the former, the classifier identifies language features that correlate with a specific person or group. The latter, misinformation detection, relies on idiosyncrasies of lies, namely, style and language characteristics that are unique to text that is false or misleading.",
"cite_spans": [
{
"start": 288,
"end": 321,
"text": "(Tweedie, Singh, and Holmes 1996;",
"ref_id": "BIBREF42"
},
{
"start": 322,
"end": 358,
"text": "Brennan, Afroz, and Greenstadt 2012;",
"ref_id": "BIBREF7"
},
{
"start": 359,
"end": 377,
"text": "Afroz et al. 2014;",
"ref_id": "BIBREF2"
},
{
"start": 378,
"end": 405,
"text": "Caliskan-Islam et al. 2015;",
"ref_id": "BIBREF8"
},
{
"start": 406,
"end": 423,
"text": "Neal et al. 2017;",
"ref_id": "BIBREF27"
},
{
"start": 424,
"end": 458,
"text": "Sari, Vlachos, and Stevenson 2017)",
"ref_id": "BIBREF33"
},
{
"start": 519,
"end": 537,
"text": "(Enos et al. 2007;",
"ref_id": "BIBREF13"
},
{
"start": 538,
"end": 568,
"text": "Mihalcea and Strapparava 2009;",
"ref_id": "BIBREF24"
},
{
"start": 569,
"end": 585,
"text": "Ott et al. 2011;",
"ref_id": "BIBREF28"
},
{
"start": 586,
"end": 616,
"text": "Feng, Banerjee, and Choi 2012;",
"ref_id": "BIBREF14"
},
{
"start": 617,
"end": 653,
"text": "Afroz, Brennan, and Greenstadt 2012)",
"ref_id": "BIBREF1"
},
{
"start": 666,
"end": 687,
"text": "(Rashkin et al. 2017;",
"ref_id": "BIBREF32"
},
{
"start": 688,
"end": 712,
"text": "P\u00e9rez-Rosas et al. 2018)",
"ref_id": "BIBREF28"
},
{
"start": 749,
"end": 770,
"text": "(Choshen et al. 2019)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "Stylometry has recently gained attention as a potential answer to concerns that language models (LMs) could be used to mass-produce malicious text (Vosoughi, Roy, and Aral 2018; Radford et al. 2019) , that (1) impersonates a human author's text and/or (2) is fallacious and misleading. Indeed, stylometry-based approaches have shown promising results for defending against human-impersonating language models (LMs) (Bakhtin et al. 2019; Zellers et al. 2019) . However, as applications of text generation such as text auto-completion (Wolf et al. 2019; House 2019 ) and automatic question answering 2 become widely used, labeling text as generated by an LM might not indicate anything at all about its trustworthiness. This motivates our core inquiry subject: Can stylometry be used to distinguish malicious uses of language models (LMs) from legitimate ones?",
"cite_spans": [
{
"start": 147,
"end": 177,
"text": "(Vosoughi, Roy, and Aral 2018;",
"ref_id": "BIBREF46"
},
{
"start": 178,
"end": 198,
"text": "Radford et al. 2019)",
"ref_id": "BIBREF31"
},
{
"start": 415,
"end": 436,
"text": "(Bakhtin et al. 2019;",
"ref_id": "BIBREF4"
},
{
"start": 437,
"end": 457,
"text": "Zellers et al. 2019)",
"ref_id": "BIBREF50"
},
{
"start": 533,
"end": 551,
"text": "(Wolf et al. 2019;",
"ref_id": "BIBREF49"
},
{
"start": 552,
"end": 562,
"text": "House 2019",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "We build the first benchmark for detection of LM-produced fake news that labels text as \"real\" or \"fake\" according to its veracity. Inspired by studies on deceitful behaviors in humans, showing that people try to diverge as little as possible from the truth when lying (Mazar, Amir, and Ariely 2008) , we focus on automatic false modifications or additions to otherwise truthful news stories.",
"cite_spans": [
{
"start": 269,
"end": 299,
"text": "(Mazar, Amir, and Ariely 2008)",
"ref_id": "BIBREF23"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "Our data sets contain articles produced by both malicious and responsible uses of language models, and the detector's task is to identify the malicious ones. In one data set, we produce text by prompting an LM to extend news articles with relevant claims. We simulate malicious user, who only accepts the LM's suggestion if the claim is factually false, and a responsible user, who only accepts correct claims. The produced sentences are short and concise statements, similarly to fake news and false claims as represented in human-generated data sets Augenstein et al. 2019) . In another data set, we modify existing news articles to include false information by inverting article statements. In this case, the LM is used to automatically identify the most plausible edit locations. This is similar to (mis-)using an autocorrect tool that suggests local modifications.",
"cite_spans": [
{
"start": 552,
"end": 575,
"text": "Augenstein et al. 2019)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "We find that with the state-of-the-art stylometry-based classifier, even a single autogenerated sentence within a wall of human-written text is detectable with high accuracy, yet the truthfulness of a single sentence remains largely undecidable. Moreover, even a relatively weak LM can be used to produce statement inversions that the state-of-the-art stylometry-based model cannot detect. Thus, stylometry fails to distinguish malicious from responsible behaviors. This indicates that, unlike humans who expose stylistic cues when writing false content (Ott et al. 2011; Frank, Menasco, and O'Sullivan 2008; Matsumoto et al. 2011) , LMs maintain consistency for both true and false content. Worse yet, while a provenance classifier can effectively detect a potentially malicious author or publication venue (given enough examples), it might not distinguish malicious from legitimate authors if they are both using the same LM to generate their text. In this regard, malicious text generated by an LM might actually be harder to detect than hand-crafted malicious text.",
"cite_spans": [
{
"start": 554,
"end": 571,
"text": "(Ott et al. 2011;",
"ref_id": "BIBREF28"
},
{
"start": 572,
"end": 608,
"text": "Frank, Menasco, and O'Sullivan 2008;",
"ref_id": "BIBREF15"
},
{
"start": 609,
"end": 631,
"text": "Matsumoto et al. 2011)",
"ref_id": "BIBREF22"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "Our human evaluation tests show that humans are also fooled by machinegenerated misinformation, but that access to external information sources can help. Therefore, we recommend future research on machine-generated misinformation to focus on non-stylometry strategies. Finally, we discuss what benchmarks are required for evaluating the performance of such detectors.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "Stylometry for Human-Written Text. The use of statistical methods for analyzing humanwritten documents has been studied extensively since the early days of the field. One common application is provenance detection. For example, Mosteller and Wallace (1963) used word counts to predict the authorship of historical documents. Tweedie, Singh, and Holmes (1996) extracted other stylistic features and applied a neural network to the same task. Although these classifiers could be fooled by an aware writer that intentionally imitates other's style (Brennan, Afroz, and Greenstadt 2012) , this approach was found useful for de-anonymizing cybercriminals in forums (Afroz et al. 2014) , identifying programmers (Caliskan-Islam et al. 2015) , and more (Neal et al. 2017) . In a related line of study, stylometry was applied to, rather than detecting the specific person, identifying characteristics of the author, such as age and gender (Goswami, Sarkar, and Rustagi 2009) , political views (Potthast et al. 2018), or IQ (Abramov and Yampolskiy 2019). Recently, as detailed later in this article, stylometry was used to distinguish machine-from human-writers.",
"cite_spans": [
{
"start": 228,
"end": 256,
"text": "Mosteller and Wallace (1963)",
"ref_id": "BIBREF25"
},
{
"start": 325,
"end": 358,
"text": "Tweedie, Singh, and Holmes (1996)",
"ref_id": "BIBREF42"
},
{
"start": 545,
"end": 582,
"text": "(Brennan, Afroz, and Greenstadt 2012)",
"ref_id": "BIBREF7"
},
{
"start": 660,
"end": 679,
"text": "(Afroz et al. 2014)",
"ref_id": "BIBREF2"
},
{
"start": 706,
"end": 734,
"text": "(Caliskan-Islam et al. 2015)",
"ref_id": "BIBREF8"
},
{
"start": 746,
"end": 764,
"text": "(Neal et al. 2017)",
"ref_id": "BIBREF27"
},
{
"start": 931,
"end": 966,
"text": "(Goswami, Sarkar, and Rustagi 2009)",
"ref_id": "BIBREF16"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background and Related Work",
"sec_num": "2."
},
{
"text": "Another common application of stylometry is detecting human-written misinformation. Mihalcea and Strapparava (2009) found specific words that are highly correlated with true and false statements. Ott et al. (2011) and Feng, Banerjee, and Choi (2012) used a richer set of features such as POS tag frequencies and constituency structure to identify deceptive writing. Following these observations and the increasing interest in fake news, recent studies applied stylometry on entire news articles (Pisarevskaya 2017), short news reports (P\u00e9rez-Rosas et al. 2018), fact and political statements (Nakashole and Mitchell 2014; Rashkin et al. 2017) , and posts in social media (Volkova et al. 2017) . The success of these studies is mostly attributed to stylistic changes in human language when lying or deceiving (Bond and Lee 2005; Frank, Menasco, and O'Sullivan 2008) . In this work, we evaluate the viability of this approach on machine-generated text, where stylistic differences between truth and lie might be more subtle.",
"cite_spans": [
{
"start": 84,
"end": 115,
"text": "Mihalcea and Strapparava (2009)",
"ref_id": "BIBREF24"
},
{
"start": 196,
"end": 213,
"text": "Ott et al. (2011)",
"ref_id": "BIBREF28"
},
{
"start": 218,
"end": 249,
"text": "Feng, Banerjee, and Choi (2012)",
"ref_id": "BIBREF14"
},
{
"start": 592,
"end": 621,
"text": "(Nakashole and Mitchell 2014;",
"ref_id": "BIBREF26"
},
{
"start": 622,
"end": 642,
"text": "Rashkin et al. 2017)",
"ref_id": "BIBREF32"
},
{
"start": 671,
"end": 692,
"text": "(Volkova et al. 2017)",
"ref_id": "BIBREF44"
},
{
"start": 808,
"end": 827,
"text": "(Bond and Lee 2005;",
"ref_id": "BIBREF6"
},
{
"start": 828,
"end": 864,
"text": "Frank, Menasco, and O'Sullivan 2008)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background and Related Work",
"sec_num": "2."
},
{
"text": "Machine-Generated Text Detection. Detecting text's provenance is similar to authorship attribution and, therefore, stylometry can be effective. Indeed, Gehrmann, Strobelt, and Rush (2019) show the existence of distributional differences between human-written texts and machine-generated ones by visualizing the per-token probability according to an LM. Bakhtin et al. (2019) learn a dedicated provenance neural classifier. Although their classifier achieves high in-domain accuracy, they find that it overfits the generated text distribution rather than detecting outliers from human texts, resulting in increased \"human-ness\" scores for random perturbations. Nevertheless, an advantage of such neural approaches over more traditional stylometry methods is that, given enough data, the model learns hidden stylistic representations without the need to manually define any features.",
"cite_spans": [
{
"start": 152,
"end": 187,
"text": "Gehrmann, Strobelt, and Rush (2019)",
"ref_id": "BIBREF16"
},
{
"start": 353,
"end": 374,
"text": "Bakhtin et al. (2019)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background and Related Work",
"sec_num": "2."
},
{
"text": "Building on this observation, Zellers et al. (2019) focus on fake news and create a Transformer-based LM (Vaswani et al. 2017) dubbed Grover and train it on a large news corpus. Grover also includes a \"neural fake news detector,\" a linear classifier on top of the hidden state of the last token of the examined article, fine-tuned to classify whether the news text was machine-generated or not. The experiments in this article are based on the Grover-Mega classifier, fine-tuned for the target task (see Section 3).",
"cite_spans": [
{
"start": 30,
"end": 51,
"text": "Zellers et al. (2019)",
"ref_id": "BIBREF50"
},
{
"start": 105,
"end": 126,
"text": "(Vaswani et al. 2017)",
"ref_id": "BIBREF43"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background and Related Work",
"sec_num": "2."
},
{
"text": "Fake News Detection Approaches Beyond Stylometry. The other most extensively studied NLP-based approach for fake news detection is based on fact-checking. This approach has recently gained increasing attention thanks to several synthetic (Thorne et al. 2018) and real-world data sets (Hanselowski et al. 2018; Popat et al. 2018; Augenstein et al. 2019) . The performance of current models is still far from that of humans (Thorne et al. 2019; Schuster et al. 2019 ), but with their advancements they can still play a positive role in detection.",
"cite_spans": [
{
"start": 238,
"end": 258,
"text": "(Thorne et al. 2018)",
"ref_id": "BIBREF39"
},
{
"start": 284,
"end": 309,
"text": "(Hanselowski et al. 2018;",
"ref_id": "BIBREF18"
},
{
"start": 310,
"end": 328,
"text": "Popat et al. 2018;",
"ref_id": "BIBREF29"
},
{
"start": 329,
"end": 352,
"text": "Augenstein et al. 2019)",
"ref_id": "BIBREF3"
},
{
"start": 422,
"end": 442,
"text": "(Thorne et al. 2019;",
"ref_id": "BIBREF38"
},
{
"start": 443,
"end": 463,
"text": "Schuster et al. 2019",
"ref_id": "BIBREF34"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background and Related Work",
"sec_num": "2."
},
{
"text": "Another line of work for fake news detection utilizes non-textual information such as how content is propagated, by which users, its originating URL, and other metadata (Castillo, Mendoza, and Poblete 2011; Gupta et al. 2014; Zhao, Resnick, and Mei 2015; Kochkina, Liakata, and Zubiaga 2018; Liu and Wu 2018) , as well as incorporation of users' explicit feedback, such as abuse reporting (Tschiatschek et al. 2018) . Social network platforms, ISPs, and even individual users can use such methods to moderate content exposure. These approaches are beset by the challenges (Shu et al. 2017) of noisy and incomplete data, especially given the need to detect fake news early (Liu and Wu 2018) (before propagation and user engagement patterns are fully formed).",
"cite_spans": [
{
"start": 169,
"end": 206,
"text": "(Castillo, Mendoza, and Poblete 2011;",
"ref_id": "BIBREF9"
},
{
"start": 207,
"end": 225,
"text": "Gupta et al. 2014;",
"ref_id": "BIBREF17"
},
{
"start": 226,
"end": 254,
"text": "Zhao, Resnick, and Mei 2015;",
"ref_id": "BIBREF51"
},
{
"start": 255,
"end": 291,
"text": "Kochkina, Liakata, and Zubiaga 2018;",
"ref_id": "BIBREF20"
},
{
"start": 292,
"end": 308,
"text": "Liu and Wu 2018)",
"ref_id": "BIBREF21"
},
{
"start": 389,
"end": 415,
"text": "(Tschiatschek et al. 2018)",
"ref_id": "BIBREF41"
},
{
"start": 572,
"end": 589,
"text": "(Shu et al. 2017)",
"ref_id": "BIBREF36"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background and Related Work",
"sec_num": "2."
},
{
"text": "\"Fake News\": Our Working Definition. Our attackers focus on automatically introducing false information into otherwise trustworthy content. We call the resultant manipulated articles fake news. This definition matches that of Zhou and Zafarani (2018) and is in line with the disinformation focused view of Wardle and Derakhshan (2017) . Also, this is in line with how false claims are represented in many human-generated fake news data sets Augenstein et al. 2019) . Conversely, Zellers et al. (2019) focus on entirely fabricated articles, a different type of fake news where the goal is mostly to create \"viral and persuasive\" content.",
"cite_spans": [
{
"start": 226,
"end": 250,
"text": "Zhou and Zafarani (2018)",
"ref_id": "BIBREF52"
},
{
"start": 306,
"end": 334,
"text": "Wardle and Derakhshan (2017)",
"ref_id": "BIBREF48"
},
{
"start": 441,
"end": 464,
"text": "Augenstein et al. 2019)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "Our choice of creating articles with only a limited number of false statements is aligned with the way humans tend to deceive or lie. Psychological studies (Mazar, Amir, and Ariely 2008) support the age-old notion that, when lying, \"the best policy for the criminal is to tell the truth as nearly as possible.\" 3 This helps preserve an honest selfimage and, perhaps more importantly, reduces the chances of the lie being detected. For example, a study on the longest-surviving known fake Wikipedia article (Benjakob 2019) 4 revealed that many of the presented facts were only slightly altered from other, true facts.",
"cite_spans": [
{
"start": 156,
"end": 186,
"text": "(Mazar, Amir, and Ariely 2008)",
"ref_id": "BIBREF23"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "SEOUL, South Korea -North Korea's leader, Kim said on Sunday that his country was making final preparations to conduct its first test of an intercontinental ballistic missile -a bold statement less than a month before the inauguration of Donald J. Trump. Although North Korea has conducted five nuclear tests in the last decade and more than 20 ballistic missile tests in 2016 alone, and although it habitually threatens to attack the United States with nuclear weapons, the country has never an intercontinental ballistic missile, or ICBM. <...> In his speech, Mr. Kim did not comment on Mr. Trump's election. Doubt still runs deep that North Korea has mastered all the technology needed to build a reliable ICBM. But analysts in the region said the North's launchings of rockets to put satellites into orbit in recent years showed that the country had cleared some key technological hurdles. After the North's satellite launch in February, South Korean defense officials said the Unha rocket used in the launch, if successfully reconfigured as a missile, could fly more than 7, 400 miles with a warhead of 1, 100 to 1, 300 pounds -far enough to reach most of the United States. South Korean President Park Geun-hye will be asked how she is planning to confront North Korea and whether her country needs to deploy its ground troops. It also is unlikely that she will deploy U.S. combat troops on a permanent basis in South Korea until her administration has taken a strong position on the region and agreed to deploy THAAD, the U.S. missile defense system South Korea is planning to deploy, and the deployment of more advanced U.S. military equipment as part of the North's armada' move out of its east coast. Mr. Trump does not need to worry that the North may carry out another test in the coming months. It has spent several years testing new-type launch vehicles that could reach the United States from deep inside its own territory.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "Title: Fernandez defends Argentine grain export tax President Cristina Fernandez on Tuesday defended an increase in export taxes on grains that has riled many farmers, and she called on them to respect the law in protesting her policies. <\u2026> In a concession to her critics, Fernandez said the increase in taxes on exports of grains that she instituted in March by decree will be debated by Congress. But there is little likelihood that the Congress will order major changes, since her party controls both houses. But Hilda Duhalde, an opponent of Fernandez, was not persuaded. \"It's true that they have a majority in both houses, but we have to put white on black and watch out for the small-and medium-sized producers, who are the ones suffering,\" she said. Argentina raised export taxes in March by more than 10 percent. Fernandez has said growers have benefited from rising world prices and the profits should be spread to help the poor. Farmers have countered that they need to reinvest the profits and that the higher taxes make it difficult for them to make a living. Fernandez said she was open to dialogue, but a dialogue that does not countenance the blocking of roads or other disruptions to the lives of Argentines. \"Democracy for the people, not the corporations,\" she said.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "We attempt to answer: Who appealed for dialogue and respect?",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "Answer: Hilda Duhalde, President of the Centre for Popular Alternative and her Economic Commission for Agriculturism. (fake; President Cristina Fernandez)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "We attempt to answer: What do farmers say higher taxes do? Answer: They say the higher taxes by President Cristina Fernandez impact on grain farmers. (real) Title: Nominee Betsy DeVos's Knowledge of Education Basics Is Open to Criticism Until Tuesday, the fight over Betsy DeVos's nomination to be secretary of education revolved mostly around her support of contentious school choice programs. But her confirmation hearing that night opened her up to new criticism: <\u2026> Ms.",
"cite_spans": [
{
"start": 150,
"end": 156,
"text": "(real)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "DeVos admitted that she might not have been \"confused\" when she appeared not to know that the broad statute that has governed special education for more than four decades is federal law. <\u2026> She appeared blank on basic education terms. Asked how school performance should be assessed, she did not know the difference between growth, which measures how much students have learned over a given period, and proficiency, which measures how many students reach a targeted score. Ms. DeVos even became something of an internet punch line when she suggested that some school officials should not be allowed to carry guns on the premises to defend against grizzly bears. <\u2026> But her statements on special education could make her vulnerable families of children with special needs are a vocal lobby, one that Republicans do not want to alienate. <\u2026> Senator Tim Kaine of Virginia, last year's Democratic nominee for vice president, asked Ms. DeVos whether schools that receive tax dollars should be required to meet the requirements of IDEA. \"I think that is a matter that's best left to the states,\" Ms. DeVos replied. Mr. Kaine came back: \"So some states might be good to kids with disabilities, and other states might not be so good, and then what? People can just move around the country if they don't like how their kids are being treated?\" Ms. DeVos repeated, \"I think that is an issue that's best left to the states. \" \"It's not federal law,\" an exasperated Mr. Kaine replied. <\u2026> \"Do you think families should have recourse in the courts if schools don't meet their needs?\" she asked. \"Senator, I assure you that if confirmed I will be very sensitive to the needs of special needs students,\" Ms. DeVos said. \"It's not about sensitivity, although that helps,\" Ms. Hassan countered. <\u2026> (c) Vanilla extension (b) Article modification (a) QA extension",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Adversarial Setting",
"sec_num": "3."
},
{
"text": "Examples of the fake class in our experiments. (a) In the news question answering (Section 4), a CNN article is presented with two examples of questions (bold) from newsQA (Trischler et al. 2017 ) and Grover's generated answer (red) . The first answer is verified by a human annotator to be false and the second as true. (b) In article modification (m = 6) (Section 4), the negations are marked with a cross-line for deletions and underline for addition. (c) In the article extension case (Section 5), the bold red text is the generation of GPT-2 Medium to extend the prefix.",
"cite_spans": [
{
"start": 172,
"end": 194,
"text": "(Trischler et al. 2017",
"ref_id": "BIBREF40"
},
{
"start": 227,
"end": 232,
"text": "(red)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 1",
"sec_num": null
},
{
"text": "Attack and Defense Capabilities. We adopt an adversarial setting similar to that of Zellers et al. (2019) . Our attacker wishes to generate fake text, that contains unverified or false claims, en masse, using a language model to automate the process. The attacker's goal is to produce fake text that the verifier classifies as real (see Figure 1 for examples). Our verifier is adaptive: It receives a limited set of examples generated by the attacker, and trains a discriminator to detect the attacker's texts from legitimately produced, real text, containing exclusively human-verified claims (news articles from relatively reputable sources, like the The New York Times, are assumed to be real). We also experiment with a non-adaptive, zero-shot setting, where the verifier does not receive the attacker's examples.",
"cite_spans": [
{
"start": 84,
"end": 105,
"text": "Zellers et al. (2019)",
"ref_id": "BIBREF50"
}
],
"ref_spans": [
{
"start": 337,
"end": 345,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Figure 1",
"sec_num": null
},
{
"text": "Training and Evaluating the Detector. In each experiment, we collected a data set with a \"real\" text class and a \"fake\" text class and used separate samples for testing and for fine-tuning. We used a Grover-Mega discriminator for all of the experiments. The model's weights were initialized from a checkpoint provided by Zellers et al. (2019) and fine-tuned for 10 epochs with our training samples. For evaluating the zero-shot defense, we applied a pretrained Grover-Mega discriminator by querying its Web interface. We report human performance on some of the attacks.",
"cite_spans": [
{
"start": 321,
"end": 342,
"text": "Zellers et al. (2019)",
"ref_id": "BIBREF50"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Figure 1",
"sec_num": null
},
{
"text": "We create two data sets, simulating two different uses of LMs to automatically produce fake news. In the first, the extension scenario, an auto-completion text generator extends a news article. A responsible user of this generator verifies the correctness of the output (producing real text), whereas an attacker verifies incorrectness (producing fake text).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Fails to Detect Machine-Generated Misinformation",
"sec_num": "4."
},
{
"text": "In the second, the modification scenario, the attacker uses a human-written news article and performs subtle modifications to semantically modify statements. Specifically, we add and remove negations. This follows the intuition that, if we take care to add negations in a syntactically correct fashion, the new sentence is a negative inversion of the original (Rudanko 1982 ). Yet, such changes are subtle enough to retain the original article's style and distributional features. See below for the full details on the creation of the data sets.",
"cite_spans": [
{
"start": 360,
"end": 373,
"text": "(Rudanko 1982",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Fails to Detect Machine-Generated Misinformation",
"sec_num": "4."
},
{
"text": "Additionally, we used about 100 examples from each data set to test human performance in detecting this form of misinformation. For the extension scenario, we assigned two subject volunteers with two different tasks. The first had the same task as Grover's detector, and the second was allowed to use external sources to verify facts (which Grover cannot do). For the modification scenario, the participants had the same task as Grover but we highlighted negations to help them focus on the relevant parts.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Fails to Detect Machine-Generated Misinformation",
"sec_num": "4."
},
{
"text": "(1) Creating the Extension Data Set. We use the newsQA data set (Trischler et al. 2017 ) that contains CNN articles with corresponding questions and answers (segments copied verbatim from the text). Following Radford et al. (2019) and von Davier (2019), we embed the question in a template that is appended to the end of the input text:",
"cite_spans": [
{
"start": 64,
"end": 86,
"text": "(Trischler et al. 2017",
"ref_id": "BIBREF40"
},
{
"start": 209,
"end": 230,
"text": "Radford et al. (2019)",
"ref_id": "BIBREF31"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Fails to Detect Machine-Generated Misinformation",
"sec_num": "4."
},
{
"text": "We attempt to answer: <question> Answer:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Fails to Detect Machine-Generated Misinformation",
"sec_num": "4."
},
{
"text": "Then, the first sentence produced by Grover's generator is used as a potential answer. We choose this template by examining common formats of questions in news articles and finding the one that empirically produces mostly reasonable answers by Grover's generator. For example, given text about a suspect caught in Washington, DC, and the question \"Where was the suspect stopped?\", Grover answered \"2 blocks from the U.S.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Fails to Detect Machine-Generated Misinformation",
"sec_num": "4."
},
{
"text": "We manually labeled the generated answers. We filtered out nonsensical ones (29%) and labeled the rest as real or fake texts by correctness. Answers containing at least one false statement (judging by the article and/or external knowledge) were labeled fake. We measured inter-annotator agreement on a subset of 100 examples labeled by two different annotators, and found it to be substantial (Cohen's kappa score of k = 0.78).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Capitol.\"",
"sec_num": null
},
{
"text": "Finally, we remove the sentence containing the answer from each article by removing the sentence with the highest TF-IDF-weighted word-count similarity with the question and answer (concatenated).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Capitol.\"",
"sec_num": null
},
{
"text": "(2) Creating the Modifications Data Set. We use articles from The New York Times (Thompson 2017) . We perform the modifications by removing and adding negations from statements: first, we randomly delete m/2 \"not\" or \"no\" occurrences from the text. Then, we automatically find statements to add m/2 negations to, using the probabilities of a GPT-2 Medium language model. We randomly sample 100 locations in the article and choose the ones with the maximal score, defined as the probability for either \"not\" or \"no,\" multiplied by the probability for the word following the negation. By both adding and removing, we keep the total number of negations unchanged, avoiding anomalous negations counts. Original New York Times articles are labeled as real and modified ones as fake.",
"cite_spans": [
{
"start": 81,
"end": 96,
"text": "(Thompson 2017)",
"ref_id": "BIBREF37"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Capitol.\"",
"sec_num": null
},
{
"text": "Results: Section 4. We report (macro) F1 score and overall accuracy, as well as precision and recall of the \"fake\" class. Zero-shot performance (not included) was very low in all cases. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Table 1",
"sec_num": null
},
{
"text": "The detector's performance figures are given in Table 1 . Against QA extension, the detector reached 71% accuracy in distinguishing true vs. false extensions generated by Grover. Although this is above a trivial majority baseline (51%), some of the success could be a side effect of the way that the data set was constructed. Labeling instances with at least one false statement as fake makes these sentences generally longer, allowing a simple linear classifier that only uses the extension's length to reach 56%. When evaluated only on the short false answers (up to 10 words), the accuracy drops from 71% to only 62%. Ultimately, however, despite performing better than simple baselines, the detection in this setting is much worse than in distinguishing human from machine text (Section 5).",
"cite_spans": [],
"ref_spans": [
{
"start": 48,
"end": 55,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "4.1"
},
{
"text": "Comparing this with the results of the human evaluation is particularly instructive: The first subject, given the same task as the detector, attained an F1 score of 0.68, only marginally below the detector. The second participant, who was allowed to use external sources, attained an F1 of 0.84. We conclude that the stylometry-based classifier does not perform much better than humans in detecting potential misinformation, and that verifying against other resources can drastically improve results.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "4.1"
},
{
"text": "For the automatic article modification attack, the classifier fails completely to detect two modifications (m = 2), and is not much better even when m = 10 statements are inverted. Even though the change includes negations that might introduce surprising factual claims, the language style remains too similar to detect it. Humans, evaluated with m = 2, had an F1 of 0.74, with 0.91 and 0.59 recall for real and modified articles, respectively. Thus, although humans perform better than the model, many of the modified statements were interpreted as real. Manually examining these, we found that the added negation actually changed the meaning in 60% of the instances that were misclassified. Overall, the results show that both attacks can produce articles containing misinformation that evade the stylometry detector and mislead humans.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "4.1"
},
{
"text": "One might suspect that the low performance in Section 4 is due to limited capacity of our detector or the small fraction of generated text. We now show that the detector performs well on the provenance task and is sensitive to differences in small text portions.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Detects Machine-Human Impersonations",
"sec_num": "5."
},
{
"text": "(1) Fully Generated Articles. We perform a first evaluation of Grover against texts generated by a different model of similar size, namely, GPT-2 XL (Radford et al. 2019) . We include examples from the WebText test set (Radford et al. 2019) , labeled real text, and the released unconditioned generations of the GPT-2 XL model, labeled fake text.",
"cite_spans": [
{
"start": 149,
"end": 170,
"text": "(Radford et al. 2019)",
"ref_id": "BIBREF31"
},
{
"start": 219,
"end": 240,
"text": "(Radford et al. 2019)",
"ref_id": "BIBREF31"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Stylometry Detects Machine-Human Impersonations",
"sec_num": "5."
},
{
"text": "Detection results on the attacks of Section 5 in a zero-shot setting and the adaptive setting (where the discriminator is fine-tuned to a specific attacker). We report (macro) F1 score and overall accuracy. Precision, and recall of the \"fake\" class are reported as well. (2) Automatic Article Extension (Vanilla). In order to experiment with mixed humanmachine articles, we build an attacker that automatically extends unfinished humanwritten articles. We simulate this attacker by querying a GPT-2 Medium model, conditioned on the first 500 words from an article. We include New York Times articles, truncated to match the length of the fake class, labeled as real text, and articles that were automatically extended, labeled fake text. We use different percentages (g) of machinegenerated text, by stopping the generation after an appropriate number of sentences.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Table 2",
"sec_num": null
},
{
"text": "(3) Automatic QA-based Extension. In this experiment, we test whether the QA defense from Section 4 can perform better if we can assume that true texts are always humanwritten. To this end, we repeat the same setting but use the gold answers from the newsQA data set for the \"real text\" class.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Provenance detection",
"sec_num": null
},
{
"text": "Zero-shot Setting. As Table 2 shows, the zero-shot classifier is effective in detecting the fully generated articles of a different model, with a 0.9 F-1 score. The extended articles, containing a substantial amount of human-written text, are mostly classified as human-written in this setting. This is unsurprising considering the data set of full articles the detector was originally trained on.",
"cite_spans": [],
"ref_spans": [
{
"start": 22,
"end": 29,
"text": "Table 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Results",
"sec_num": "5.1"
},
{
"text": "Adaptive Setting. After fine-tuning, Grover improves on the full articles and performs well against article extension generations, reaching 0.94 F1 score for articles with a single generated sentence (g = 1%). This stylometry detector is thus highly effective in distinguishing human from machine.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.1"
},
{
"text": "In the QA setting, also containing a single machine-generated sentence, the detector performs worse by 0.11 F1 points. This might be due to using the (stronger) Grover-Mega generator for the QA setting, and because the appended template for QA was selected to empirically maximize human \"reasonableness\" scores. Still, this setting allows the stylometry detector to reach much higher performance than in the veracitybased setting (Section 4). This indicates that a restricted benchmark, which assumes no legitimate use of LMs, might not reflect the misinformation detection performance if the model is deployed in a world where LMs are used for both legitimate and malicious purposes.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Results",
"sec_num": "5.1"
},
{
"text": "Advancements in LM technology and their various applications have introduced a new challenge: distinguishing truthful text from misinformation, when the text is generated or edited by an LM. Our experiments indicate that LM-generated falsified texts are very similar in style to LM-generated texts containing true content. As a result, stylometrybased classifiers cannot identify auto-generated intentionally misleading content.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "6."
},
{
"text": "We conclude with the following recommendations:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "6."
},
{
"text": "(1) Extending Veracity-based Benchmarks. In order to better evaluate detectors against LM-generated misinformation, we recommend extending our benchmarks by creating other veracity-oriented data sets, that represent a wide range of LM applications, from whole-article generation to forms of hybrid writing and editing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "6."
},
{
"text": "(2)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "6."
},
{
"text": "Improving Non-stylometry Methods. Other detection approaches, as surveyed at the end of Section 2, are less affected by the use of LMs. Therefore, advancements in such methods can improve the detection of both human-and machine-generated misinformation. Notably, the fact-checking setting makes fewer assumptions on the available auxiliary information and can be applied even if the text was sent to the verifier through a private channel such as e-mail. However, because fact-checking requires advanced inference capabilities, incorporating non-textual information, when available, can yield better results.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "6."
},
{
"text": "The potential use of LMs in creating fake news calls for a re-evaluation of current defense strategies. We examine the state-of-the-art stylometry model, and find it effective in preventing impersonation, but limited in detecting LM-generated misinformation. This new kind of misinformation could be created by the same model that is used by legitimate writers as a writing-assistance tool, hiding stylistic differences between falsified and truthful content. This motivates (1) constructing more instructive benchmarks for NLP-based approaches and improving non-stylistic methods, and (2) addressing a set of challenges that spans many disciplines beyond NLP, including social networks, information security, human-computer interaction, and others.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "7."
},
{
"text": "https://github.com/re-search/DocProduct; https://openai.com/blog/openai-api/.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Raskolnikov, \"Crime and Punishment\"(Dostoyevsky 1866). 4 https://en.wikipedia.org/wiki/Wikipedia:List of hoaxes on Wikipedia.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "We thank the anonymous reviewers and the members of the MIT NLP group for their helpful comments. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgments",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "In Automatic IQ estimation using stylometric methods",
"authors": [
{
"first": "Polina",
"middle": [],
"last": "Abramov",
"suffix": ""
},
{
"first": "Roman",
"middle": [
"V"
],
"last": "Shafran",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Yampolskiy",
"suffix": ""
}
],
"year": 2019,
"venue": "Handbook of Research on Learning in the Age of Transhumanism. IGI Global",
"volume": "",
"issue": "",
"pages": "32--45",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Abramov, Polina Shafran and Roman V. Yampolskiy. 2019. In Automatic IQ estimation using stylometric methods. In Handbook of Research on Learning in the Age of Transhumanism. IGI Global, pages 32-45, Hershey, PA.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Detecting hoaxes, frauds, and deception in writing style online",
"authors": [
{
"first": "Sadia",
"middle": [],
"last": "Afroz",
"suffix": ""
},
{
"first": "Michael",
"middle": [],
"last": "Brennan",
"suffix": ""
},
{
"first": "Rachel",
"middle": [],
"last": "Greenstadt",
"suffix": ""
}
],
"year": 2012,
"venue": "Proceedings of 2012 IEEE Symposium on Security and Privacy",
"volume": "",
"issue": "",
"pages": "461--475",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Afroz, Sadia, Michael Brennan, and Rachel Greenstadt. 2012. Detecting hoaxes, frauds, and deception in writing style online. In Proceedings of 2012 IEEE Symposium on Security and Privacy. pages 461-475, San Francisco, CA, USA.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Doppelg\u00e4nger finder: Taking stylometry to the underground",
"authors": [
{
"first": "Sadia",
"middle": [],
"last": "Afroz",
"suffix": ""
},
{
"first": "Aylin",
"middle": [],
"last": "Caliskan",
"suffix": ""
},
{
"first": "Rachel",
"middle": [],
"last": "Greenstadt",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Mcoy",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of 2014 IEEE Symposium on Security and Privacy",
"volume": "",
"issue": "",
"pages": "212--226",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Afroz, Sadia, Aylin Caliskan, Rachel Greenstadt, and Dan Mcoy. 2014. Doppelg\u00e4nger finder: Taking stylometry to the underground. In Proceedings of 2014 IEEE Symposium on Security and Privacy, pages 212-226, IEEE Computer Society, San Jose, CA, USA.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "MultiFC: A real-world multi-domain dataset for evidence-based fact checking of claims",
"authors": [
{
"first": "Isabelle",
"middle": [],
"last": "Augenstein",
"suffix": ""
},
{
"first": "Christina",
"middle": [],
"last": "Lioma",
"suffix": ""
},
{
"first": "Dongsheng",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "Lucas",
"middle": [
"Chaves"
],
"last": "Lima",
"suffix": ""
},
{
"first": "Casper",
"middle": [],
"last": "Hansen",
"suffix": ""
},
{
"first": "Christian",
"middle": [],
"last": "Hansen",
"suffix": ""
},
{
"first": "Jakob",
"middle": [
"Grue"
],
"last": "Simonsen",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
"volume": "",
"issue": "",
"pages": "4685--4697",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Augenstein, Isabelle, Christina Lioma, Dongsheng Wang, Lucas Chaves Lima, Casper Hansen, Christian Hansen, and Jakob Grue Simonsen. 2019. MultiFC: A real-world multi-domain dataset for evidence-based fact checking of claims. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4685-4697, Hong Kong.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Real or fake? Learning to discriminate machine from human generated text",
"authors": [
{
"first": "Anton",
"middle": [],
"last": "Bakhtin",
"suffix": ""
},
{
"first": "Sam",
"middle": [],
"last": "Gross",
"suffix": ""
},
{
"first": "Myle",
"middle": [],
"last": "Ott",
"suffix": ""
},
{
"first": "Yuntian",
"middle": [],
"last": "Deng",
"suffix": ""
},
{
"first": "Marc' Aurelio",
"middle": [],
"last": "Ranzato",
"suffix": ""
},
{
"first": "Arthur",
"middle": [],
"last": "Szlam",
"suffix": ""
}
],
"year": 2019,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1906.03351"
]
},
"num": null,
"urls": [],
"raw_text": "Bakhtin, Anton, Sam Gross, Myle Ott, Yuntian Deng, Marc' Aurelio Ranzato, and Arthur Szlam. 2019. Real or fake? Learning to discriminate machine from human generated text. arXiv preprint arXiv:1906.03351.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "The fake Nazi death camp: Wikipedia's longest hoax",
"authors": [
{
"first": "Omer",
"middle": [],
"last": "Benjakob",
"suffix": ""
}
],
"year": 2019,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Benjakob, Omer. 2019. The fake Nazi death camp: Wikipedia's longest hoax, exposed. https://tinyurl.com/y3m2wd9k. Haaretz.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Language of lies in prison: Linguistic classification of prisoners' truthful and deceptive natural language",
"authors": [
{
"first": "Gary",
"middle": [
"D"
],
"last": "Bond",
"suffix": ""
},
{
"first": "Adrienne",
"middle": [
"Y"
],
"last": "Lee",
"suffix": ""
}
],
"year": 2005,
"venue": "Applied Cognitive Psychology",
"volume": "19",
"issue": "3",
"pages": "313--329",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bond, Gary D. and Adrienne Y. Lee. 2005. Language of lies in prison: Linguistic classification of prisoners' truthful and deceptive natural language. Applied Cognitive Psychology, 19(3):313-329.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Adversarial stylometry: Circumventing authorship recognition to preserve privacy and anonymity",
"authors": [
{
"first": "Michael",
"middle": [],
"last": "Brennan",
"suffix": ""
},
{
"first": "Sadia",
"middle": [],
"last": "Afroz",
"suffix": ""
},
{
"first": "Rachel",
"middle": [],
"last": "Greenstadt",
"suffix": ""
}
],
"year": 2012,
"venue": "ACM Transactions on Information and System Security (TISSEC)",
"volume": "15",
"issue": "3",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Brennan, Michael, Sadia Afroz, and Rachel Greenstadt. 2012. Adversarial stylometry: Circumventing authorship recognition to preserve privacy and anonymity. ACM Transactions on Information and System Security (TISSEC), 15(3):12.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "De-anonymizing programmers via code stylometry",
"authors": [
{
"first": "Aylin",
"middle": [],
"last": "Caliskan-Islam",
"suffix": ""
},
{
"first": "Richard",
"middle": [],
"last": "Harang",
"suffix": ""
},
{
"first": "Andrew",
"middle": [],
"last": "Liu",
"suffix": ""
},
{
"first": "Arvind",
"middle": [],
"last": "Narayanan",
"suffix": ""
},
{
"first": "Clare",
"middle": [],
"last": "Voss",
"suffix": ""
},
{
"first": "Fabian",
"middle": [],
"last": "Yamaguchi",
"suffix": ""
},
{
"first": "Rachel",
"middle": [],
"last": "Greenstadt",
"suffix": ""
}
],
"year": 2015,
"venue": "24th USENIX Security Symposium (USENIX Security 15)",
"volume": "",
"issue": "",
"pages": "255--270",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Caliskan-Islam, Aylin, Richard Harang, Andrew Liu, Arvind Narayanan, Clare Voss, Fabian Yamaguchi, and Rachel Greenstadt. 2015. De-anonymizing programmers via code stylometry. In 24th USENIX Security Symposium (USENIX Security 15), pages 255-270, Washington, D.C.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Information credibility on Twitter",
"authors": [
{
"first": "Carlos",
"middle": [],
"last": "Castillo",
"suffix": ""
},
{
"first": "Marcelo",
"middle": [],
"last": "Mendoza",
"suffix": ""
},
{
"first": "Barbara",
"middle": [],
"last": "Poblete",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the 20th International Conference on World Wide Web",
"volume": "",
"issue": "",
"pages": "67--684",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Castillo, Carlos, Marcelo Mendoza, and Barbara Poblete. 2011. Information credibility on Twitter. In Proceedings of the 20th International Conference on World Wide Web, pages 67-684.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "The language of legal and illegal activity on the Darknet",
"authors": [
{
"first": "Leshem",
"middle": [],
"last": "Choshen",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Eldad",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Hershcovich",
"suffix": ""
},
{
"first": "Elior",
"middle": [],
"last": "Sulem",
"suffix": ""
},
{
"first": "Omri",
"middle": [],
"last": "Abend",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
"volume": "",
"issue": "",
"pages": "4271--4279",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Choshen, Leshem, Dan Eldad, Daniel Hershcovich, Elior Sulem, and Omri Abend. 2019. The language of legal and illegal activity on the Darknet. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4271-4279, Florence.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Generating medical certification items by fine-tuning OpenAI's gpt2 transformer model",
"authors": [
{
"first": "",
"middle": [],
"last": "Von Davier",
"suffix": ""
},
{
"first": ". ; M",
"middle": [
"D"
],
"last": "Matthias",
"suffix": ""
}
],
"year": 2019,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1908.08594"
]
},
"num": null,
"urls": [],
"raw_text": "von Davier, Matthias. 2019. Training optimus prime, M.D.: Generating medical certification items by fine-tuning OpenAI's gpt2 transformer model. arXiv preprint arXiv:1908.08594.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Crime and Punishment. The Russian Messenger",
"authors": [
{
"first": "Fyodor",
"middle": [],
"last": "Dostoyevsky",
"suffix": ""
}
],
"year": null,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Dostoyevsky, Fyodor. 1866. Crime and Punishment. The Russian Messenger.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Detecting deception using critical segments",
"authors": [
{
"first": "Frank",
"middle": [],
"last": "Enos",
"suffix": ""
},
{
"first": "Elizabeth",
"middle": [],
"last": "Shriberg",
"suffix": ""
},
{
"first": "Martin",
"middle": [],
"last": "Graciarena",
"suffix": ""
},
{
"first": "Julia",
"middle": [],
"last": "Hirschberg",
"suffix": ""
},
{
"first": "Andreas",
"middle": [],
"last": "Stolcke",
"suffix": ""
}
],
"year": 2007,
"venue": "International Speech Communication Association",
"volume": "",
"issue": "",
"pages": "2281--2284",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Enos, Frank, Elizabeth Shriberg, Martin Graciarena, Julia Hirschberg, and Andreas Stolcke. 2007. Detecting deception using critical segments. In International Speech Communication Association, pages 2281-2284, Antwerp, Belgium.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Syntactic stylometry for deception detection",
"authors": [
{
"first": "Song",
"middle": [],
"last": "Feng",
"suffix": ""
},
{
"first": "Ritwik",
"middle": [],
"last": "Banerjee",
"suffix": ""
},
{
"first": "Yejin",
"middle": [],
"last": "Choi",
"suffix": ""
}
],
"year": 2012,
"venue": "Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics",
"volume": "2",
"issue": "",
"pages": "171--175",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Feng, Song, Ritwik Banerjee, and Yejin Choi. 2012. Syntactic stylometry for deception detection. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Short Papers -Volume 2, ACL '12, pages 171-175.",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Handbook of Science and Technology for Homeland Security",
"authors": [
{
"first": "Mark",
"middle": [
"G"
],
"last": "Frank",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Melissa",
"suffix": ""
},
{
"first": "Maureen O'",
"middle": [],
"last": "Menasco",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Sullivan",
"suffix": ""
}
],
"year": 2008,
"venue": "",
"volume": "",
"issue": "",
"pages": "1--13",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Frank, Mark G., Melissa A. Menasco, and Maureen O'Sullivan. 2008. Human behavior and deception detection. In John G. Voeller, editor, Wiley Handbook of Science and Technology for Homeland Security. Wiley & Sons, Inc., pages 1-13.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "GLTR: Statistical detection and visualization of generated text",
"authors": [
{
"first": "Sebastian",
"middle": [],
"last": "Gehrmann",
"suffix": ""
},
{
"first": "Hendrik",
"middle": [],
"last": "Strobelt",
"suffix": ""
},
{
"first": "Alexander",
"middle": [],
"last": "Rush",
"suffix": ""
},
{
"first": ";",
"middle": [],
"last": "Florence",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Goswami",
"suffix": ""
},
{
"first": "Sudeshna",
"middle": [],
"last": "Sumit",
"suffix": ""
},
{
"first": "Mayur",
"middle": [],
"last": "Sarkar",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Rustagi",
"suffix": ""
}
],
"year": 2009,
"venue": "Proceedings of the 57th Conference of the Association for Computational Linguistics: System Demonstrations",
"volume": "",
"issue": "",
"pages": "111--116",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gehrmann, Sebastian, Hendrik Strobelt, and Alexander Rush. 2019. GLTR: Statistical detection and visualization of generated text. In Proceedings of the 57th Conference of the Association for Computational Linguistics: System Demonstrations, pages 111-116, Florence. Goswami, Sumit, Sudeshna Sarkar, and Mayur Rustagi. 2009. Stylometric analysis of bloggers' age and gender. In International AAAI Conference on Web and Social Media, San Jose, CA.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "TweetCred: Real-time credibility assessment of content on Twitter",
"authors": [
{
"first": "Aditi",
"middle": [],
"last": "Gupta",
"suffix": ""
},
{
"first": "Ponnurangam",
"middle": [],
"last": "Kumaraguru",
"suffix": ""
},
{
"first": "Carlos",
"middle": [],
"last": "Castillo",
"suffix": ""
},
{
"first": "Patrick",
"middle": [],
"last": "Meier",
"suffix": ""
}
],
"year": 2014,
"venue": "International Conference on Social Informatics",
"volume": "",
"issue": "",
"pages": "228--243",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gupta, Aditi, Ponnurangam Kumaraguru, Carlos Castillo, and Patrick Meier. 2014. TweetCred: Real-time credibility assessment of content on Twitter. In International Conference on Social Informatics, pages 228-243, Cham.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "A retrospective analysis of the fake news challenge stance-detection task",
"authors": [
{
"first": "Andreas",
"middle": [],
"last": "Hanselowski",
"suffix": ""
},
{
"first": "P",
"middle": [
"V S"
],
"last": "Avinesh",
"suffix": ""
},
{
"first": "Benjamin",
"middle": [],
"last": "Schiller",
"suffix": ""
},
{
"first": "Felix",
"middle": [],
"last": "Caspelherr",
"suffix": ""
},
{
"first": "Debanjan",
"middle": [],
"last": "Chaudhuri",
"suffix": ""
},
{
"first": "Christian",
"middle": [
"M"
],
"last": "Meyer",
"suffix": ""
},
{
"first": "Iryna",
"middle": [],
"last": "Gurevych",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the 27th International Conference on Computational Linguistics",
"volume": "",
"issue": "",
"pages": "1859--1874",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hanselowski, Andreas, P. V. S. Avinesh, Benjamin Schiller, Felix Caspelherr, Debanjan Chaudhuri, Christian M. Meyer, and Iryna Gurevych. 2018. A retrospective analysis of the fake news challenge stance-detection task. In Proceedings of the 27th International Conference on Computational Linguistics, pages 1859-1874, Santa Fe, NM.",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "2019. I, language robot",
"authors": [
{
"first": "Patrick",
"middle": [],
"last": "House",
"suffix": ""
}
],
"year": null,
"venue": "Los Angeles Review of Books",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "House, Patrick. 2019. I, language robot. https://lareviewofbooks.org/article/ i-language-robot/. Los Angeles Review of Books.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "All-in-one: Multi-task learning for rumour verification",
"authors": [
{
"first": "Elena",
"middle": [],
"last": "Kochkina",
"suffix": ""
},
{
"first": "Maria",
"middle": [],
"last": "Liakata",
"suffix": ""
},
{
"first": "Arkaitz",
"middle": [],
"last": "Zubiaga",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the 27th International Conference on Computational Linguistics",
"volume": "",
"issue": "",
"pages": "3402--3413",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kochkina, Elena, Maria Liakata, and Arkaitz Zubiaga. 2018. All-in-one: Multi-task learning for rumour verification. In Proceedings of the 27th International Conference on Computational Linguistics, pages 3402-3413, Santa Fe, NM.",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "Early detection of fake news on social media through propagation path classification with recurrent and convolutional networks",
"authors": [
{
"first": "Yang",
"middle": [],
"last": "Liu",
"suffix": ""
},
{
"first": "Yi-Fang",
"middle": [],
"last": "Wu",
"suffix": ""
}
],
"year": 2018,
"venue": "AAAI Conference on Artificial Intelligence",
"volume": "",
"issue": "",
"pages": "354--361",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Liu, Yang and Yi-Fang Wu. 2018. Early detection of fake news on social media through propagation path classification with recurrent and convolutional networks. In AAAI Conference on Artificial Intelligence, pages 354-361, New Orleans, LA.",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Evaluating truthfulness and detecting deception",
"authors": [
{
"first": "David",
"middle": [],
"last": "Matsumoto",
"suffix": ""
},
{
"first": "Hyi",
"middle": [
"Sung"
],
"last": "Hwang",
"suffix": ""
},
{
"first": "Lisa",
"middle": [],
"last": "Skinner",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Frank",
"suffix": ""
}
],
"year": 2011,
"venue": "FBI Law Enforcement Bulletin",
"volume": "",
"issue": "80",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Matsumoto, David,, Hyi Sung Hwang, Lisa Skinner, and Mark Frank. 2011. Evaluating truthfulness and detecting deception. FBI Law Enforcement Bulletin, (80):1.",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "The dishonesty of honest people: A theory of self-concept maintenance",
"authors": [
{
"first": "Nina",
"middle": [],
"last": "Mazar",
"suffix": ""
},
{
"first": "On",
"middle": [],
"last": "Amir",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Ariely",
"suffix": ""
}
],
"year": 2008,
"venue": "Journal of Marketing Research",
"volume": "45",
"issue": "6",
"pages": "633--644",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mazar, Nina, On Amir, and Dan Ariely. 2008. The dishonesty of honest people: A theory of self-concept maintenance. Journal of Marketing Research, 45(6):633-644.",
"links": null
},
"BIBREF24": {
"ref_id": "b24",
"title": "The lie detector: Explorations in the automatic recognition of deceptive language",
"authors": [
{
"first": "Rada",
"middle": [],
"last": "Mihalcea",
"suffix": ""
},
{
"first": "Carlo",
"middle": [],
"last": "Strapparava",
"suffix": ""
}
],
"year": 2009,
"venue": "Proceedings of the ACL-IJCNLP 2009 Conference Short Papers",
"volume": "",
"issue": "",
"pages": "309--312",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mihalcea, Rada and Carlo Strapparava. 2009. The lie detector: Explorations in the automatic recognition of deceptive language. In Proceedings of the ACL-IJCNLP 2009 Conference Short Papers, pages 309-312, Suntec, Singapore.",
"links": null
},
"BIBREF25": {
"ref_id": "b25",
"title": "Inference in an authorship problem",
"authors": [
{
"first": "Frederick",
"middle": [],
"last": "Mosteller",
"suffix": ""
},
{
"first": "David",
"middle": [
"L"
],
"last": "Wallace",
"suffix": ""
}
],
"year": 1963,
"venue": "Journal of the American Statistical Association",
"volume": "58",
"issue": "302",
"pages": "275--309",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mosteller, Frederick and David L. Wallace. 1963. Inference in an authorship problem. Journal of the American Statistical Association, 58(302):275-309.",
"links": null
},
"BIBREF26": {
"ref_id": "b26",
"title": "Language-aware truth assessment of fact candidates",
"authors": [
{
"first": "Ndapandula",
"middle": [],
"last": "Nakashole",
"suffix": ""
},
{
"first": "Tom",
"middle": [
"M"
],
"last": "Mitchell",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics",
"volume": "1",
"issue": "",
"pages": "1009--1019",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Nakashole, Ndapandula and Tom M. Mitchell. 2014. Language-aware truth assessment of fact candidates. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1009-1019, Baltimore, MD.",
"links": null
},
"BIBREF27": {
"ref_id": "b27",
"title": "Surveying stylometry techniques and applications",
"authors": [
{
"first": "Tempestt",
"middle": [
"J"
],
"last": "Neal",
"suffix": ""
},
{
"first": "Aneez",
"middle": [],
"last": "Kalaivani Sundararajan",
"suffix": ""
},
{
"first": "Yiming",
"middle": [],
"last": "Fatima",
"suffix": ""
},
{
"first": "Yingfei",
"middle": [],
"last": "Yan",
"suffix": ""
},
{
"first": "Damon",
"middle": [
"L"
],
"last": "Xiang",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Woodard",
"suffix": ""
}
],
"year": 2017,
"venue": "ACM Computing Surveys",
"volume": "50",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Neal, Tempestt J., Kalaivani Sundararajan, Aneez Fatima, Yiming Yan, Yingfei Xiang, and Damon L. Woodard. 2017. Surveying stylometry techniques and applications. ACM Computing Surveys, 50:86:1-86:36.",
"links": null
},
"BIBREF28": {
"ref_id": "b28",
"title": "Deception detection in news reports in the Russian language: Lexics and discourse",
"authors": [
{
"first": "Myle",
"middle": [],
"last": "Ott",
"suffix": ""
},
{
"first": "Yejin",
"middle": [],
"last": "Choi",
"suffix": ""
},
{
"first": "Claire",
"middle": [],
"last": "Cardie",
"suffix": ""
},
{
"first": "Jeffrey",
"middle": [
"T"
],
"last": "Hancock ; Portland",
"suffix": ""
},
{
"first": "O",
"middle": [
"R"
],
"last": "P\u00e9rez-Rosas",
"suffix": ""
},
{
"first": "Bennett",
"middle": [],
"last": "Ver\u00f3nica",
"suffix": ""
},
{
"first": "Alexandra",
"middle": [],
"last": "Kleinberg",
"suffix": ""
},
{
"first": "Rada",
"middle": [],
"last": "Lefevre",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Mihalcea",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies",
"volume": "",
"issue": "",
"pages": "74--79",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ott, Myle, Yejin Choi, Claire Cardie, and Jeffrey T. Hancock. 2011. Finding deceptive opinion spam by any stretch of the imagination. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pages 309-319, Portland, OR. P\u00e9rez-Rosas, Ver\u00f3nica, Bennett Kleinberg, Alexandra Lefevre, and Rada Mihalcea. 2018. Automatic detection of fake news. In Proceedings of the 27th International Conference on Computational Linguistics, pages 3391-3401, Santa Fe, NM. Pisarevskaya, Dina. 2017. Deception detection in news reports in the Russian language: Lexics and discourse. In Proceedings of the 2017 EMNLP Workshop: Natural Language Processing meets Journalism, pages 74-79, Copenhagen.",
"links": null
},
"BIBREF29": {
"ref_id": "b29",
"title": "DeClarE: Debunking fake news and false claims using evidence-aware deep learning",
"authors": [
{
"first": "Kashyap",
"middle": [],
"last": "Popat",
"suffix": ""
},
{
"first": "Subhabrata",
"middle": [],
"last": "Mukherjee",
"suffix": ""
},
{
"first": "Andrew",
"middle": [],
"last": "Yates",
"suffix": ""
},
{
"first": "Gerhard",
"middle": [],
"last": "Weikum",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Popat, Kashyap, Subhabrata Mukherjee, Andrew Yates, and Gerhard Weikum. 2018. DeClarE: Debunking fake news and false claims using evidence-aware deep learning. In Proceedings of the 2018",
"links": null
},
"BIBREF30": {
"ref_id": "b30",
"title": "A stylometric inquiry into hyperpartisan and fake news",
"authors": [
{
"first": "",
"middle": [],
"last": "Brussels",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Potthast",
"suffix": ""
},
{
"first": "Johannes",
"middle": [],
"last": "Martin",
"suffix": ""
},
{
"first": "Kevin",
"middle": [],
"last": "Kiesel",
"suffix": ""
},
{
"first": "Janek",
"middle": [],
"last": "Reinartz",
"suffix": ""
},
{
"first": "Benno",
"middle": [],
"last": "Bevendorff",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Stein",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics",
"volume": "1",
"issue": "",
"pages": "231--240",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Conference on Empirical Methods in Natural Language Processing, pages 22-32, Brussels. Potthast, Martin, Johannes Kiesel, Kevin Reinartz, Janek Bevendorff, and Benno Stein. 2018. A stylometric inquiry into hyperpartisan and fake news. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 231-240, Melbourne.",
"links": null
},
"BIBREF31": {
"ref_id": "b31",
"title": "Language models are unsupervised multitask learners",
"authors": [
{
"first": "Alec",
"middle": [],
"last": "Radford",
"suffix": ""
},
{
"first": "Jeffrey",
"middle": [],
"last": "Wu",
"suffix": ""
},
{
"first": "Rewon",
"middle": [],
"last": "Child",
"suffix": ""
},
{
"first": "David",
"middle": [],
"last": "Luan",
"suffix": ""
},
{
"first": "Dario",
"middle": [],
"last": "Amodei",
"suffix": ""
},
{
"first": "Ilya",
"middle": [],
"last": "Sutskever",
"suffix": ""
}
],
"year": 2019,
"venue": "OpenAI Blog",
"volume": "",
"issue": "8",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI Blog, 1(8). https://openai.com/blog/ better-language-models/.",
"links": null
},
"BIBREF32": {
"ref_id": "b32",
"title": "Towards a description of negatively conditioned subject operator inversion in English",
"authors": [
{
"first": "Hannah",
"middle": [],
"last": "Rashkin",
"suffix": ""
},
{
"first": "Eunsol",
"middle": [],
"last": "Choi",
"suffix": ""
},
{
"first": "Jin",
"middle": [
"Yea"
],
"last": "Jang",
"suffix": ""
},
{
"first": "Svitlana",
"middle": [],
"last": "Volkova",
"suffix": ""
},
{
"first": "Yejin",
"middle": [],
"last": "Choi",
"suffix": ""
}
],
"year": 1982,
"venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing",
"volume": "63",
"issue": "",
"pages": "348--359",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Rashkin, Hannah, Eunsol Choi, Jin Yea Jang, Svitlana Volkova, and Yejin Choi. 2017. Truth of varying shades: Analyzing language in fake news and political fact-checking. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 2931-2937, Copenhagen. Rudanko, Juhani. 1982. Towards a description of negatively conditioned subject operator inversion in English. English Studies, 63(4):348-359.",
"links": null
},
"BIBREF33": {
"ref_id": "b33",
"title": "Continuous n-gram representations for authorship attribution",
"authors": [
{
"first": "Yunita",
"middle": [],
"last": "Sari",
"suffix": ""
},
{
"first": "Andreas",
"middle": [],
"last": "Vlachos",
"suffix": ""
},
{
"first": "Mark",
"middle": [],
"last": "Stevenson",
"suffix": ""
}
],
"year": 2017,
"venue": "Proceedings of the 15th Conference of the European Chapter",
"volume": "2",
"issue": "",
"pages": "267--273",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Sari, Yunita, Andreas Vlachos, and Mark Stevenson. 2017. Continuous n-gram representations for authorship attribution. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pages 267-273, Valencia.",
"links": null
},
"BIBREF34": {
"ref_id": "b34",
"title": "Towards debiasing fact verification models",
"authors": [
{
"first": "Tal",
"middle": [],
"last": "Schuster",
"suffix": ""
},
{
"first": "Darsh",
"middle": [],
"last": "Shah",
"suffix": ""
},
{
"first": "Yun Jie Serene",
"middle": [],
"last": "Yeo",
"suffix": ""
},
{
"first": "Daniel",
"middle": [],
"last": "Roberto Filizzola",
"suffix": ""
},
{
"first": "Enrico",
"middle": [],
"last": "Ortiz",
"suffix": ""
},
{
"first": "Regina",
"middle": [],
"last": "Santus",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Barzilay",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 2019",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Schuster, Tal, Darsh Shah, Yun Jie Serene Yeo, Daniel Roberto Filizzola Ortiz, Enrico Santus, and Regina Barzilay. 2019. Towards debiasing fact verification models. In Proceedings of the 2019",
"links": null
},
"BIBREF35": {
"ref_id": "b35",
"title": "Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
"authors": [],
"year": null,
"venue": "",
"volume": "",
"issue": "",
"pages": "3419--3425",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3419-3425, Hong Kong.",
"links": null
},
"BIBREF36": {
"ref_id": "b36",
"title": "Fake news detection on social media: A data mining perspective",
"authors": [
{
"first": "Kai",
"middle": [],
"last": "Shu",
"suffix": ""
},
{
"first": "Amy",
"middle": [],
"last": "Sliva",
"suffix": ""
},
{
"first": "Suhang",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "Jiliang",
"middle": [],
"last": "Tang",
"suffix": ""
},
{
"first": "Huan",
"middle": [],
"last": "Liu",
"suffix": ""
}
],
"year": 2017,
"venue": "ACM SIGKDD Explorations Newsletter",
"volume": "19",
"issue": "1",
"pages": "22--36",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Shu, Kai, Amy Sliva, Suhang Wang, Jiliang Tang, and Huan Liu. 2017. Fake news detection on social media: A data mining perspective. ACM SIGKDD Explorations Newsletter, 19(1):22-36.",
"links": null
},
"BIBREF37": {
"ref_id": "b37",
"title": "All the news dataset",
"authors": [
{
"first": "Andrew",
"middle": [],
"last": "Thompson",
"suffix": ""
}
],
"year": 2017,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Thompson, Andrew. 2017. All the news dataset. https://www.kaggle.com/ snapcrack/all-the-news. Kaggle.",
"links": null
},
"BIBREF38": {
"ref_id": "b38",
"title": "Evaluating adversarial attacks against multiple fact verification systems",
"authors": [
{
"first": "James",
"middle": [],
"last": "Thorne",
"suffix": ""
},
{
"first": "Andreas",
"middle": [],
"last": "Vlachos",
"suffix": ""
},
{
"first": "Christos",
"middle": [],
"last": "Christodoulopoulos",
"suffix": ""
},
{
"first": "Arpit",
"middle": [],
"last": "Mittal",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
"volume": "",
"issue": "",
"pages": "2944--2953",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Thorne, James, Andreas Vlachos, Christos Christodoulopoulos, and Arpit Mittal. 2019. Evaluating adversarial attacks against multiple fact verification systems. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2944-2953, Hong Kong.",
"links": null
},
"BIBREF39": {
"ref_id": "b39",
"title": "The fact extraction and verification (fever) shared task",
"authors": [
{
"first": "James",
"middle": [],
"last": "Thorne",
"suffix": ""
},
{
"first": "Andreas",
"middle": [],
"last": "Vlachos",
"suffix": ""
},
{
"first": "Oana",
"middle": [],
"last": "Cocarascu",
"suffix": ""
}
],
"year": 2018,
"venue": "Proceedings of the First Workshop on Fact Extraction and VERification (FEVER)",
"volume": "",
"issue": "",
"pages": "1--9",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Thorne, James, Andreas Vlachos, Oana Cocarascu, Christos Christodoulopoulos, and Arpit Mittal. 2018. The fact extraction and verification (fever) shared task. In Proceedings of the First Workshop on Fact Extraction and VERification (FEVER), pages 1-9, Brussels.",
"links": null
},
"BIBREF40": {
"ref_id": "b40",
"title": "NewsQA: A machine comprehension dataset",
"authors": [
{
"first": "Adam",
"middle": [],
"last": "Trischler",
"suffix": ""
},
{
"first": "Tong",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "Xingdi",
"middle": [],
"last": "Yuan",
"suffix": ""
},
{
"first": "Justin",
"middle": [],
"last": "Harris",
"suffix": ""
},
{
"first": "Alessandro",
"middle": [],
"last": "Sordoni",
"suffix": ""
},
{
"first": "Philip",
"middle": [],
"last": "Bachman",
"suffix": ""
},
{
"first": "Kaheer",
"middle": [],
"last": "Suleman",
"suffix": ""
}
],
"year": 2017,
"venue": "Proceedings of the 2nd Workshop on Representation Learning for NLP",
"volume": "",
"issue": "",
"pages": "191--200",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Trischler, Adam, Tong Wang, Xingdi Yuan, Justin Harris, Alessandro Sordoni, Philip Bachman, and Kaheer Suleman. 2017. NewsQA: A machine comprehension dataset. In Proceedings of the 2nd Workshop on Representation Learning for NLP, pages 191-200, Vancouver.",
"links": null
},
"BIBREF41": {
"ref_id": "b41",
"title": "Fake news detection in social networks via crowd signals",
"authors": [
{
"first": "Sebastian",
"middle": [],
"last": "Tschiatschek",
"suffix": ""
},
{
"first": "Adish",
"middle": [],
"last": "Singla",
"suffix": ""
},
{
"first": "Manuel",
"middle": [
"Gomez"
],
"last": "Rodriguez",
"suffix": ""
},
{
"first": "Arpit",
"middle": [],
"last": "Merchant",
"suffix": ""
},
{
"first": "Andreas",
"middle": [],
"last": "Krause",
"suffix": ""
}
],
"year": 2018,
"venue": "Companion Proceedings of the The Web Conference",
"volume": "",
"issue": "",
"pages": "517--524",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Tschiatschek, Sebastian, Adish Singla, Manuel Gomez Rodriguez, Arpit Merchant, and Andreas Krause. 2018. Fake news detection in social networks via crowd signals. In Companion Proceedings of the The Web Conference 2018, pages 517-524.",
"links": null
},
"BIBREF42": {
"ref_id": "b42",
"title": "Neural network applications in stylometry: The Federalist Papers. Computers and the Humanities",
"authors": [
{
"first": "Fiona",
"middle": [
"J"
],
"last": "Tweedie",
"suffix": ""
},
{
"first": "Shreyan",
"middle": [],
"last": "Singh",
"suffix": ""
},
{
"first": "David",
"middle": [
"I"
],
"last": "Holmes",
"suffix": ""
}
],
"year": 1996,
"venue": "",
"volume": "30",
"issue": "",
"pages": "1--10",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Tweedie, Fiona J., Shreyan Singh, and David I. Holmes. 1996. Neural network applications in stylometry: The Federalist Papers. Computers and the Humanities, 30:1-10.",
"links": null
},
"BIBREF43": {
"ref_id": "b43",
"title": "Attention is all you need",
"authors": [
{
"first": "Ashish",
"middle": [],
"last": "Vaswani",
"suffix": ""
},
{
"first": "Noam",
"middle": [],
"last": "Shazeer",
"suffix": ""
},
{
"first": "Niki",
"middle": [],
"last": "Parmar",
"suffix": ""
},
{
"first": "Jakob",
"middle": [],
"last": "Uszkoreit",
"suffix": ""
},
{
"first": "Llion",
"middle": [],
"last": "Jones",
"suffix": ""
},
{
"first": "Aidan",
"middle": [
"N"
],
"last": "Gomez",
"suffix": ""
},
{
"first": "\u0141ukasz",
"middle": [],
"last": "Kaiser",
"suffix": ""
},
{
"first": "Illia",
"middle": [],
"last": "Polosukhin",
"suffix": ""
}
],
"year": 2017,
"venue": "Advances in Neural Information Processing Systems",
"volume": "",
"issue": "",
"pages": "5998--6008",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems, pages 5998-6008.",
"links": null
},
"BIBREF44": {
"ref_id": "b44",
"title": "Separating facts from fiction: Linguistic models to classify suspicious and trusted news posts on",
"authors": [
{
"first": "Svitlana",
"middle": [],
"last": "Volkova",
"suffix": ""
},
{
"first": "Kyle",
"middle": [],
"last": "Shaffer",
"suffix": ""
},
{
"first": "Jin",
"middle": [
"Yea"
],
"last": "Jang",
"suffix": ""
},
{
"first": "Nathan",
"middle": [],
"last": "Hodas",
"suffix": ""
}
],
"year": 2017,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Volkova, Svitlana, Kyle Shaffer, Jin Yea Jang, and Nathan Hodas. 2017. Separating facts from fiction: Linguistic models to classify suspicious and trusted news posts on",
"links": null
},
"BIBREF45": {
"ref_id": "b45",
"title": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics",
"authors": [
{
"first": "",
"middle": [],
"last": "Twitter",
"suffix": ""
}
],
"year": null,
"venue": "",
"volume": "2",
"issue": "",
"pages": "647--653",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Twitter. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 647-653,Vancouver.",
"links": null
},
"BIBREF46": {
"ref_id": "b46",
"title": "The spread of true and false news online",
"authors": [
{
"first": "Soroush",
"middle": [],
"last": "Vosoughi",
"suffix": ""
},
{
"first": "Deb",
"middle": [],
"last": "Roy",
"suffix": ""
},
{
"first": "Sinan",
"middle": [],
"last": "Aral",
"suffix": ""
}
],
"year": 2018,
"venue": "Science",
"volume": "359",
"issue": "6380",
"pages": "1146--1151",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Vosoughi, Soroush, Deb Roy, and Sinan Aral. 2018. The spread of true and false news online. Science, 359(6380):1146-1151.",
"links": null
},
"BIBREF47": {
"ref_id": "b47",
"title": "Liar, liar pants on fire\": A new benchmark dataset for fake news detection",
"authors": [
{
"first": "William",
"middle": [],
"last": "Wang",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Yang",
"suffix": ""
}
],
"year": 2017,
"venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics",
"volume": "2",
"issue": "",
"pages": "422--426",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Wang, William Yang. 2017. \"Liar, liar pants on fire\": A new benchmark dataset for fake news detection. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 422-426, Vancouver.",
"links": null
},
"BIBREF48": {
"ref_id": "b48",
"title": "Information disorder: Toward an interdisciplinary framework for research and policy making",
"authors": [
{
"first": "Claire",
"middle": [],
"last": "Wardle",
"suffix": ""
},
{
"first": "Hossein",
"middle": [],
"last": "Derakhshan",
"suffix": ""
}
],
"year": 2017,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Wardle, Claire and Hossein Derakhshan. 2017. Information disorder: Toward an interdisciplinary framework for research and policy making. https://tinyurl.com/sy64l6s.",
"links": null
},
"BIBREF49": {
"ref_id": "b49",
"title": "Huggingface's transformers: State-of-the-art natural language processing",
"authors": [
{
"first": "Thomas",
"middle": [],
"last": "Wolf",
"suffix": ""
},
{
"first": "Lysandre",
"middle": [],
"last": "Debut",
"suffix": ""
},
{
"first": "Victor",
"middle": [],
"last": "Sanh",
"suffix": ""
},
{
"first": "Julien",
"middle": [],
"last": "Chaumond",
"suffix": ""
},
{
"first": "Clement",
"middle": [],
"last": "Delangue",
"suffix": ""
},
{
"first": "Anthony",
"middle": [],
"last": "Moi",
"suffix": ""
},
{
"first": "Pierric",
"middle": [],
"last": "Cistac",
"suffix": ""
},
{
"first": "Tim",
"middle": [],
"last": "Rault",
"suffix": ""
},
{
"first": "R'emi",
"middle": [],
"last": "Louf",
"suffix": ""
},
{
"first": "Morgan",
"middle": [],
"last": "Funtowicz",
"suffix": ""
},
{
"first": "Jamie",
"middle": [],
"last": "Brew",
"suffix": ""
}
],
"year": 2019,
"venue": "ArXiv",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Wolf, Thomas, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, R'emi Louf, Morgan Funtowicz, and Jamie Brew. 2019. Huggingface's transformers: State-of-the-art natural language processing. ArXiv, abs/1910.03771.",
"links": null
},
"BIBREF50": {
"ref_id": "b50",
"title": "Defending against neural fake news",
"authors": [
{
"first": "Rowan",
"middle": [],
"last": "Zellers",
"suffix": ""
},
{
"first": "Ari",
"middle": [],
"last": "Holtzman",
"suffix": ""
},
{
"first": "Hannah",
"middle": [],
"last": "Rashkin",
"suffix": ""
},
{
"first": "Yonatan",
"middle": [],
"last": "Bisk",
"suffix": ""
},
{
"first": "Ali",
"middle": [],
"last": "Farhadi",
"suffix": ""
},
{
"first": "Franziska",
"middle": [],
"last": "Roesner",
"suffix": ""
},
{
"first": "Yejin",
"middle": [],
"last": "Choi",
"suffix": ""
}
],
"year": 2019,
"venue": "Advances in Neural Information Processing Systems",
"volume": "32",
"issue": "",
"pages": "9054--9065",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Zellers, Rowan, Ari Holtzman, Hannah Rashkin, Yonatan Bisk, Ali Farhadi, Franziska Roesner, and Yejin Choi. 2019. Defending against neural fake news. In Advances in Neural Information Processing Systems 32, pages 9054-9065, Curran Associates, Inc.",
"links": null
},
"BIBREF51": {
"ref_id": "b51",
"title": "Enquiring minds: Early detection of rumors in social media from enquiry posts",
"authors": [
{
"first": "Zhe",
"middle": [],
"last": "Zhao",
"suffix": ""
},
{
"first": "Paul",
"middle": [],
"last": "Resnick",
"suffix": ""
},
{
"first": "Qiaozhu",
"middle": [],
"last": "Mei",
"suffix": ""
}
],
"year": 2015,
"venue": "Proceedings of the 24th International Conference on World Wide Web",
"volume": "",
"issue": "",
"pages": "1395--1405",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Zhao, Zhe, Paul Resnick, and Qiaozhu Mei. 2015. Enquiring minds: Early detection of rumors in social media from enquiry posts. In Proceedings of the 24th International Conference on World Wide Web, pages 1395-1405, Florence.",
"links": null
},
"BIBREF52": {
"ref_id": "b52",
"title": "Fake news: A survey of research, detection methods, and opportunities",
"authors": [
{
"first": "Xinyi",
"middle": [],
"last": "Zhou",
"suffix": ""
},
{
"first": "Reza",
"middle": [],
"last": "Zafarani",
"suffix": ""
}
],
"year": 2018,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1812.00315"
]
},
"num": null,
"urls": [],
"raw_text": "Zhou, Xinyi and Reza Zafarani. 2018. Fake news: A survey of research, detection methods, and opportunities. arXiv preprint arXiv:1812.00315.",
"links": null
}
},
"ref_entries": {}
}
}