Next Article in Journal
CRI-Based Smart Lighting System That Provides Characteristics of Natural Light
Next Article in Special Issue
Bibliometric Analysis of IoT Lightweight Cryptography
Previous Article in Journal
Collaboration System for Multidisciplinary Research with Essential Data Analysis Toolkit Built-In
Previous Article in Special Issue
Predicting Abnormal Respiratory Patterns in Older Adults Using Supervised Machine Learning on Internet of Medical Things Respiratory Frequency Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The PolitiFact-Oslo Corpus: A New Dataset for Fake News Analysis and Detection

1
Department of Literature, Area Studies and European Languages, University of Oslo, 0315 Oslo, Norway
2
Sintef Digital, 0373 Oslo, Norway
*
Author to whom correspondence should be addressed.
Information 2023, 14(12), 627; https://doi.org/10.3390/info14120627
Submission received: 14 September 2023 / Revised: 16 November 2023 / Accepted: 16 November 2023 / Published: 23 November 2023

Abstract

:
This study presents a new dataset for fake news analysis and detection, namely, the PolitiFact-Oslo Corpus. The corpus contains samples of both fake and real news in English, collected from the fact-checking website PolitiFact.com. It grew out of a need for a more controlled and effective dataset for fake news analysis and detection model development based on recent events. Three features make it uniquely placed for this: (i) the texts have been individually labelled for veracity by experts, (ii) they are complete texts that strictly correspond to the claims in question, and (iii) they are accompanied by important metadata such as text type (e.g., social media, news and blog). In relation to this, we present a pipeline for collecting quality data from major fact-checking websites, a procedure which can be replicated in future corpus building efforts. An exploratory analysis based on sentiment and part-of-speech information reveals interesting differences between fake and real news as well as between text types, thus highlighting the importance of adding contextual information to fake news corpora. Since the main application of the PolitiFact-Oslo Corpus is in automatic fake news detection, we critically examine the applicability of the corpus and another PolitiFact dataset built based on less strict criteria for various deep learning-based efficient approaches, such as Bidirectional Long Short-Term Memory (Bi-LSTM), LSTM fine-tuned transformers such as Bidirectional Encoder Representations from Transformers (BERT) and RoBERTa, and XLNet.

1. Introduction

Fake news is a societal issue that does not seem to be slowing down. If anything, the growing levels of distrust in mainstream media and government agencies [1] have added fuel to the fire of fake news and allowed for false and misleading information to spread through a variety of digital channels. From COVID-19 conspiracy theories to the bogeyman of “Nazism” in Ukraine, fact-checking websites such as PolitiFact in the US and others around the world have their work cut out for them. Automatic fake news detection can support these efforts, as shown by the large amount of research on this topic in natural language processing and machine learning, e.g., [2,3,4,5,6,7,8,9]. However, the results obtained from these studies are only as good as the data on which they are trained. Who determines what is fake and what is real? How much data are enough? How can we control for confounding variables? These are only some of the questions that developers of fake news datasets need to consider for valid and reliable outcomes.
In this study, we present a new dataset of fake and real news in English with a focus on the US, namely, the PolitiFact-Oslo Corpus. The corpus is built based on a pipeline for collecting quality data from major fact-checking websites, which we describe in detail to facilitate future corpus building efforts in this area. We also examine the linguistic characteristics of the corpus and its suitability for automatic fake news detection using a series of case studies in natural language processing and machine learning. The latter are applied both to the PolitiFact-Oslo Corpus and the DeClarE dataset, based on data from the same fact checker (see below), in order to compare their characteristics. Moreover, we run separate analyses on the full datasets and the main text types of the PolitiFact-Oslo Corpus (social media, news and blog) to demonstrate the importance of contextual information in fake news corpora.
The PolitiFact-Oslo Corpus grew out of a need for a more controlled and effective dataset for fake news analysis and detection model development based on recent events. So far, fake news researchers have relied on three main types of data: (i) simulated data, (ii) natural data that have been classified for veracity based on source reputation, and (iii) natural data that have been individually labelled for veracity by experts. The first data source has a clear disadvantage: the training data have been artificially generated and therefore have little external validity. Natural data based on source reputation allow for large-scale data collection procedures, but since there is no one-to-one correspondence between truth content and source reputation, then such data are not very reliable. Therefore, the best approach to collecting data for automatic fake news detection is by using natural data that have been individually labelled for veracity by experts, such as journalists working for reputable fact-checking organizations.
Fact-checking websites focusing on news in the US have served as a basis for several existing fake news datasets. Early examples are PolitiFact and Channel 4 in [10] and Emergent in [11], with 221 statements and 300 rumors, respectively. In fact, PolitiFact.com has been a favorite among fake news researchers, possibly due to its good reputation and a fine-grained system of ratings: False, Mostly False and Pants on Fire to True, Mostly True and Half True. The LIAR dataset, for example, includes as many as 12,800 statements from PolitiFact [12], while MisInfoText [13] provides access to a large, but unverified, dataset based on the fact checker. PolitiFact was also used in [14]’s study of fake news around the 2016 US presidential election, which, according to the authors, is when fake news went mainstream. The dataset that we use to compare with the PolitiFact-Oslo Corpus in this study, DeClarE [15], contains a subset from PolitiFact collected before 2017.
The ready availability of fake news datasets based on PolitiFact thus begs the question: why the PolitiFact-Oslo Corpus? We argue that our corpus, a structured collection of natural texts in machine-readable form, addresses several limitations of the previous datasets. This has important implications, not only for our understanding of the best practices of language data collection and management, but also for being able to identify real differences between fake and real news. Automatic extraction methods have meant that many of the datasets contain news items that do not in fact correspond to the claim fact-checked by PolitiFact. Often, these are multimodal resources such as photos and videos (e.g., “An image shows gas prices on 6 January 2021”). In these cases, what is extracted is not the content of the image but the text that accompanies it, which may be irrelevant or incomplete. At other times, what is extracted is not the news item corresponding to the claim but the claim itself, which is very short and may not be enough for robust text classification. Finally, there is a problem with the lack of access to metadata about the news items. An important piece of information that is often missing and that has negative consequences for language models is text type. Since most fake news in PolitiFact seems to be from social media, and most real news seems to be from news websites, then a model trained on them is more likely to identify differences between the text types of social media and news websites than fake and real news (cf. [16]). This limitation of the previous studies calls into question the validity and effectiveness of the fake news detection models developed so far. The DeClarE dataset, for instance, provides access to the news items rather than the claims, but as far as we know, no steps were taken to ensure a direct and valid relationship between the news items and the claims or to determine their text type. As will be shown, the PolitiFact-Oslo Corpus meets all the requirements above, making it highly suited for fake news detection.
The case studies are a mix of approaches from natural language processing and machine learning, drawing on insights from, e.g., [17,18,19,20,21,22,23]. In the former, we carry out exploratory analyses of the sentiment and part-of-speech features of the corpus and, in the latter, we examine the applicability of the corpus for a deep learning-based efficient approach for fake news detection by means of various state-of-the-art approaches. These are: Deep Belief Networks (DBN), Long Short-Term Memory (LSTM), Bidirectional LSTM, Bidirectional Encoder Representations from Transformers (BERT), RoBERTa, DistilBERT, and XLNet. The approaches in this study are applied both separately and in combination to achieve better machine learning performance.

2. The PolitiFact-Oslo Corpus

As the name implies, the PolitiFact-Oslo Corpus relies on the fact-checking website PolitiFact.com for its data. The data collection procedure was a collaborative effort between the University of Oslo and Sintef Digital in Oslo, Norway. In what follows, we describe the procedure in detail and then present an overview of how the key features of the corpus make it uniquely placed for practical applications in fake news analysis and detection model development.

2.1. Data Collection, Design and Metadata

The procedure for collecting data for the PolitiFact-Oslo Corpus was a combination of automatic and manual techniques to have greater control over what is included. The automatic techniques involved extracting the fact-checking articles and their metadata from PolitiFact, including all the external links. To reduce later manual work, at this stage, we also filtered out those fact-checking articles that corresponded to spoken text types such as ‘ad’, ‘press conference’, ‘radio’, ‘TV’, ‘debate’, etc. This is because transcripts created for general popular use come with no assurances that they are detailed enough for robust text analysis.
Then, we manually identified the one link that corresponded to the PolitiFact’s claim to extract its content. We extracted the complete text, its title, the author/poster, and the date. Several important decisions were made at this stage. For example, we decided not to include content that was clearly satirical (e.g., The Onion), because satire is disseminated for very different reasons than fake news in its strict sense. Thus, adding satire would have negatively affected our results, in line with the functionalist idea that differences in communicative intent are reflected in linguistic structure [24]. In the case of images, the text was extracted only if it sufficiently described the content of the image. One example of an image that was excluded was of Dwayne “The Rock” Johnson, whose photo of him sitting on a beach was entitled “Epstein Island” by a Facebook user. The claim that PolitiFact then sought to verify was: “A photo shows Dwayne “The Rock” Johnson on Epstein Island” (of course, it was rated Pants on Fire). However, since the text itself did not contain any critical information about the content of the photo (not even who was in it), then it would have added little value to text classification models trained on the data. Instead, we decided to go with content that was substantial enough for robust text analysis. Finally, we came across several news items that had been corrected after being fact-checked by PolitiFact. In these cases, we accessed the original version of the news item using Wayback Machine, a digital archive, to make sure that the item indeed corresponded to the PolitiFact’s claim. Thus, the manual approach used to build the PolitiFact-Oslo Corpus was crucial for ensuring an effective dataset.
To ensure a controlled dataset, we added important metadata information to the corpus. In addition to the above metadata (author/poster, date), the PolitiFact-Oslo Corpus contains information about the text type (e.g., social media, news and blog) and source (e.g., Facebook, The Gateway Pundit) of the news items. This information is more consistent than that provided by PolitiFact, where, for instance, an image may be tagged either as a ‘viral image’ or ‘Instagram post’ (depending on the source). Within text type, we conflated news articles and blog posts because of the fuzzy boundary between the two. This is because most of the news websites in our corpus are run by highly specialized news organizations with limited budget and resources, and therefore their function and design are very similar to blogs run by individuals. As will be shown, taking into consideration where the news item is published has important implications for fake news analysis and detection model development, due to contextual variation between different types of texts.
Figure 1 summarizes the data collection procedure implemented in this study. The pipeline in the figure is meant to be general enough to facilitate future corpus building efforts with data from any of the major fact-checking websites (e.g., PolitiFact, Snopes, Emergent in English). It consists of four main steps and two optional steps, depending on the structure of the fact-checking website and the researchers’ goals. While the pipeline itself is agnostic about the nature of the data collection procedure (automatic or manual), we strongly believe that manual techniques are key to ensuring a more controlled and effective dataset.
The outcome of the data collection procedure for the PolitiFact-Oslo Corpus is given in Table 1. It should be noted that the development of the corpus is a work-in-progress and therefore the figures in the table are preliminary. However, they provide a large enough sample for the natural language processing and machine learning models developed in this study. At present, the corpus contains 428,917 words stored in 2745 texts, covering a time period between January 2019 and December 2022. The average number of words in the texts is 158, which means that the texts are relatively short. Table 1 presents the distribution of the text types in terms of the number of texts, words and average text length across the broad labels of False and True, or Fake and Real, where the former subsumes the PolitiFact ratings of False, Mostly False and Pants on Fire, and the latter subsumes True, Mostly True and Half True. We have excluded the ratings of Full Flop and No Flip due to their low occurrences. As can be seen in the table, over 83% of the texts are from social media, and 14% are from news and blog, making these two the largest text types in the corpus. In addition, the corpus contains 249 different sources, with Facebook being the most common one. The labels are also not uniformly distributed, with over 80% of them being False. This is a limitation of the PolitiFact-Oslo Corpus to which we return in Section 5.

2.2. Key Features and Applications

As previously mentioned, the PolitiFact-Oslo Corpus is characterized by three key features that make it uniquely placed for practical applications in fake news analysis and detection model development.
(i)
The texts have been individually labelled for veracity by expert journalists working for a reputable fact-checking organization.
(ii)
They are complete texts, rather than short claims or statements, which strictly correspond to the claims in question.
(iii)
They are accompanied by important metadata about the news items.
The first feature ensures that the news items, based on real-world data, are reliably classified for automatic detection using a fine-grained system. The second feature of having complete texts is important for building robust models, since “text classification relies mainly on the linguistic characteristics of longer text” [13]. Moreover, the texts of the news items correspond strictly to the claims in question to make sure that there is a direct and valid relationship between the two. The third feature allows us to control for well-known confounding variables in fake news research, such as text type, authorship, time, etc. Text type, in particular, is important for teasing apart the differences in variation in how news is disseminated across a variety of digital channels. In addition, the PolitiFact-Oslo Corpus gives access to new data about recent events such as COVID-19, the 2020 US presidential election and the Russian invasion of Ukraine, when fake news is believed to have reached new dimensions [25].
The corpus, therefore, has clear applications in fake news research and detection model development. Research in style-based fake news classification, for instance, has generated plenty of evidence that there are systematic differences in the language and style of fake and real news; the former is often characterized by more negative and intense emotions, exemplified by intensifiers and negative sentiment words as well as the more frequent use of verbs, adjectives and adverbs [7,8,16,26]. This information could then be used by the very same fact-checking organizations that provided the data for an initial screening of potentially false and misleading information. Due to its focus on text and the language of fake news in the strict sense, the PolitiFact-Oslo Corpus is less suited for the identification of multimodal features such as those found in deepfakes (or even emojis). Although multimodal data may have played a role in the fact-checkers’ decisions, only the text corresponding to the data is represented in a reliable way in the corpus (but see the Fakeddit dataset for an alternative [27]).

3. Fake News Analysis Using Natural Language Processing

In this section, we present a couple of case studies in natural language processing, sentiment and part-of-speech (POS), to explore the specific linguistic characteristics of the PolitiFact-Oslo Corpus across the broad labels of True and False. For this, we used the two largest text types which together constitute over 97% of the dataset, namely, social media, and news and blog (a total of 2674 news items). The idea is to reveal potential similarities and differences between the text types.

3.1. Sentiment

Firstly, we conducted sentiment analysis of the news items using the VADER sentiment analysis tool [28] from the Stanford CoreNLP natural language processing toolkit [29]. We compared the number of news items with positive and negative sentiment. We considered a news item to have positive sentiment if the VADER ‘compound’ score is greater than or equal to 0. A news item was considered to have negative sentiment if the ‘compound’ score is less than 0. We plot the differences in sentiment for the classes of True and False in Figure 2.
The first thing that one notices in Figure 2 is the difference between social media, on the one hand, and news and blog, on the other. Specifically, there is a higher proportion of items with negative sentiment in the news articles and blog posts than in the social media posts, regardless of whether the items are True or False. This is perhaps surprising, considering that news articles and blog posts tend to be distributed to wider audiences to shape public opinion. An important difference between fake and real news, however, is that, in both text types, the percentage of items with positive sentiment is higher in the True class than it is in the False class. The higher percentage of positive sentiment in the former could be because of factual reporting and a generally more positive and favorable depiction of the news events. By contrast, False items often contain deliberately deceptive content, i.e., disinformation, which may lead to “non-conscious leakage of anxiety” [30] about lying and, consequently, to a higher negative sentiment score. Examples (1) and (2), snippets of news articles and blog posts from the PolitiFact-Oslo Corpus, illustrate this difference. While both texts are concerned with COVID-19, they view the topic through different evaluative lenses. The news article from which (1) is extracted was labelled Mostly True by PolitiFact and received a highly positive sentiment score (0.9148).
(1)
It sounds like science fiction. A fabric that smothers coronaviruses in less than a minute, its electrokinetic superpowers choking the life out of COVID-19 like Thanos squeezing the life out of Loki? It’s real, says Chandan Sen, the director of the Indiana Center for Regenerative Medicine and Engineering at the Indiana University School of Medicine in Indianapolis. And, with the number of novel coronavirus infections sitting at 4 million as of 10 May, it holds tantalizing potential for the future of personal protective equipment.
A highly negative sentiment score (−0.9993) was obtained for the blog post in (2), labelled False by PolitiFact.
(2)
The “novel Coronavirus” outbreak affecting China and many other countries right now, has been determined to be a military BIO-WEAPON, which was being worked on at the Wuhan Virology Laboratory by China’s People’s Liberation Army, Nanjiang Command. Somehow, it got out. The world is now facing a massive wipe-out of humanity as a result.
We also investigated the average of the sentiment scores in all the False and True items in the two text types. These values are reported in Table 2. In the table, we can make a similar observation as above: the average sentiment scores of True items are higher than the average sentiment scores of False items. The higher average sentiment scores could be another indication that accurate and credible news content tends to evoke more positive emotions than false and deceptive content.

3.2. Part-of-Speech

Next, we looked at the distribution of POS tags in the corpus. Focusing again on the two largest text types, Figure 3 shows the top five POS tags across the False and True classes. By far the most frequent POS tag in all cases is nouns, followed by verbs. There is some variation in the rest of the tags. Pronouns, for example, only feature frequently in the False class of social media posts. This is in line with the more involved and informal style of social media posts compared to news articles and blog posts, as well as compared to the more informational and formal style of factual reporting as identified in previous research [16]. An example of a False social media post from Facebook is given in (3). Note the frequent use of pronouns, particularly personal pronouns such as she, her, you, me, us, they and I, which together make up a total of 14% of the complete social media post (not shown here due to copyright and privacy reasons).
(3)
Queen Pelosi wasn’t happy with the small USAF C-20B jet, Gulfstream III, that comes with the Speaker’s job … OH NO! Queen Pelosi was aggravated that this little jet had to stop to refuel, so she ordered a Big Fat, 200-seat, USAF C-32, Boeing 757 jet that could get her back to California without stopping […] Queen Pelosi wants you and me to conserve our carbon footprint. She wants us to buy smaller cars and Obama wants us to get a bicycle pump and air up our tires. Who do these people think they are??? Their motto is … Don’t do as I do … JUST DO AS I SAY!
The higher frequency of determiners in the True class is an indication of the presence of complex noun phrases (e.g., the new bill, the state’s stay-at-home order), which contribute further to the more formal style of truthful news items.
Within news and blog, there are more proper nouns in the False class than in the True class. This could be an indication of an attempt to make the false content more credible by adding recognizable names and entities (e.g., a long list of schools that are to be affected by an alleged ban by Joe Biden on school choice vouchers in Wisconsin). By contrast, the True class stands out with regard to its use of adjectives to add detail and richness to the description of the news events. Considering that previously adjectives have been found to characterize fake news rather than real news (see Section 2.2), we need to look closer into their function in the PolitiFact-Oslo Corpus in the future.

4. Machine Learning for Fake News Detection

4.1. Overview

As previously mentioned, the main application of the PolitiFact-Oslo Corpus is in automatic fake news detection. Therefore, the rest of the case studies are concerned with machine learning and a deep learning-based efficient approach for fake news detection by means of various state-of-the-art approaches. Deep learning has contributed a lot recently in many fields such as pattern analysis and artificial intelligence [31,32,33], with important applications in fake news detection model development [34,35,36,37,38]. However, deep learning has two major disadvantages: the first disadvantage is the overfitting problem most of the time and the second one is that it takes a lot of time to model the underlying data.
The pioneer deep learning algorithm was Deep Belief Networks (DBN). It consisted of Restricted Boltzmann Machines (RBMs) that made the training faster than previous learning methods. Recurrent Neural Networks (RNNs) are, however, a better choice than DBN, although they have gradient vanishing point problems [32]. To avoid that, the Long Short-Term Memory (LSTM) deep learning model was introduced. Basically, it consists of some gates over input sequences to remember the history inside the data. To be able to follow both directions to model sequential information via LSTM, Bidirectional LSTM may be used. In addition, some studies have applied pre-trained models for sequence classification such as Bidirectional Encoder Representations from Transformers (BERT), RoBERTa, DistilBERT, and XLNet [33]. Transformer models are cutting-edge natural language processing (NLP) models that have played a pivotal role in advancing the capabilities of language understanding and generation in the field of artificial intelligence. BERT, developed by Google AI, was a groundbreaking model that marked a significant shift in NLP. It introduced the concept of bidirectional context, allowing the model to understand the meaning of a word in a sentence by considering the surrounding words. BERT’s pretraining on vast text corpora, followed by fine-tuning on specific tasks, has set the standard for a wide range of NLP applications. A Robustly Optimized BERT Pretraining Approach (RoBERTa), an improvement upon BERT, was developed by Facebook AI. It refines BERT’s training methodology by employing larger batch sizes and more extensive data, resulting in remarkable performance improvements on various NLP tasks. RoBERTa is celebrated for its robustness and efficiency. DistilBERT, created by Hugging Face, is a distilled version of BERT designed for efficiency. It retains BERT’s performance while being smaller and faster to train and deploy. DistilBERT is particularly suitable for resource-constrained applications, making it a popular choice for real-world usage. Transformer-XL Network (XLNet), developed by Google AI and Carnegie Mellon University, extends the transformer architecture and introduces a novel permutation-based training approach. This method allows XLNet to capture complex dependencies in text by considering all possible word permutations, surpassing the performance of previous models on various NLP benchmarks. In short, BERT, RoBERTa, DistilBERT, and XLNet have significantly advanced the field of NLP, each bringing unique contributions in terms of bidirectional context understanding, robustness, efficiency, and complex dependency modeling, respectively. These models have set the foundation for a new era of language understanding in artificial intelligence.
In this study, we focus on processing text data, features, and fake news prediction as a smart application. Figure 4 shows a schematic setup of a text-based fake news prediction system in a smart application where a user provides a query text, and a server processes the text to apply feature extraction and deep learning for predicting whether the text is fake or not. Figure 5 shows the basic architecture of the proposed fake news prediction system consisting of training and testing procedures. In the training part, text data from all the users are obtained and the features are trained using machine learning models such as Bidirectional LSTM and BERT-based fine-tuned models. In the testing part, features from a sample test are applied to the trained models to make the decision whether the text contains fake news or not.
As previously mentioned, machine learning models capable of handling text data sequences are applied in this study. To achieve high accuracy, we propose to combine Bidirectional LSTM and BERT, because of their capabilities to model underlying events in text information. LSTM has recurrent connections between its hidden units, which connect history to the present state. LSTM was derived from typical RNNs which generally cause vanishing gradient problems during the processing of long-term information, which LSTM can overcome. Bidirectional LSTM should perform even better than typical unidirectional LSTM to model time-series information from both directions. Figure 6 shows a sample Bidirectional LSTM model.
Each LSTM block contains a cell state and three gates, namely, input, forget and the output gate. The input gate It is determined as
I t = β ( W P I P t + W H I H t 1 + b I )
where W is the weight matrix, b represents the bias vectors, and β is the logistic sigmoid function. The forget gate P can be expressed as
P t = β ( W P F I t + W H F H t 1 + b F ) .
The long-term memory is stored in a cell state vector S expressed as
S t = P t S t 1 + I t tanh ( W P S V t + W H S H t 1 + b S ) .
The output gate N determines what is going to be an output expressed as
N t = β ( W P O P t + W H O H t 1 + b O ) .
The hidden state H is expressed as
H t = O t tanh ( S t ) .
Finally, the output O can be determined as
O = softmax ( W U H l + b U )
where l indicates the last LSTM number. Figure 7 shows the fine-tuned transformer model using LSTM and dense layers applied in this study, where the softmax function is applied to the final layer weights to make the decision. Figure 7 shows the transfer learning using the LSTM-based fine-tuned transformer-based (BERT, RoBERTa, DistlilBERT, XLNet) basic architecture model applied in his study. We have compared the fine-tuned transformer-based model with three other typical algorithms: DBN, LSTM, and Bidirectional LSTM.

4.2. Experimental Results and Discussion

The experiments were applied both to the PolitiFact-Oslo Corpus and the DeClarE dataset in order to compare their characteristics. The classes were both fine-grained (i.e., True, Half True, Mostly True, False, Mostly False, Pants on Fire) and broad (i.e., Real or True combining True, Half True and Mostly True as well as Fake or False combining False, Mostly False and Pants on Fire); see Table 1 for the distribution of the classes. Moreover, in the PolitiFact-Oslo Corpus, we carried out separate experiments on the full dataset as well as the two largest text types: news and blog, and social media. Five-fold cross-validation was applied to evaluate the machine learning models. Twenty percent of the data were used for testing each fold and ten percent of the data extracted from the training data were used for validation.

4.2.1. Full Dataset

We started out with the full dataset of the PolitiFact-Oslo Corpus. Figure 8 shows the word clouds of the True and False classes based on the most common words in the corpus. As can be seen in the figure, the word clouds are different in terms of lexical and topical features. False news, in particular, is clearly focused on topics that, over the last few years, have been highly vulnerable to manipulation and misinformation (e.g., COVID-19, vaccines).
Table 3 and Table 4 show the results (in average recall rates (%) and their means) of traditional embedding features and various machine learning approaches to the corpus based on two and six classes, respectively. As expected, the fined-tuned transformer models (e.g., XLNet) show overall better performance. Moreover, two classes provide higher accuracy rates than six classes since the number of samples for training decreases when the number of classes increases. Figure 9 shows the loss vs. epoch plots of the Bidirectional LSTM and BERT models, which show quick convergence to lower loss when the number of epochs increases. Although transformer models are more time-consuming than Bidirectional LSTM, the former also shows convergence considering the number of epochs. It is clear, however, that in the case of both two and six classes, the models have been biased towards the False class. This is due to the considerably lower number of texts in the True class compared to the False class (see Table 1).
Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 show the results of the different machine learning model architectures with layers and parameters as well as example results from a fold to show the characteristics of the models on the corpus. The models are: Bidirectional LSTM, fine-tuned BERT (bert-base-uncased), RoBERTa (roberta-base), DistilBERT (distilbert-base-uncased), and XLNet (XLNet-base-cased). The results from the folds include the precision, recall, f1-score, support, and accuracy of the fold. For all the machine learning models, we used 0.001 as the learning rate.
Figure 15 shows the confusion matrix of a fold during cross-fold validation experiments of two classes with the BERT-based model on the corpus. It can be observed in Figure 15 that during two-class experiments, texts from the True class were highly confused with texts from the False class due to misclassification. By contrast, most of the texts in the False class were well-classified with less confusion in the True class. Figure 16 shows the bidirectional LSTM architecture and results in detail using that on the PolitiFact-Oslo Corpus for a fold of six classes, while Figure 17 shows the LSTM-based fine-tuned BERT architecture and results using that on the same fold. We followed similar layers for RoBERTa, DistilBERT, and XLNet. Figure 18 shows the results of the same fold using RoBERTa, DistilBERT, and XLNet, respectively. Figure 19 represents the confusion matrix of a fold from cross-fold validation experiments of six classes with the BERT-based model. It can be observed in Figure 19 that when the True and False classes were further divided into sub-classes, the performance of the machine learning algorithm decreased, with increased confusion among the sub-classes. The overall accuracies from the figures confirm that the transformer-based models have better accuracies (e.g., 43% using XLNet) and recall rates (e.g., 23.65% using RoBERTa) compared to others (e.g., 26% accuracy and 20.98% recall rates using Bidirectional LSTM). Thus, all the experimental results reported in this sub-section point to the fact that the recommended machine learning models (i.e., LSTM-based fine-tuned transformers) of this study generate better performance than any of the other traditional approaches (DBN, typical LSTM, Bidirectional LSTM).

4.2.2. Full DeClarE Dataset

Next, we turned to the DeClarE dataset. DeClarE contains 15,018 True and 14,537 False texts, so the dataset is larger and more balanced than the PolitiFact-Oslo Corpus in its current form. The fine-grained labels are also distributed more evenly: True (4393), Half True (5466), Mostly True (5159), False (6052), Mostly False (5024), and Pants on Fire (3461). Moreover, DeClarE does not provide access to the complete texts but rather to the snippets of the complete texts around the claim in question. The snippets are clearly longer than the claims themselves and therefore are more suited for text classification based on linguistic characteristics. Access to the snippets only means that the texts do not contain any irrelevant information in the rest of the news item that might negatively affect the experimental results; this is a concern when complete texts are used.
We applied the same cross-fold validation and machine learning approaches to DeClarE as to the PolitiFact-Oslo Corpus. Figure 20 shows the word clouds of the True and False classes in the DeClarE dataset. As can be seen in the figure, the lexical and topical features in DeClarE are very different from those in the PolitiFact-Oslo Corpus due to the different time period covered by the former dataset (pre-2017). (Barack) Obama figures in both types of news, while the fake news sample also includes (Hilary) Clinton and health as major topics. Presumably, the former is due to the various disinformation campaigns launched against Clinton during the 2016 US presidential election [14].
Table 5 and Table 6 show the results (in average recall rates (%) and their means) of traditional embedding features and various machine learning approaches to the dataset based on two and six classes, respectively. The same architectures as in Section 4.2.1 were applied. In the case of both classes, the fined-tuned transformer models show better performance than the others. Similar to the PolitiFact-Oslo Corpus, two classes provide higher accuracy rates than six classes. Figure 21 and Figure 22 show the results of a single fold applying different machine learning models. Each figure includes the precision, recall, f1-score, support, and accuracy of the fold. Figure 23 shows the loss vs. epoch plots of the Bi-LSTM and BERT models. The plots indicate that the models take more epochs to converge towards lower error rates compared to the models applied to the PolitiFact-Oslo Corpus. The main reason is that the DeClarE dataset contains more texts.
Overall, the accuracy rates from the machine learning models are comparable across the PolitiFact-Oslo Corpus and the DeClarE dataset, with the latter showing slightly better performance. However, we are hesitant to use accuracy rates as a point of comparison between the datasets, due to differences in the sizes of the datasets as well as the heavy class imbalances in the PolitiFact-Oslo Corpus (but see Section 5 for a possible solution). Therefore, we now turn to a feature of our corpus that is not present in DeClarE and that provides a more appropriate point of comparison, namely, access to information about where the news items come from, i.e., their text type.

4.2.3. News and Blog

As previously mentioned, the two largest text types in the PolitiFact-Oslo Corpus are news and blog, and social media. The two text types differ from each other in important ways pertaining mainly to conventional features such as specialized expressions, rhetorical organization and formatting, as well as aspects related to variation in the use of linguistic features [16]. To gauge these differences, we have carried out separate analyses based on the text types. Figure 24 and Figure 25 report the results of a fold applying different machine learning models. Each figure includes the precision, recall, f1-score, support, and accuracy of the fold. Starting with news and blog, Table 7 and Table 8 show the results of traditional embedding features and various machine learning approaches to the text type based on two and six classes, respectively. Again, it is the fine-tuned transformer models that show better performance, most clearly when applied to two classes compared to six classes. In both cases, however, the accuracy rates are slightly lower than for the full dataset.

4.2.4. Social Media

When applied to social media, we can observe some improvement in terms of the performance of the experiments. Figure 26 and Figure 27 report the results of a fold applying different machine learning models. Each figure includes the precision, recall, f1-score, support, and accuracy of the fold. Table 9 and Table 10 show the results of traditional embedding features and various machine learning approaches to the text type based on two and six classes, respectively. Here, almost all the models show better performance than for the full dataset or for news and blog separately. In fact, the combined model shows the highest accuracy obtained for the PolitiFact-Oslo Corpus so far. This might be explained by the greater homogeneity of the social media posts, most of which are from a handful of platforms (e.g., Facebook, Twitter, and Instagram). News articles and blog posts, by contrast, are more diverse in terms of publishers and therefore their lexical and topical features may be more difficult to capture.
In summary, the fact that we obtain different results for the full dataset vs. the different text types is an indication that the language of fake news is not the same across all situations and contexts of use. Also, the lack of access to information about text type in other fake news datasets such as DeClarE means that what is being captured by the machine learning algorithms may not be the difference between fake and real news, but rather a difference between the text types that dominate each of the news samples. Therefore, for the best result possible, text type variation needs to be built into future machine learning models to ensure that we compare like with like: real social media posts with fake social media posts, and so on. With this study, we have taken an important step forward in this regard. In addition to featuring important metadata information, the PolitiFact-Oslo Corpus shows improvement in terms of several other aspects of corpus design, in line with the pipeline developed in this study (see Figure 1). For example, thanks to manual work, the corpus does not include any texts that do not strictly correspond to the claim in question, whereas the automatic procedures used in DeClarE may have let such texts seep in. Finally, a lot has changed in the digital news media landscape since many of the PolitiFact datasets were collected, as suggested by the lexical and topical differences between DeClarE and the PolitiFact-Oslo Corpus in Section 4.2.1 and Section 4.2.2. Malign actors, both human and artificial, have become increasingly good at simulating the speech and style of real news, potentially narrowing the window of opportunity for finding any differences at all. The more recent the data, the more likely we are to keep up with developments in this area.

5. Conclusions

This study has presented the PolitiFact-Oslo Corpus, a new dataset of fake and real news in English based on recent events, and critically examined its suitability for fake news analysis and detection model development by means of a series of case studies in natural language processing and machine learning. The case studies in machine learning were applied both to the PolitiFact-Oslo Corpus and to an existing PolitiFact dataset, namely, DeClarE. While DeClarE shows several advantages over our own corpus, it has nevertheless not been built based on strict criteria. These criteria were presented in this study in the form of a pipeline for collecting quality data from major fact-checking websites such as PolitiFact, which may be used in future corpus building efforts. Besides being individually labelled for veracity by experts, in the PolitiFact-Oslo Corpus the complete texts correspond strictly to the claims in question and are accompanied by important metadata information, thus ensuring a more controlled and effective dataset. Access to information about text type is particularly crucial for valid and reliable outcomes since this kind of variation is pervasive in natural language.
In the case studies, we identified interesting differences between fake and real news in terms of sentiment and POS information. Fake news was characterized by more negative sentiment and a greater use of pronouns, which contribute to a more emotional and informal style of this type of news. However, these features varied across the text types of social media, and news and blog, thus highlighting the importance of adding contextual information to fake news corpora. Access to information about text type also turned out to be an important feature of the PolitiFact-Oslo Corpus in the machine learning models applied in this study; the accuracy rates were different in the case of the full corpus and when the text types were considered separately. The lack of ready access to information about where the news items came from in other fake news datasets, including DeClarE, is a potential limitation that may negatively affect the automatic fake news detection models trained on the data. The PolitiFact-Oslo Corpus meets this requirement. As for the machine learning models that we applied, the LSTM fined-tuned transformer models generally showed better performance over non-transformer-based models by achieving higher accuracy rates. However, at this point, the use of the models only pays off in relation to the broad distinction between fake and real news, rather than to finer-grained distinctions. Therefore, a more powerful combination of various ensemble models based on larger sample sizes will have to be explored in the future.
This study has brought to light several limitations and directions for the future development of the PolitiFact-Oslo Corpus. Firstly, it has drawn attention to the relatively uneven distribution of fake and real news in the corpus due to the preference for PolitiFact and other fact-checkers to debunk false information rather than to find support for true information. One solution that we are currently implementing is to extend the scope of fact-checkers (as found in Google’s Fact Check Explorer) to find more instances of True news without having to stretch out the timeline. Secondly, the case studies are just a taster of what can be performed with the PolitiFact-Oslo Corpus. With a focus on text and the language of fake news in the strict sense, the corpus provides an excellent resource for the analysis of specific linguistic features of fake news. Early models based on grammatical features and relations from linguistic theory are showing promising results. Finally, due to ethical and legal restrictions, the corpus texts are currently available upon request only. Inspired by the DeClarE dataset, however, we are exploring opportunities to release the text snippets via an online interface.

Author Contributions

All authors worked on the concept of the work. More specifically N.P. and A.T. worked on the dataset. Z.U. worked on machine learning experiments and reporting. Finally, all authors revised the work. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by The Research Council of Norway under the project-ID 302573.

Data Availability Statement

Dataset used in this work is available based on the request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Newman, N.; Fletcher, R.; Schulz, A.; Andi, S.; Robertson, C.T.; Nielsen, R.K. Reuters Institute Digital News Report 2021; Reuters Institute: Oxford, UK, 2021. [Google Scholar]
  2. Capuano, N.; Fenza, G.; Loia, V.; Nota, F.D. Content-based fake news detection with machine and deep learning: A systematic review. Neurocomputing 2023, 530, 91–103. [Google Scholar] [CrossRef]
  3. Conroy, N.K.; Rubin, V.L.; Chen, Y. Automatic deception detection: Methods for finding fake news. In Proceedings of the 78th ASIS&T Annual Meeting, St. Louis, MO, USA, 6–10 November 2015; Association for Information Science and Technology: St. Louis, MO, USA, 2015; pp. 1–4. [Google Scholar]
  4. Ibrishimova, M.D.; Li, K. A machine learning approach to fake news detection using knowledge verification and natural language processing. In Advances in Intelligent Networking and Collaborative Systems, INCoS 2019; Barolli, L., Nishino, H., Miwa, H., Eds.; Springer: Cham, Switzerland, 2020; Volume 1035, pp. 223–234. [Google Scholar]
  5. Oshikawa, R.; Qian, J.; Wang, W.Y. A survey of natural language processing for fake news detection. In Proceedings of the 12th Language Resources and Evaluation Conference (LREC 2020), Marseille, France, 11–16 May 2018; European Language Resources Association: Marseille, France, 2018; pp. 6086–6093. [Google Scholar]
  6. Villela, H.F.; Correa, F.; Ribeiro, J.S.A.N.; Rabelo, A.; Carvalho, D.B.F. Fake news detection: A systematic literature review of machine learning algorithms and datasets. J. Interact. Syst. 2023, 14, 47–58. [Google Scholar] [CrossRef]
  7. Rashkin, H.; Choi, E.; Jang, J.Y.; Volkova, S.; Choi, Y. Truth of varying shades: Analyzing language in fake news and political fact-checking. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; Association for Computational Linguistics: Copenhagen, Denmark, 2017; pp. 2931–2937. [Google Scholar]
  8. Volkova, S.; Shaffer, K.; Jang, J.Y.; Hodas, N. Separating facts from fiction: Linguistic models to classify suspicious and trusted news posts on Twitter. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; Association for Computational Linguistics: Vancouver, BC, Canada, 2017; pp. 647–653. [Google Scholar]
  9. Siino, M.; Di Nuovo, E.; Tinnirello, I.; La Cascia, M. Fake News Spreaders Detection: Sometimes Attention Is Not All You Need. Information 2022, 13, 426. [Google Scholar] [CrossRef]
  10. Vlachos, A.; Riedel, S. Fact checking: Task definition and dataset construction. In Proceedings of the ACL 2014 Workshop on Language Technologies and Computational Social Science, Baltimore, MD, USA, 26 June 2014; Association for Computational Linguistics: Baltimore, MD, USA, 2014; pp. 18–22. [Google Scholar]
  11. Ferreira, W.; Vlachos, A. Emergent: A novel data-set for stance classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; Association for Computational Linguistics: San Diego, CA, USA, 2016; pp. 1163–1168. [Google Scholar]
  12. Wang, W.Y. “Liar, Liar Pants on Fire”: A new benchmark dataset for fake news detection. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; Association for Computational Linguistics: Vancouver, BC, Canada, 2017; pp. 422–426. [Google Scholar]
  13. Asr, F.T.; Taboada, M. Big Data and quality data for fake news and misinformation detection. Big Data Soc. 2019, 6, 3310. [Google Scholar] [CrossRef]
  14. Allcott, H.; Gentzkow, M. Social media and fake news in the 2016 election. J. Econ. Perspect. 2017, 31, 211–236. [Google Scholar] [CrossRef]
  15. Popat, K.; Mukherjee, S.; Yates, A.; Weikum, G. DeClarE: Debunking fake news and false claims using evidence-aware deep learning. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; Association for Computational Linguistics: Brussels, Belgium, 2018; pp. 22–32. [Google Scholar]
  16. Grieve, J.; Woodfield, H. The Language of Fake News; Cambridge University Press: Cambridge, UK, 2023. [Google Scholar]
  17. Subba, B.; Kumari, S. A heterogeneous stacking ensemble based sentiment analysis framework using multiple word embeddings. Comput. Intell. 2021, 38, 530–559. [Google Scholar] [CrossRef]
  18. Rodriguez, P.L.; Spirling, A. Word Embeddings: What Works, What Doesn’t, and How to Tell the Difference for Applied Research. J. Polit. 2022, 84, 101–115. [Google Scholar] [CrossRef]
  19. Mangione, S.; Siino, M.; Garbo, G. Improving Irony and Stereotype Spreaders Detection using Data Augmentation and Convolutional Neural Network. In CEUR Workshop Proceedings; Università degli Studi di Palermo, Dipartimento di Ingegneria: Palermo, Italy, 2022; Volume 3180, pp. 2585–2593. [Google Scholar]
  20. Saleh, H.; Alhothali, A.; Moria, K. Detection of Hate Speech using BERT and Hate Speech Word Embedding with Deep Model. Appl. Artif. Intell. 2023, 37, 2166719. [Google Scholar] [CrossRef]
  21. Daniele, C.; Garlisi, D.; Siino, M. An SVM Ensamble Approach to Detect Irony and Stereotype Spreaders on Twitter. In CEUR Workshop Proceedings; Sun SITE Central Europe: Aachen, Germany, 2022; Volume 3180. [Google Scholar]
  22. Incitti, F.; Urli, F.; Snidaro, L. Beyond word embeddings: A survey. Inf. Fusion 2023, 89, 418–436. [Google Scholar] [CrossRef]
  23. Espinosa, D.; Sidorov, G. Using BERT to profiling cryptocurrency influencers. In Working Notes of CLEF; Sun SITE Central Europe: Aachen, Germany, 2023. [Google Scholar]
  24. Biber, D. Variation across Speech and Writing; Cambridge University Press: Cambridge, UK, 1988. [Google Scholar]
  25. Association for Progressive Communications: Disinformation and Freedom of Expression. 2021. Available online: https://www.apc.org/sites/default/files/APCSubmissionDisinformationFebruary2021.pdf (accessed on 2 November 2023).
  26. Sousa-Silva, R. Fighting the fake: A forensic linguistic analysis to fake news detection. Int. J. Semiot. Law 2022, 35, 2409–2433. [Google Scholar] [CrossRef] [PubMed]
  27. Nakamura, K.; Levy, S.; Wang, W.Y. r/Fakeddit: A new multimodal benchmark dataset for fine-grained fake news detection. In Proceedings of the 12th Language Resources and Evaluation Conference (LREC 2020), Marseille, France, 11–16 May 2018; European Language Resources Association: Marseille, France, 2020; pp. 6149–6157. [Google Scholar]
  28. Hutto, C.; Gilbert, E. Vader: A parsimonious rule-based model for sentiment analysis of social media text. In Proceedings of the International AAAI Conference on Web and Social Media, Ann Arbor, MI, USA, 1–4 June 2014; Volume 8. [Google Scholar]
  29. Manning, C.D.; Surdeanu, M.; Bauer, J.; Finkel, J.R.; Bethard, S.; McClosky, D. The Stanford CoreNLP natural language processing toolkit. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Baltimore, MD, USA, 23–24 June 2014; pp. 55–60. [Google Scholar]
  30. Markowitz, D.M.; Hancock, J.T. Linguistic traces of a scientific fraud: The case of Diederik Stapel. PLoS ONE 2014, 9, e105937. [Google Scholar] [CrossRef] [PubMed]
  31. Uddin, Z. Applied Machine Learning for Assisted Living; Springer Science and Business Media LLC: Dordrecht, The Netherlands, 2022. [Google Scholar] [CrossRef]
  32. Uddin, Z.; Hassan, M.M.; Alsanad, A.; Savaglio, C. A body sensor data fusion and deep recurrent neural network-based behavior recognition approach for robust healthcare. Inf. Fusion 2020, 55, 105–115. [Google Scholar] [CrossRef]
  33. Patwardhan, N.; Marrone, S.; Sansone, C. Transformers in the Real World: A Survey on NLP Applications. Information 2023, 14, 242. [Google Scholar] [CrossRef]
  34. Masciari, E.; Moscato, V.; Picariello, A.; Sperli, G. A deep learning approach to fake news detection. In Proceedings of the Foundations of Intelligent Systems: 25th International Symposium, ISMIS 2020, Graz, Austria, 23–25 September 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  35. Konkobo, P.M.; Zhang, R.; Huang, S.; Minoungou, T.T.; Ouedraogo, J.A.; Li, L. A deep learning model for early detection of fake news on social media. In Proceedings of the 2020 7th International Conference on Behavioural and Social Computing (BESC), Bournemouth, UK, 5–7 November 2020. [Google Scholar]
  36. Alghamdi, J.; Lin, Y.; Luo, S. A Comparative Study of Machine Learning and Deep Learning Techniques for Fake News Detection. Information 2022, 13, 576. [Google Scholar] [CrossRef]
  37. Palani, B.; Elango, S.; Viswanathan, K.V. CB-Fake: A multimodal deep learning framework for automatic fake news detection using capsule neural network and BERT. Multimed. Tools Appl. 2022, 81, 5587–5620. [Google Scholar] [CrossRef] [PubMed]
  38. Ali, A.M.; Ghaleb, F.A.; Al-Rimy, B.A.S.; Alsolami, F.J.; Khan, A.I. Deep Ensemble Fake News Detection Model Using Sequential Deep Learning Technique. Sensors 2022, 22, 6970. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A pipeline for collecting quality data from PolitiFact (or any other fact-checking website).
Figure 1. A pipeline for collecting quality data from PolitiFact (or any other fact-checking website).
Information 14 00627 g001
Figure 2. Distribution of news items with positive and negative sentiment in True and False classes across the text types of social media (upper plot), and news and blog (lower plot).
Figure 2. Distribution of news items with positive and negative sentiment in True and False classes across the text types of social media (upper plot), and news and blog (lower plot).
Information 14 00627 g002
Figure 3. Distribution of POS tags in False and True classes across the text types of social media (upper plots), and news and blog (lower plots).
Figure 3. Distribution of POS tags in False and True classes across the text types of social media (upper plots), and news and blog (lower plots).
Information 14 00627 g003
Figure 4. A schematic setup of a text-based fake news prediction system.
Figure 4. A schematic setup of a text-based fake news prediction system.
Information 14 00627 g004
Figure 5. Flow chart of a text-based fake news prediction system.
Figure 5. Flow chart of a text-based fake news prediction system.
Information 14 00627 g005
Figure 6. Bidirectional LSTM model.
Figure 6. Bidirectional LSTM model.
Information 14 00627 g006
Figure 7. Transfer learning model fine-tuned LSTM and dense layers.
Figure 7. Transfer learning model fine-tuned LSTM and dense layers.
Information 14 00627 g007
Figure 8. Word clouds of (a) True and (b) False classes in the full dataset of the PolitiFact-Oslo Corpus.
Figure 8. Word clouds of (a) True and (b) False classes in the full dataset of the PolitiFact-Oslo Corpus.
Information 14 00627 g008
Figure 9. (a) Bidirectional LSTM’s and (b) BERT model’s loss vs. epoch plots based on the PolitiFact-Oslo Corpus.
Figure 9. (a) Bidirectional LSTM’s and (b) BERT model’s loss vs. epoch plots based on the PolitiFact-Oslo Corpus.
Information 14 00627 g009
Figure 10. (a) Bidirectional LSTM architecture (b) results using that on the PolitiFact-Oslo Corpus for a fold.
Figure 10. (a) Bidirectional LSTM architecture (b) results using that on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g010
Figure 11. (a) LSTM fine-tuned BERT architecture (b) results using BERT on the PolitiFact-Oslo Corpus for a fold.
Figure 11. (a) LSTM fine-tuned BERT architecture (b) results using BERT on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g011
Figure 12. (a) LSTM fine-tuned RoBERTa architecture (b) results using RoBERTa on the PolitiFact-Oslo Corpus for a fold.
Figure 12. (a) LSTM fine-tuned RoBERTa architecture (b) results using RoBERTa on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g012
Figure 13. (a) LSTM fine-tuned distilBERT architecture (b) results using distilBERT on the PolitiFact-Oslo Corpus for a fold.
Figure 13. (a) LSTM fine-tuned distilBERT architecture (b) results using distilBERT on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g013
Figure 14. (a) LSTM fine-tuned XLNet architecture (b) results using XLNet on the PolitiFact-Oslo Corpus for a fold.
Figure 14. (a) LSTM fine-tuned XLNet architecture (b) results using XLNet on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g014
Figure 15. Confusion matrix of test results (BERT) of two classes from a fold.
Figure 15. Confusion matrix of test results (BERT) of two classes from a fold.
Information 14 00627 g015
Figure 16. (a) Bidirectional LSTM architecture (b) results using Bidirectional-LSTM on the PolitiFact-Oslo Corpus for a fold.
Figure 16. (a) Bidirectional LSTM architecture (b) results using Bidirectional-LSTM on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g016
Figure 17. (a) LSTM fine-tuned BERT (b) results using BERT on the PolitiFact-Oslo Corpus for a fold.
Figure 17. (a) LSTM fine-tuned BERT (b) results using BERT on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g017
Figure 18. Results using LSTM fine-tuned (a) RoBERTa, (b) DistilBERT, and (c) XLNet on the PolitiFact-Oslo Corpus for a fold.
Figure 18. Results using LSTM fine-tuned (a) RoBERTa, (b) DistilBERT, and (c) XLNet on the PolitiFact-Oslo Corpus for a fold.
Information 14 00627 g018
Figure 19. Confusion matrix of test results (BERT) of six classes from a fold.
Figure 19. Confusion matrix of test results (BERT) of six classes from a fold.
Information 14 00627 g019
Figure 20. Word clouds for (a) True and (b) False classes in the DeClarE dataset.
Figure 20. Word clouds for (a) True and (b) False classes in the DeClarE dataset.
Information 14 00627 g020aInformation 14 00627 g020b
Figure 21. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the DeClarE dataset for a fold for two classes.
Figure 21. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the DeClarE dataset for a fold for two classes.
Information 14 00627 g021
Figure 22. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the DeClarE dataset for a fold for six classes.
Figure 22. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the DeClarE dataset for a fold for six classes.
Information 14 00627 g022
Figure 23. (a) Bidirectional LSTM’s and (b) BERT model’s loss vs. epoch plot based on the DeClarE dataset.
Figure 23. (a) Bidirectional LSTM’s and (b) BERT model’s loss vs. epoch plot based on the DeClarE dataset.
Information 14 00627 g023aInformation 14 00627 g023b
Figure 24. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (news and blog) for a fold for two classes.
Figure 24. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (news and blog) for a fold for two classes.
Information 14 00627 g024
Figure 25. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (news and blog) for a fold for six classes.
Figure 25. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (news and blog) for a fold for six classes.
Information 14 00627 g025
Figure 26. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (social media) for a fold for two classes.
Figure 26. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (social media) for a fold for two classes.
Information 14 00627 g026
Figure 27. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (social media) for a fold for six classes.
Figure 27. Results using (a) Bidirectional LSTM, (b) LSTM fine-tuned BERT, (c) RoBERTa, (d) DistilBERT, and (e) XLNet on the PolitiFact-Oslo Corpus (social media) for a fold for six classes.
Information 14 00627 g027
Table 1. Distribution of text types in the PolitiFact-Oslo Corpus (as of November 2023).
Table 1. Distribution of text types in the PolitiFact-Oslo Corpus (as of November 2023).
Text TypeTextsWordsAverage Text Length *
FalseTrueFalseTrueFalseTrue
Social media1878403119,00220,4986350
News and blog35142215,56333,156614789
Press release272516,105 14,699 778588
Personal82951 1348 119674
Academic504352 0 8690
Political312543 700 848700
Sub-total2272473358,516 70,401160148
Corpus2745428,917158
* Average text length is given in words; all the decimal values have been rounded up to the nearest integer.
Table 2. Average sentiment scores of False and True items across the text types of social media, and news and blog.
Table 2. Average sentiment scores of False and True items across the text types of social media, and news and blog.
Text TypeAverage Sentiment ScoreNews Items
FalseTrueFalseTrue
Social media−0.102−0.031878403
News and blog−0.228−0.11535142
Sub-total−0.122−0.0382229445
Total−0.162674
Table 3. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the full dataset of the PolitiFact-Oslo Corpus based on two classes.
Table 3. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the full dataset of the PolitiFact-Oslo Corpus based on two classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True19.2126.3333.1533.2439.2328.1428.61
False70.3290.5292.6995.1492.7092.6398.59
Mean44.7658.4262.4561.8862.6460.3863.60
Table 4. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the full dataset of the PolitiFact-Oslo Corpus based on six classes.
Table 4. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the full dataset of the PolitiFact-Oslo Corpus based on six classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True2.783.213.284.584.4715.133.12
Half True10.5412.3113.7215.2325.7216.5910.14
Mostly True3.214.736.245.778.335.0212.37
False33.4747.9350.7865.2965.1556.1972.66
Mostly False19.4822.3424.3518.7111.0915.0715.13
Pants on Fire27.3428.5727.5328.7627.1123.1522.57
Mean16.1419.8520.9823.0523.6521.8622.66
Table 5. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the DeClarE dataset based on two classes.
Table 5. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the DeClarE dataset based on two classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True51.2972.1776.5475.2373.8976.9274.15
False14.2760.1870.1373.1976.1873.9070.57
Mean32.7866.1773.3574.2175.0675.4172.36
Table 6. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the DeClarE dataset based on six classes.
Table 6. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the DeClarE dataset based on six classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True23.6140.18154.1457.1247.9355.1851.02
Half True30.2350.0252.7154.2450.0957.9754.83
Mostly True24.1245.7151.7445.9155.1847.1856.18
False50.1851.1755.1855.0861.1859.9157.15
Mostly False40.2345.3055.4854.9454.1148.1449.61
Pants on Fire33.1248.5852.1760.0357.1852.1353.02
Mean33.5848.8253.5754.5554.2753.4153.64
Table 7. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (news and blog) based on two classes.
Table 7. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (news and blog) based on two classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True19.0322.1913.1318.5825.5114.1325.61
False60.1584.1895.1394.1795.6295.6186.57
Mean39.6053.1954.1456.3760.5654.8756.09
Table 8. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (news and blog) based on six classes.
Table 8. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (news and blog) based on six classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True0.020.761.210.872.151.690.98
Half True0.742.372.912.2621.252.131.56
Mostly True0.251.183.585.473.872.1534.15
False12.7111.3510.158.1374.7683.9746.12
Mostly False8.776.0209.2210.0910.723.619.16
Pants on Fire25.6327.2873.9773.9117.2116.1559.69
Mean8.028.1616.8416.7821.6618.2825.27
Table 9. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (social media) based on two classes.
Table 9. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (social media) based on two classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True18.2325.1626.2733.8141.7136.4740.15
False77.1490.0691.3793.9690.3793.1985.91
Mean47.6957.6158.8263.8866.0464.8363.03
Table 10. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (social media) based on six classes.
Table 10. Experimental results (recall rates in %) of traditional embedding features and various machine learning approaches to the PolitiFact-Oslo Corpus (social media) based on six classes.
DBNLSTMBidirectional LSTMBERTRoBERTaDistilBERTXLNet
True10.9211.4715.4423.185.144.6124.47
Half True12.1715.7917.2914.618.613.1719.28
Mostly True2.193.2310.017.1315.173.834.15
False38.1441.1649.0172.1565.3786.1972.32
Mostly False18.1721.2725.0910.3111.1521.5215.81
Pants on Fire17.2818.8420.1427.5112.5710.1516.27
Mean16.4718.6322.8325.8119.6621.5825.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Põldvere, N.; Uddin, Z.; Thomas, A. The PolitiFact-Oslo Corpus: A New Dataset for Fake News Analysis and Detection. Information 2023, 14, 627. https://doi.org/10.3390/info14120627

AMA Style

Põldvere N, Uddin Z, Thomas A. The PolitiFact-Oslo Corpus: A New Dataset for Fake News Analysis and Detection. Information. 2023; 14(12):627. https://doi.org/10.3390/info14120627

Chicago/Turabian Style

Põldvere, Nele, Zia Uddin, and Aleena Thomas. 2023. "The PolitiFact-Oslo Corpus: A New Dataset for Fake News Analysis and Detection" Information 14, no. 12: 627. https://doi.org/10.3390/info14120627

APA Style

Põldvere, N., Uddin, Z., & Thomas, A. (2023). The PolitiFact-Oslo Corpus: A New Dataset for Fake News Analysis and Detection. Information, 14(12), 627. https://doi.org/10.3390/info14120627

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop