Next Article in Journal
Dynamic Game Analysis on Cooperative Advertising Strategy in a Manufacturer-Led Supply Chain with Risk Aversion
Next Article in Special Issue
A Concise Relation Extraction Method Based on the Fusion of Sequential and Structural Features Using ERNIE
Previous Article in Journal
A Bimodal Extension of the Epsilon-Skew-Normal Model
Previous Article in Special Issue
Contextual Urdu Lemmatization Using Recurrent Neural Network Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

It’s All in the Embedding! Fake News Detection Using Document Embeddings

by
Ciprian-Octavian Truică
*,† and
Elena-Simona Apostol
*,†
Computer Science and Engineering Department, Faculty of Automatic Control and Computers, University Politehnica of Bucharest, RO-060042 Bucharest, Romania
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2023, 11(3), 508; https://doi.org/10.3390/math11030508
Submission received: 27 December 2022 / Revised: 12 January 2023 / Accepted: 16 January 2023 / Published: 18 January 2023

Abstract

:
With the current shift in the mass media landscape from journalistic rigor to social media, personalized social media is becoming the new norm. Although the digitalization progress of the media brings many advantages, it also increases the risk of spreading disinformation, misinformation, and malformation through the use of fake news. The emergence of this harmful phenomenon has managed to polarize society and manipulate public opinion on particular topics, e.g., elections, vaccinations, etc. Such information propagated on social media can distort public perceptions and generate social unrest while lacking the rigor of traditional journalism. Natural Language Processing and Machine Learning techniques are essential for developing efficient tools that can detect fake news. Models that use the context of textual data are essential for resolving the fake news detection problem, as they manage to encode linguistic features within the vector representation of words. In this paper, we propose a new approach that uses document embeddings to build multiple models that accurately label news articles as reliable or fake. We also present a benchmark on different architectures that detect fake news using binary or multi-labeled classification. We evaluated the models on five large news corpora using accuracy, precision, and recall. We obtained better results than more complex state-of-the-art Deep Neural Network models. We observe that the most important factor for obtaining high accuracy is the document encoding, not the classification model's complexity.

1. Introduction

With the increase in the digitalization of mass media, new journalistic paradigms for information distribution have emerged. These new paradigms have substantially changed the way society consumes information. By trying to be ahead of the competition, sometimes people who report on world events leave behind the rigors of classical journalism and publish their content as soon as possible in order to “go viral” by obtaining as many views, likes, comments, and shares as possible in a short amount of time [1]. This new paradigm centers on the users, catering to their needs, behavior, and interests. Along with the advantages the digitalization of mass media brings, it also increases the risk of misinformation, with potentially detrimental consequences for society, by facilitating the spread of misinformation [2,3] in the form of fake news (which influenced the Brexit referendum [4], the 2016 US presidential election [5], COVID-19 vaccinations [6], etc.).
Fake news consists of news articles that are intentionally and verifiably false. This type of information aims to mislead readers by presenting alleged, real-seeming facts about social, economic, and political subjects of interest [7]. However, the current technological trends make this type of content harmful, with potentially dire consequences to the community (e.g., public polarization regarding elections). This has become a major challenge for democracy. Information propagated online may lack the rigor of classic journalism, and can, therefore, distort public perceptions, cause false alarms, and generate social unrest. Furthermore, the president of the EU, Ursula von der Leyen, has repeatedly condemned and asked for immediate action to be taken against the spread of fake news that undermines democracy and public health [8]. Thus, the ideological polarization of readers through the spread of fake news is an important issue and requires scholarly attention. We believe that designing and building tools and methods for accurately detecting fake news is of great relevance, and thus, our results will have an overall positive impact.
In this paper, we propose a new approach that uses document embeddings (i.e., DocEmb) for detecting fake news. We also present our benchmark on different architectures that detect fake news using binary or multi-labeled classification. The document embeddings are constructed using several (1) word embeddings trained on each dataset selected for experiments, and (2) pre-trained transformers. We employ TFIDF, word embeddings (i.e., Word2Vec, FastText, and GloVe), and transformers (i.e., BERT, RoBERTa, and BART) to create DocEmb, our new document embedding approach. For classification, we train both classical Machine Learning models (i.e., Naïve Bayes, Gradient Boosted Trees) and Deep Learning models (i.e., Perceptron, Multi-Layer Perceptron, [Bi]LSTM, [Bi]GRU).
In our experiments, we analyze the performance of the DocEmb based detection solution on multiple datasets annotated with either binary or multi-class labels. We use 4 binary datasets, i.e., a sample of 20 000 manually annotated news articles from the Fake News Corpus, Liar, Kaggle, and Buzz Feed News. Finally, we use 2 multi-labeled datasets, i.e., Liar with 6 labels and TSHP-17 with 3 labels. As evaluation metrics, we use accuracy, precision, and recall.
We compare our results with state-of-the-art Deep Neural Networks models. Our method outperforms these models on each dataset. The most important takeaway from our experiments is that we empirically show that:
(1) 
A simpler neural architecture offers better or at least similar results compared to complex architectures that employ multiple layers, and
(2) 
The difference in performance lies in the embeddings used to vectorize the textual data and how well these performs in encoding contextual and linguistic features.
The main contributions of this article are as follows:
( C 1 )
We propose a new document embedding (DocEmb) constructed using word embeddings and transformers. We specifically trained the proposed DocEmb on the five datasets used in the experiments.
( C 2 )
We show empirically that simple Machine Leaning algorithms trained with our proposed DocEmb obtain similar results or better results than deep learning architectures specifically developed for the task of binary and multi-class fake news detection. This contribution is important in the machine learning literature because it changes the focus from the classification architecture to the document encoding architecture.
( C 3 )
We present a new manually filtered dataset. The original dataset is the widely used Fake News Corpus that was annotated with an automatic process.
This paper is structured as follows. Section 2 discusses current research on the topic of fake news detection. Section 3 introduces our approach and presents the different modules and models employed. Section 4 presents the datasets and analyzes the results. In Section 5, we summarize our key findings and discuss the major implications. Section 6 presents the conclusions and outlines directions for future work.

2. Related Work

As views and clicks monetize online media, for some publishers, it is most important to provide news that might interest their audience, to the detriment of the quality of the facts reported [9]. Thus, proper journalistic rigor has come under threat through the online spread of fake news.
Wang [10] employed SVM (Support Vector Machine), LogReg (Logistic Regression), BiLSTM (Bidirectional Long Short-Term Memory), and CNN (Convolutional Neural Network), to detect the veracity of ∼13 K short statements. The preprocessing was done using Google News’ pre-trained Word2Vec embeddings. Conroy et al. [11] present analysis methods based on linguistic and syntactic features for discovering fake news.
Many current approaches employ complex Deep Neural Network architectures, e.g., based on CNN (Convolutional Neural Network) [12,13,14], BiLSTM (Bidirectional Long Short-Term Memory) [15], and others. Ilie et al. [16] used multiple deep neural networks to determine how models that use pre-trained and specific trained word embeddings perform in the task of fake news detection. Further, some solutions use advanced document embeddings based on encoder architectures [17]. Kaliyar et al. [18] propose FakeBERT, an extension of FNDNet that uses BERT instead of GloVe embeddings. Kula et al. [19] used a hybrid architecture for fake news detection that connects BERT with recurrent networks while Mersinias et al. [20] introduced CLDF, a new vectorization technique for extracting feature vectors. The results for CLDF, FNDNet, and FakeBERT were obtained using the Kaggle dataset with ∼21 K news articles.
Different ensemble models have also been used for this task, with good results [21,22]. Mondal et al. [21] used a voting-based ensemble method that relies on the voting of the collective majority. The authors employ only non-deep learning models and TF-IDF as the vectorization technique. Aslam et al. [22] used an ensemble-based deep learning model that combines two architectures, i.e., Bi-LSTM-GRU-Dense and Dense. Truică and Apostol [23] propose MisRoBÆRTa, a BERT- and RoBERTa-based ensemble model for fake news detection.
Sedik et al. [24] propose a deep learning approach that uses both sequential and recurrent layers. The sequential models employ stacked CNNs (i.e., CNN model) or concatenated CNN (i.e., C-CNN model), while the recurrent models use stacked CNN with LSTM and Dense layers (i.e., CNN-LSTM model) or simple GRU with a Dense layer (i.e., GRU model). The experimental results using the binary labeled Kaggle and Fake News Challenge dataset show that C-CNN and CNN-LSTM have the best performance, i.e., C-CNN obtains an accuracy of 99.90% on the Kaggle dataset, and CNN-LSTM obtains an accuracy of 96% on the Fake News Challenge dataset.
Several current solutions are based on linguistic and syntactic features, e.g., WELFake [25], which uses word embedding over linguistic features. In other current directions, multimodal learning that integrates comments [26], images [27], and the social and network context has been used [26,28,29]. Wang et al. [30] propose a knowledge-driven Multimodal Graph Convolutional Network model for detecting fake news from textual and visual information. This solution models posts from social media as graph data structures that combine textual and visual data with knowledge concepts.
Le and Mikolov [31] propose Doc2Vec as an extension to Word2Vec. Doc2Vec computes a feature vector for every document in the dataset, as opposed to Word2Vec, which computes every word in the dataset. Several articles have discussed the use of Doc2Vec for fake news detection, but it is used only as a baseline combined with traditional Machine Learning solutions. Cui et al. [32] use as a baseline Doc2Vec with SVM and compare it with graph-based Deep Learning solutions. Singh [33] presents several experiments on LIAR and Kaggle datasets using different vector space representations, i.e., one-hot encoding, TFIDF, Word2Vec, and Doc2Vec. Truică et al. [7] propose a BiLSTM architecture with Sentence Transformer for the fake news detection challenge at CheckThatLab! 2022. The proposed architecture uses BART for a monolingual fake news detection task and XML-RoBERTa for the multilingual task. For the multilingual task, the model relies on transfer learning. Thus, the BiLSTM XML-RoBERTa model is trained on English and tested on a German dataset. The proposed model managed to obtain an accuracy of 0.53 for the first task and an accuracy of 0.28 for the second task.

3. Methodology

Figure 1 presents the pipeline of our proposed solution. A labeled corpus of news articles is first preprocessed to extract the tokens. Then, the tokens are transformed into a vector model using term weighting schemes (TFIDF) and word/transformer embeddings. These vectors are used to create document embeddings. We also use the raw corpus to create document embeddings using transformers. The vectorized documents are then passed to the classification module. Finally, the classification is evaluated using accuracy, precision, and recall.

3.1. Text Preprocessing

To prepare the text for vectorization, we use the following preprocessing steps to minimize the vocabulary and remove terms that bring no information gain [34]: (1) removal of punctuation and stopwords, and (2) word lemmatization. We chose to lemmatize the words to minimize the vocabulary and remove any language inflections. We do not apply these preprocessing steps when using the transformer embeddings.

3.2. Term Weighting

To vectorize the preprocessed documents, we employ the TFIDF (Equation (1)). To compute this metric, we first need to compute the (1) term-frequency TF (Equation (2)) and (2) the inverse document frequency (Equation (3)). For a set of n documents D = { d i i 1 , n ¯ } , we extract the set of m unique terms V = { t j j 1 , m ¯ } . This set of unique terms is called a vocabulary. For each term, we compute the raw frequency ( f t j , d i ), which counts the number of occurrences of a term t j in a document d i . The f t j , d i does not store context and is biased towards longer documents. Thus, to remove the bias, we normalize the frequency with the length of the document ( t d i f t , d i ) and obtain TF [35]. Furthermore, to minimize the importance of common terms that bring no information value, the IDF (Equation (3)) is used to reduce the TF weight by a factor that grows with the collection frequency n j of a term t j , i.e., n j is the number of documents where there is at least one occurrence of term t j . Finally, to normalize TFIDF in the [ 0 , 1 ] range, we use the 2 -norm (Equation (4)).
T F I D F ( t j , d i , D ) = T F ( t j , d i ) · I D F ( t j , D ) 2 ( d i )
T F ( t j , d i ) = f t j , d i t d i f t , d i
I D F ( t j , D ) = log n n j
2 ( d i ) = t d i T F ( t , d i ) · I D F ( t , D ) 2
Using TFIDF, we construct a n × m document-term matrix X = { x i j i = 1 , n ¯ j = 1 , m ¯ } ( X R n × m ) where rows correspond to documents and columns to terms. The value x i j = T F I D F ( t j , d i , D ) represents the weight of term t j in document d i . Thus, each document d i is represented by a vector x i = { x i j j 1 , m ¯ } . For ease of notation, we use x i for denoting lines in X.

3.3. Word Embeddings

Each word from the vocabulary is transformed into its vector representation. This module employs Word2Vec [36,37], FastText [38], and GloVe [39]. For Word2Vec and FastText, we use both the CBOW (Continuous Bag-of-Words) and SG (Skip-gram) models. By using these models, we obtain the embedding W o r d E m b ( t ) for each term t V .

3.3.1. Word2Vec

The Word2Vec [36,37] embedding model is used to create vectorized representations of the words in a dataset within the same vector space. This representation measures the distance between the corresponding vectors in this space to determine the context similarity. For Word2Vec, there are two models for representing the words in this vector space: Continuous Bag-Of-Words (CBOW) or Skip-Gram.

CBOW Model

The CBOW model attempts to predict a word using the context given by its surrounding words. Each word t i V ( i 1 , m ¯ ) is defined by two d-dimensional (with d 2 a natural number, i.e., d N ) vectors depending of its function in training: (1) v t i R d is defined when t i is used as the center word, and (2) u t i R d is defined when t i is used as a context word. The conditional probability of generating any center word t c given its surrounding context words T o = { t 1 , , t c 1 , t c + 1 , , t s } within a context window of size s can be modeled by a probability distribution p ( t c T o ) (Equation (5)) that considers the average of the context vectors v ¯ o = 1 s ( v t 1 + + v t c 1 + v t c + 1 + + v t s ) .
p ( t c T c ) = e u c v ¯ o i = 1 m e u i v ¯ o

Skip-Gram Model

The Skip-Gram model starts with the context word t c as input and tries to generate its context. As in the CBOW case, the two d-dimensional vectors ( d N and d 2 ), i.e., v t i R d and u t i R d , are defined for each word t i V ( i 1 , m ¯ ). The conditional probability of generating any context word t o given the center word t c can be modeled by a softmax operation (Equation (6)).
p ( t o t c ) = e u o v c i = 1 m e u i v c

3.3.2. FastText

FastText [38] is an extension to Word2Vec and follows a similar approach to construct word embeddings [40]. The main difference between FastText and Word2Vec is that FastText does not consider the word as the basic unit, but rather considers a bag of character n-grams. Using such an approach, the accuracy is improved, and the training time is decreased when compared to Word2Vec. As in the case of Word2Vec, FastText employs both CBOW and Skip-Gram models.

3.3.3. GloVe

GloVe (Global Vectors) [39] is another model used for creating word embeddings. To create the vector representation of words, GloVe uses the word co-occurrences matrix. This matrix manages to encapsulate local and global corpus statistics regarding word–word co-occurrences. Thus, GloVe for each word stores the frequency of its appearance in the same context as another word by employing a term co-occurrence matrix. Using the ratio of co-occurrence probability, GloVe captures the relationship between words. Furthermore, GloVe identifies word analogies and synonyms within the same contexts using this probability ratio.

3.4. Transformers Embeddings

To create transformer embeddings, we use BERT [41], RoBERTa [42], and BART [43]. By using these models, we obtain the word embedding by transformer W o r d E m b ( t ) for each term t V .

3.4.1. BERT

BERT (bidirectional encoder representations from transformers) [41] is a deep bidirectional transformer architecture used for natural language understanding. Thus, in contrast to classic language models that treat textual data as unidirectional or bidirectional sequences of words, BERT learns contextual relations between the words by employing this deep bidirectional transformer architecture. Using the surrounding words of a given word, the model learns and creates a vector representation for each word that also encapsulates its context. Thus, BERT reads the entire sequence of words at once using the transformer encoder to create contextual word embeddings. By employing transfer learning, BERT can directly be used for various natural language processing operations, understanding, and generation. Furthermore, it can be fine-tuned by using new datasets to adapt to specific tasks. Experimental results on various tasks [41] show that the language models built with BERT manage to improve language context detection more than the models that use static word embeddings, which only see textual data as sequences of words.

3.4.2. RoBERTa

RoBERTa (a Robustly optimized BERT pre-training Approach) [42] is a training optimizing method for BERT. This model improves the language masking strategy of BERT by modifying the following key training aspects: (1) more data are used for training, (2) dynamic masking patterns instead of static masking patterns are employed, (3) the next-sentence pre-training objective is removed and replaced with full sentences without NSP (Next Sentence Prediction), (4) training is performed on longer sequences, (5) the mini-batches are improved, and (6) the learning rates are improved. Thus, all these modifications lead to improving RoBERTa’s downstream task performance and mitigate some of the shortcomings encountered by the significantly undertrained BERT model.

3.4.3. BART

BART (bidirectional and autoregressive transformer) [43] is a transformer model that employs the standard transformer-based neural machine translation architecture, i.e., a generalized BERT architecture. The pre-training process of BART uses an arbitrary noising function to corrupt the textual data within the dataset in order to make the transformer learn how to recreate the original text during training. During the pre-training of BART, two key techniques are used to improve the words’ contextual representations. Firstly, the order of original sentences is randomly shuffled. Secondly, using a novel in-filling scheme, a single mask token is used to replace the spans of text. Experimental results [43] show that a fine-tuned BART works better than BERT for both text generation and comprehension tasks.

3.5. Document Embeddings

We create a vector for each document by averaging all the word or transformer embeddings for the words appearing in the document. Thus, if we have m i terms in a document d i , we obtain the document embedding (DocEmb) x i by summing all the embeddings w ( t ) of the terms t that are present in document d i as well as in the vocabulary V, and dividing the sum by m i (Equation (7)). Each document embedding creates a context for words in a document and becomes an extension of the presented word embeddings.
x i = t d i w ( t ) m i
Similarly to TFIDF, we construct a document-embedding matrix X = { x i i = 1 , n ¯ } where each row corresponds to the document embedding x i . For this matrix, the columns are not associated to terms in the vocabulary V, and the number of columns is different from the total number of terms in V. In this case, m is the size of the embedding vector. For ease of notation, we use m as the number of columns, although it is different than the number of terms in the vocabulary, as in the case of the document-term matrix. Thus, X R n × m is a n × m matrix.

3.6. Fake News Detection

Classification is used to determine the veracity of a news article, i.e, fake news detection. Given a set of documents D represented by the matrix X R n × m (either the document-terms or the document embedding matrix), a set of classes Y = { y 1 , , y n } with values in a discrete domain C = { c k k = 1 , κ ¯ } ( Y C ) of size κ (i.e., the number of classes is κ ), and an implication x i y i ( i 1 , n ) ¯ , the objective of classification is to predict y ^ i = f ( x i ) ( y ^ i Y ^ C ) that best approximates y i . For the fake news detection task, we employ the following algorithms to construct models: Naïve Bayes (NB), Gradient Boosted Trees (XGBTrees), Perceptron, Multi-Layer Perceptron (MLP), Long Short-Term Memory Network (LSTM), Bidirectional LSTM (BiLSTM), Gated Recurrent Units (GRU), and Bidirectional GRU (BiGRU). For comparison, we use MisRoBÆRTa [23]. In the original article that presents MisRoBÆRTa, the authors fine-tune both BART and RoBERTa. In this work, we use the pre-trained BART (facebook/bart-large) and RoBERTa (roberta-base) from HuggingFace [44].

3.6.1. Naïve Bayes

The Naïve Bayes (NB) model is a probabilistic classification algorithm that computes the probability of x i given a class c k (Equation (8)), where p ( x i ) and p ( c k ) are the probability of a document and a class, respectively, and p ( x i c k ) is the probability of class c k given x i . Expending x i by its components { x i 1 , , x i m } , we can rewrite Equation (8) as Equation (9).
p ( c k x i ) = p ( c k ) p ( x i c k ) p ( x i )
p ( c k x i 1 , , x i m ) = p ( c k ) p ( x i 1 , , x i m c k ) p ( x i )
The denominator p ( x i ) is constant, while p ( x i 1 , , x i m c k ) is equivalent to the joint probability p ( c k , x i 1 , , x i m ) . Furthermore, all the terms are conditionally independent given a class c k . Thus, p ( c k , x i 1 , , x i m ) = j = 1 m p ( x i j c k ) . Using these assumptions, the Naïve Bayes classifier tries to estimate the class y ^ i using Equation (10).
y ^ i = argmax c k C p ( c k ) j = 1 m p ( x i j c k )
There are various types of Naïve Bayes classifiers; the most common ones are Multinomial Naïve Bayes and Gaussian Naïve Bayes.

Multinomial Naïve Bayes

Multinomial Naïve Bayes (MNB) models the distribution of words in a document by using a multinomial representation for the distribution of probabilities that a word appears for a certain class (Equation (11)). The assumption for this model is that a document is handled as a sequence of words. Also, it is assumed that each word position is generated independently of every other [45].
p ( x i c k ) = ( j = 1 m x i j ) ! j = 1 m x i j ! j = 1 m p ( x i j c k ) x i j
Equation (12) presents the Multinomial Naïve Bayes classification model.
y ^ i = argmax c k C p ( c k ) ( j = 1 m x i j ) ! j = 1 m x i j ! j = 1 m p ( x i j c k ) x i j

Gaussian Naïve Bayes

The Gaussian Naïve Bayes (GNB) model is used when dealing with continuous data. The model is based on the assumption that continuous values correlated with each class are distributed according to a Gaussian distribution. Thus, given column j 1 , m ¯ from X and a class c k , we employ the following steps:
  • Segment the data by class c k .
  • Compute the associated means μ j and variances σ j of dimension j using the values x i j ( i 1 , n ¯ ), only for the lines x i X labeled with class c k .
  • Compute the probability p ( x i j c k ) (Equation (13)).
p ( x i j c k ) = 1 2 π σ j 2 e x i j μ j 2 2 σ j 2
Equation (14) presents the Gaussian Naïve Bayes classifier.
y ^ i = argmax c k C p ( c k ) j = 1 m 1 2 π σ j 2 e x i j μ j 2 2 σ j 2

3.6.2. Gradient Boosted Trees

Gradient boosting is an ensemble method that uses multiple weak predictions learners. In the case of Gradient Boosted Trees, the weak learners are Decision Trees. Similar to other classification methods, the method tries to predict y ^ i = f ( x i ) = w · x i + b that best approximates the true class y i of x i by minimizing an objective function L ( Y ^ i , Y i ) that represents the training loss, e.g., the mean score L ( y i , y ^ i ) = 1 n i = 1 n ( y ^ i y i ) 2 = 1 n i = 1 n l ( y ^ i , y i ) . As the model builds T weak learners f ( t ) ( x i ) ( t 1 , T ¯ ), at each stage t T the model tries to determine y ^ i ( t ) = y ^ i ( t 1 ) + f ( t ) ( x i ) using the previously estimated value y ^ i ( t 1 ) and the function f ( t ) ( x i ) determined by the current weak learner that best fits the residuals, i.e., f ( t ) ( x i ) = y i y ^ i ( t ) . As the objective is to minimize training loss and to obtain the specific objective at step t,we can take the Taylor expansion of the loss function up to the second order for each learner and remove all the constants to obtain L ( t ) ( Y ^ i ( t ) , Y i ) (Equation (15)).
L ( t ) ( Y ^ i ( t ) , Y i ) = i = 1 n [ g i f ( t ) ( x i ) + 1 2 h i ( f ( t ) ( x i ) ) 2 ] g i = y ^ i ( t 1 ) l ( y i , y ^ i ( t 1 ) ) h i = y ^ i ( t 1 ) 2 l ( y i , y ^ i ( t 1 ) )

3.6.3. Perceptron

The Perceptron model (Equation (16)) is a simple non-linear processing unit that tries to predict the label y ^ i for a given input x i by adjusting a weight vector w R m using the sigmoid activation δ s ( z ) = 1 1 + e z [ 0 , 1 ] . The objective for a good prediction is to minimize the average cross-entropy loss function between the set of prediction Y ^ and the set of true labels Y (Equation (17)).
y ^ i = δ s ( w · x i + b )
L ( Y ^ , Y ) = 1 n i = 1 n ( y i log y ^ i + ( 1 y i ) log ( 1 y ^ i ) )

3.6.4. Multi-Layer Perceptron

The Multi-Layer Perceptron (MLP) model is a Deep Learning architecture that stacks multiple layers j 1 , l ¯ of fully-connected Perceptron units. The MLP architecture can be divided into three components: (1) the input i layer ( j = 1 ), (2) the hidden layers j 2 , l 1 ¯ , and (3) the output layer o = y ^ ( j = l ). Each node in layer j connects to every node in the following layer j + 1 with a certain weight W j . Because the connections between the layers are directed from the input i to the output o by passing information through the hidden layers h j , the MLP model is a feed-forward architecture. Equation (18) presents the MLP classification model at a given iteration t.
i ( t ) = δ s ( W 1 · x i ( t ) + b 1 ) h j ( t ) = δ s ( W j · h j 1 ( t ) + b j ) o ( t ) = δ s ( W l · h l 1 ( t ) + b l )

3.6.5. Long Short-Term Memory

Long Short-Term Memory (LSTM) [46] is a Recurrent Artificial Neural Network that uses two state components for classification. The first component, represented by a hidden state, is the short-term memory that learns the short-term dependencies between the previous and current states. The second component, represented by an internal cell state, is the long-term memory which stores long-term dependencies between the previous and current states. The model uses three gates to preserve the long-term memory within the state: (1) input gate ( i R n ), (2) forget gate ( f R n ), and (3) output gate ( o R n ). Equation (19) presents the compact forms for the state updates of the LSTM unit for a given iteration t, where:
  • x i ( t ) R m is the input vector of dimension m at step t, with x i ( 0 ) = x i X ;
  • h ( t ) R n is the hidden state vector as well as the unit’s output vector of dimension n, where the initial value is h ( 0 ) = 0 ;
  • c ˜ ( t ) R n is the input activation vector;
  • c ( t ) R n is the cell state vector, with the initial value c ( 0 ) = 0 ;
  • W i , W f , W o , W c R n × m are the weight matrices corresponding to the current input of the input gate, output gate, forget gate, and the cell state;
  • V i , V f , V o , V c R n × n are the weight matrices corresponding to the hidden output of the previous state for the current input of the input gate, output gate, forget gate, and the cell state;
  • b i , b f , b o , b c R n are the bias vectors corresponding to the current input of the input gate, output gate, forget gate, and the cell state;
  • δ h ( z ) = e z e z e z + e z [ 1 , 1 ] is the hyperbolic tangent activation function;
  • ⊙ is the Hadamard Product, i.e., element wise product.
i ( t ) = δ s ( W i x i ( t ) + V i h ( t 1 ) + b i ) f ( t ) = δ s ( W f x i ( t ) + V f h ( t 1 ) + b f ) o ( t ) = δ s ( W o x i ( t ) + V o h ( t 1 ) + b o ) c ˜ ( t ) = δ h ( W c x i ( t ) + V c h ( t 1 ) + b c ) c ( t ) = i ( t ) c ˜ ( t ) + f ( t ) c ( t 1 ) h ( t ) = o ( t ) δ h ( c ( t ) )
We chose LSTM because it manages to avoid the vanishing and the exploding gradient issues by regulating the way the recurrent weights are learned.

3.6.6. Bidirectional LSTM

As the LSTM model processes sequence data, it is able to capture past information. To take into consideration future information as well, we use the Bidirectional LSTM (BiLSTM). The BiLSTM encapsulates past and future information through the use of two hidden states (Equation (20)), where (1) h ( t ) processes the input in a forward manner using the past information provided by the forward LSTM ( L S T M F ), and (2) h ( t ) processes the input in a backward manner using the future information provided by the backward LSTM ( L S T M B ).
h ( t ) = L S T M F ( x i ( t ) ) h ( t ) = L S T M B ( x i ( t ) )
At every time-step, the hidden states, i.e., h ( t ) and h ( t ) , are concatenated into one hidden state h ( t ) (Equation (21)). This approach enables the encoding of information from both past and future contexts in the hidden state h ( t ) .
h ( t ) = [ h ( t ) h ( t ) ]

3.6.7. Gated Recurrent Unit

The Gated Recurrent Unit (GRU) [47] is a Recurrent Artificial Neural Network that simplifies the LSTM unit and improves performance considerably. Instead of three gates as in the case of LSTM, the GRU has only two gating mechanisms. The first gating mechanism is the update gate ( u R n ). This gate encodes both the forget gate and the input gate that are present in the LSTM cell. The second gating mechanism is the reset gate ( r R n ). This gate determines the percentage of information from the previous hidden state that contributes to the candidate state of the new step [48] Furthermore, the GRU uses the hidden state as the only state component. Equation (22) presents the compact forms for the state updates of the GRU unit at a given iteration step t, where:
  • x i ( t ) R m is the input vector of dimension m at step t, with x i ( 0 ) = x i X ;
  • i ( t ) R n is the input and output of the cell at step t;
  • h ˜ ( t ) R n is the candidate hidden state with a cell dimension of n;
  • h ( t ) R n is the current hidden state with a cell dimension of n;
  • W u , W r , W h R n × m are the weight matrices corresponding to the current input of the update gate, reset gate, and the hidden state;
  • V u , V r , V h R n × m are the weight matrices corresponding to the hidden output of the previous state for the current input of the update gate, reset gate, and the hidden state;
  • b u , b r , b h R n are the bias vectors corresponding to the current input of the update gate, reset gate, and the hidden state;
  • ⊙ is the Hadamard Product.
u ( t ) = δ s ( W u x i ( t ) + V u h ( t 1 ) + b u ) r ( t ) = δ s ( W r x i ( t ) + V r h ( t 1 ) + b r ) h ˜ ( t ) = δ h ( W h x i ( t ) + V h h ( t 1 ) + b h ) h ( t ) = u ( t ) h ˜ ( t ) + ( 1 u ( t ) ) h ( t 1 )

3.6.8. Bidirectional GRU

Similar to the BiLSTM, the Bidirectional GRU (BiGRU) considers both past and future information by employing a forward and backward GRU, i.e., G R U F and G R U B , respectively. The G R U F and G R U B are associated to two hidden states (Equation (23)): (1) h ( t ) which processes the input in a forward manner using G R U F , and (2) h ( t ) which processes the input in a backwards manner using G R U B . As for BiLSTM, the hidden states h ( t ) and h ( t ) are concatenated at every time-step to encode the information from both past and future contexts into one hidden state h ( t ) = [ h ( t ) h ( t ) ] .
h ( t ) = G R U F ( x i ) h ( t ) = G R U B ( x i )

3.7. Evaluation Module

We use accuracy, precision, and recall [49] to evaluate the models. For binary classification with the classes positive and negative, the following information is used to construct a confusion matrix that is afterward used to compute the evaluation metrics:
  • t p (True Positive) is the number of positive observations that are correctly classified;
  • f n (False Negative) is the number of positive observations that are incorrectly classified as negative;
  • f p (False Positive) is the number of false observations that are incorrectly classified as positive;
  • t n (True Negative) is the number of false observations that are correctly classified.
Accuracy (Equation (24)) measures the overall effectiveness of a classifier. Precision (Equation (25)) measures the class agreement of the data labels within the positive labels. Recall (Equation (26)) measures the effectiveness of a classifier in identifying positive labels.
A = t p + t n t p + t n + f p + f n
P = t p t p + f p
R = t p t p + f n

4. Experimental Results

In this section, we present the experimental results obtained using our methodology. Firstly, we introduce a human-verified sample from the Fake News Corpus [50] and present the results of the exploratory data analysis performed on it. Secondly, we present the experimental setup for our experiments as well as the hyperparameters and implementation packages for the models. Thirdly, we present the experimental results using the different sentence embeddings and classification methods on the Fake News Corpus sample. Lastly, we show the generalization of our observation by performing additional experiments on 5 additional datasets: LIAR multiclass [10], LIAR binary [51], Kaggle [12,18], Buzz Feed News [52], and TSHP-17 [53,54].

4.1. Dataset Details

For the experiments, we used a set of 20K English language news articles (10K reliable and 10K fake) selected from the Fake News Corpus [50] as it is widely used in current research [16,23,55,56,57]. Some of the labels might not be correct because the original dataset is not manually annotated. However, this shortcoming should not pose a practical issue for classification models, as ML/DL models generalize better when noise is added [58]. Instead, this should help the models to better generalize and remove overfitting. Additionally, we made sure that URL for the selected article point to the correct article by matching the titles and authors.
Before performing the experiments, we verified the label correctness for the sampled news articles using computer science students as annotators. In total, there were 40 student annotators to annotate 25K articles (12.5K reliable and 12.5K fake). We sampled more articles to mitigate any inconsistencies between two annotators as well as between the final annotation and the original label. For their annotation work, the students obtained credits for different courses.
Before annotating the articles, the students received an instruction list that explains the annotation task. The annotation task included the following steps:
(1)
Verify that the title matches the title from the URL;
(2)
Verify that the content matches the content from the URL;
(3)
Verify that the authors match the authors from the URL;
(4)
Verify that the source matches the source from the URL;
(5)
Verify if the information is false or reliable;
Each article was manually verified by two annotators. If there was no consensus between the two, a third annotator was used to break the tie. In 99% of the cases, there was no requirement for adding a third annotator. In the experiments, we removed all the articles where no consensus was found as well as where a difference between the human annotation and the original label was found. In the end, we scaled down the sample to 20 K news articles.
Table 1 presents the corpus statistics and information before and after preprocessing. We observe that, although there is a small imbalance in the number of tokens between the classes, this imbalance is small enough not to add bias to the classification task. We also extracted the top 10 unigrams and the top topic using the NMF algorithm for topic modeling [59]. We used the NLTK [60] for extracting unigrams and scikit-learn [61] for NMF. We computed the average similarity with PolyFuzz [62] by employing the pre-trained FastText embedding on news articles (sim(FT)) and the base-case BERT (sim(BERT)). Analyzing both similarities, we conclude that the documents discuss the same topics (Table 1). For the neural networks (i.e., Perceptron, MLP, LSTM, BiLSTM, GRU, BiGRU), we use one-hot encoders to represent the labels.

4.2. Experimental Setup

In our experiments, we analyzed how well we can predict if an article is fake or reliable using multiple vectorizations: (1) the TFIDF vector space model, and (2) eight document embeddings (DocEmb) constructed using five word embeddings (Word2Vec CBOW, Word2Vec SG, FastText CBOW, FastText SG, and GloVe), and three transformers embeddings (BERT, RoBERTa, and BART). We trained our own word embeddings and TFIDF vectorizer. We trained each word embedding for 100 epochs using a window size of 10, a learning rate of 0.05, and a size of 128. For TFIDF, we ignored words that appeared in less than 4 documents and kept the top 5K relevant features. We used SpaCy for preprocessing, scikit-learn for implementing TFIDF, gensim [63] for the Word2Vec and FastText models, and the python-glove [64] package for GloVe. For the transformer embeddings, we used the pre-trained uncased large versions from HuggingFace [44] together with SimpleTransformers [65] and SentenceTransformers [66].
For the experiments, we used the following algorithms for classification: Naïve Bayes (NB), Perceptron, Multi-layer Perceptron (MLP), LSTM, Bidirectional LSTM (BiLSTM), GRU, and Bidirectional GRU (BiGRU). We used scikit-learn for implementing NB. We applied the Multinomial NB for the TFIDF experiments as documents can get sparse, and the Gaussian NB for the transformer and word embeddings experiments as the embeddings can have negative values. For the Gradient Boosted Trees, we used the XGBoost Python library [67].
The neural-based fake news detection module used Multi-layer Perceptron, LSTM, Bidirectional LSTM, GRU, and Bidirectional GRU. Each layer consists of 100 cells. The LSTM was configured as in [46], while the GRU was configured as presented in [47]. A Dense layer with 2 units and a sigmoid function as activation was used as the output.
For the LSTM, BiLSTM, GRU, and BiGRU models, we used the ADAM optimizer and a 64-batch size. All the neural network models were trained for 100 epochs with an Early-Stopping mechanism to mitigate overfitting. We employed Keras for implementing the neural models. For comparison, we used the free implementation of MisRoBÆRTa [23], made available by the authors on GitHub.
The code is publicly available on GitHub at https://github.com/DS4AI-UPB/It-s-all-in-the-Embedding.

4.3. Fake News Detection

For the experiments, we used an NVIDIA® DGX Station. Table 2 presents the results for the fake news detection task. We tested the models in 10 rounds and used a 70%–10%–20% train–validation–test split ratio. Each split shuffles the dataset and extracts a stratified sample initialized with different random seeds. We report the average and standard deviation for each metric.
The overall best performance (i.e., an accuracy of 99.36 % ) was obtained with the BiGRU model when employing the document embeddings obtained with BART. The BiLSTM model obtained similar accuracy as when using BART, i.e., 99.36 % . We observe that, regardless of the model, the best results are obtained when employing the document embeddings created with BART.
Regardless of the model, TFIDF obtained very good results. For some models, it even outperformed the document embeddings obtained from the word or transformer embeddings, e.g., Naïve Bayes with TFIDF obtained an accuracy of 92.69 % while Naïve Bayes with RoBERTa obtained an accuracy of 91.98 % , Gradient Boosted Trees with TFIDF obtained an accuracy of 98.76 % while Gradient Boosted Trees with RoBERTa obtained an accuracy of 98.72 % . It is worth noting that the worst results were obtained with Naïve Bayes when employing the document embeddings created using the word embeddings.
Taking a closer look at the neural networks, we observe that the model ranking when considering accuracy is as follows: (1) BiGRU, (2) BiLSTM, (3) GRU, (4) LSTM, (5) Multi-Layer Perceptron, and (6) Perceptron. We note that LSTM and GRU obtained similar results with regard to document embedding, i.e., the results obtained by LSTM with the document embedding employing Word2Vec CBOW are very similar to the results obtained by GRU with the same document embedding model. The same observation holds when comparing BiLSTM and BiGRU. Moreover, the neural models that use document embeddings created using word embeddings were outperformed by the ones using transformer embeddings and TFIDF. The highest accuracy ( 99.36 % ) was obtained by BiGRU with the document embeddings created using BART.
The TFIDF approach proves that the relevance of calculating the importance of each word from a document is an important factor for the fake news detection problem. This result is a direct consequence of the size of the document-term matrix used as the input. Moreover, the transformer embeddings obtained the best results among the document embeddings experiments as they manage to encode and preserve the context within the vector representation. When compared to the state-of-the-art model MisRoBÆRTa [23], the BiLSTM with BART obtained similar results, while the BiGRU with BART marginally outperformed the model with a 0.02 % difference in accuracy. We hypothesize that this difference in performance is due to the use of pre-trained transformers instead of fine-tuned versions.
The recall metric is the most relevant one for the fake news detection task because it calculates the documents correctly classified as fake relative to all the actual fake documents, regardless of the predicted label. No clear pattern emerges among the document embeddings to determine which has the overall best performance. For example, when using LSTM, the best performance was obtained with document embedding DocEmb Word2Vec CBOW ( 96.84 % ), followed closely by DocEmbFastText SG ( 96.72 % ), while, when using Multi-Layer Perceptron, the best performance was obtained by DocEmbFastText SG ( 95.48 % ), followed closely by DocEmbWord2Vec SG ( 95.40 % ).
Finally, by analyzing the results, we observed the following:
(1)
A simpler neural architecture offers similar or better results compared to complex deep learning architectures that employ multiple layers, i.e., in our comparison, we obtained similar results as the complex MisRoBÆRTa [23] architecture without fine-tuning the transformers;
(2)
The embeddings used to vectorize the textual data make all the difference in performance, i.e., the right embedding must be selected to obtain good results with a given model;
(3)
We need a data-driven approach to select the best model and the best embedding for our dataset.

4.4. Additional Experiments

In this section, we present more experiments using four additional datasets that are analyzed in detail in [68]. For this set of experiments, we compared our results with existing results from the current literature. Furthermore, we trained our own model for each dataset using MisRoBÆRTa [23], but we did not fine-tune the transformers. We used the pre-trained BART (facebook/bart-large) and RoBERTa (roberta-base) versions from HuggingFace [44]. We hypothesize that this is the reason we obtained similar results to the ones obtained with the models that use document embeddings with this state-of-the-art architecture.
Table 3 and Table 4 present experimental results obtained on the LIAR dataset [10]. For our experiments, we used the dataset as it was initially released, with 6 labels [10] (Table 3), and by balancing the dataset’s labels (Table 4) as proposed in [51]. To balance the labels, we created binary labels, i.e., all the texts that are not labeled with true are considered false. Using the same experimental configurations as presented in Section 4.2, we obtained results that are aligned with our original observations on the proposed dataset. Further, we obtained results similar to state-of-the-art results for the multi-label dataset, e.g., Wang [10] and Alhindi et al. [69] obtained an accuracy of ∼20%. For the binary classification, we obtained results that go beyond the the state of the art, e.g.,Upadhayay and Behzadan [51] obtains an accuracy of 70 % while we obtain an accuracy of 83.99 % with the LSTM model that employs the document embeddings constructed with GloVe.
Table 3 presents the results obtained by the different machine and deep learning algorithms on the LIAR dataset [10]. The dataset contains approximately 12.8K human annotated short statements collected using POLITIFACT.COM’s API. In this set of experiments, we used all the 6 labels of LIAR, i.e., pants-fire, false, barely-true, half-true, mostly-true, and true, to build our classification models. The dataset is highly imbalanced, as there are more news articles labeled with true than news articles labeled with the other five classes combined. Due to this high degree of imbalance, the models performed poorly. We observe that the best-performing models employ document embedding constructed with BART. The overall best performance model was Multi-Layer Perceptron with BART-built document embedding, with an accuracy of 25.89 % . The overall difference between the worst- and best-performing models is approximately 7 % . We note that for Naïve Bayes, the model trained with TFIDF obtained better scores than the models trained with document embedding. We observed no real difference in performance among the models trained with the document embeddings using word embeddings. This low performance is also present in the current literature [10,69], with accuracy scores very similar to the ones obtained by the models we trained.
To mitigate the poor performance obtained using all 6 labels of the LIAR dataset and to minimize the imbalance between the classes, we employed a binarization approach to the dataset. This approach is also used in the current literature. For example, Upadhayay and Behzadan [51] and Yang et al. [29] also use the LIAR dataset with 2 labels, i.e., true and false, to train their models. On this dataset, we observed that the performance of all the models improved. Naïve Bayes trained on document embeddings obtained the worst results. The overall best results were obtained by LSTM with the document embedding constructed with GloVe, with an accuracy score of 83.99 % . Furthermore, Naïve Bayes, Gradient Boosted Trees, and Perceptron obtained better results with the TFIDF vectorization. The performance of these models is directly impacted by TFIDF’s features. The proposed approach outperforms more complex models proposed in the current literature, e.g., CNN with BERT-base embeddings [51] obtained an accuracy of 70 % and UFD [29] obtained an accuracy of 75.90 % .
Table 5, Table 6 and Table 7 present the experimental results obtained on the Kaggle [12,18], Buzz Feed News [52], and TSHP-17 datasets as presented in [53,54]. Both Kaggle and Buzz Feed News are binary datasets, i.e., with the levels reliable and false. To emphasize that the embedding makes the main difference and that the models can generalize when we move from binary to multi-class classification, we used the multilabel dataset TSHP-17, which has the following 3 classes: satire, hoax, and propaganda. For this set of experiments, we used the same experimental setup and algorithm configurations as presented above. Again, we obtained results that are aligned with our original observations, reinforcing our claims.
Table 5 presents the results obtained on the Kaggle dataset [12,18]. We observed that only for the Gradient Boosted Trees and Multi-Layer Perceptron models, the document embeddings obtained with Word2Vec SG and FastText SG outperformed their CBOW counterparts. When analyzing the same document embedding, i.e., DocEmb, we observed very little difference in performance among the neural models. As we used early stopping mechanisms, the neural network models did not overfit. Among the document embeddings employing transformers, the ones that use BART obtained the best results across all experiments. With an accuracy of 99.80 % , the overall best-performing model is the Bidirectional LSTM with document embeddings constructed with BART, i.e., DocEmb BART. The results show that our approach outperforms more complex models proposed in the current literature, e.g., FNDNet [12] obtained an accuracy of 98.36 % , FakeBERT [18] obtained an accuracy of 98.90 % , and C-CNN [24] obtained an accuracy of 99.90 % . We observe that C-CNN, a large neural model with multiple layers that also concatenates the results of three CNN models, outperforms the Bidirectional LSTM in terms of average accuracy on the Kaggle dataset by only 0.10 % . We also want to emphasize that the results in Table 5 present the mean over 10 runs for each metric per model and embedding pair. Thus, if we only take the best-performing model as in the case of the C-CNN results presented by Sedik et al. [24], then the Bidirectional LSTM model manages to obtain an accuracy of 99.92 % ( = 99.80 % (mean accuracy) + 0.12 % (standard deviation)).
Table 6 presents the results obtained on the Buzz Feed News dataset [52]. On this dataset, we observed that all the models obtained good results with TFIDF, such that some models that employ the TFIDF vectorization outperformed the document embeddings constructed with word and transformer embeddings, see, e.g., the results for LSTM, Bidirectional LSTM, GRU, and Bidirectional GRU. With an accuracy of 79.78 % , the overall best-performing model is Perceptron with BART document embeddings. For all the models, there is very little difference between the document embeddings that employ CBOW and their Skip-Gram counterparts. The results show that our approach outperforms more complex models proposed in the current literature, e.g., SVM [52] obtained an accuracy of 78.00 % and UFD [29] obtained an accuracy of 67.90 % .
Table 7 presents our final experiments, on the TSHP-17 dataset [53,54]. We observed that the document embeddings (DocEmbs) obtained with Word2Vec SG and FastText SG outperformed their CBOW counterparts only for the Gradient Boosted Trees model. All the models that employ TFIDF outperformed their counterparts that employ document embedding built with word embeddings. Among the document embeddings employing a transformer, the ones built with BART obtained the best results with regards to the model, while the ones that employ RoBERTa obtained the worst results. With an accuracy of 99.65 % , the overall best-performing model is the Bidirectional LSTM with BART document embeddings. For the same DocEmb, we observed very little difference in performance among the neural models. The results show that our approach outperforms more complex models proposed in the current literature, e.g., Proppy’s [70] accuracy is 98.36 % .
Thus, in conclusion, we show, on five additional datasets, that:
(1)
A simpler neural architecture offers at least similar or better results as complex architectures that employ multiple layers, and
(2)
The difference in performance lies in the embeddings used to vectorize the textual data.
Furthermore, we generally obtained better results than in other current state-of-the-art work:
(1)
On the LIAR dataset with 6 labels, Wang [10] obtained an F 1 -Score of 27.7 % using Hybrid CNNs and Alhindi et al. [69] obtained an F 1 -Score of 26 % using BiLSTM, while we obtained an accuracy of 25.89 % using Multi-Layer Perceptron with the document embeddings employing BART;
(2)
On the LIAR dataset with 2 labels, Upadhayay and Behzadan [51] obtained an accuracy of 70 % using CNN with BERT-base embeddings, while we obtained an accuracy of 83.99 % using LSTM with the document embeddings employing GloVe;
(3)
On the Kaggle dataset, the large deep learning model FakeBERT [18] obtained an accuracy of 98.90 % and C-CNN [24] obtained an accuracy of 99.90 % , while we obtained an accuracy of 99.80 % using a simple Bidirectional LSTM with the document embeddings employing BART;
(4)
On the Buzz Feed News dataset, Horne and Adali [52] obtained an accuracy of 78 % using a linear SVM, while we obtained an accuracy of 79.78 % using Perceptron with the document embeddings employing BART;
(5)
On the TSHP-17 dataset, Barrón-Cedeño et al. [54] obtained an accuracy of 97.58 % using Proppy [70], while we obtained 99.65 % using Bidirectional LSTM with the document embeddings employing BART;
To sum up, this set of experiments again enforces our observations that the embedding is more important than the complexity of the classification architecture. Furthermore, there is no generic model that offers the best performance regardless of the dataset. Thus, a data-driven approach together with hyper-parameter tuning and ablation testing should be considered when the goal is to determine the best-performing model for a given dataset.

5. Discussion

Word embeddings manage to capture both local and global contexts as defined in Truică et al. [71]. These help the machine learning algorithms to model and learn the text context, syntax, and semantics, but fail to differentiate among the words’ grammatical functions, i.e., the same word embedding is computed for a word regardless of its part-of-speech. On the other hand, transformer embeddings manage to learn the linguistic meaning of words, as they manage to preserve context by design. Thus, the same word has a different embedding depending on its lexical sense and concept as well as part-of-speech. Based on this, we can observe that the experiments that use document embeddings that employ transformers perform better than those that employ word embedding on average. The most interesting results, however, are those obtained with the document representation obtained with TFIDF. We observed that only the frequency-based importance of a word to a document within a textual corpus has a high impact on the models’ performance. As a general observation, we observed very little difference in performance among the neural models when using the same document embedding.
The experimental results show that the DocEmbs that use Word2Vec and FastText obtain very similar results, with a difference of ∼±2% when using Perceptron, and ∼±0.5% when using LSTM. GloVe based DocEmb obtains the best results together with the LSTM model on the LIAR dataset when using 2 labels. For the sample extracted from the Fake News Corpus as well as the LIAR with 6 labels, Kaggle, and TSHP-17 datasets, the BART DocEmb obtains the best results with different classification algorithms. We can conclude that there is no clear result for a classification model that generalizes well regardless of the dataset.
From the experimental results, we could not determine a clear winner with regards to document embedding and classification model. We observed empirically that the best-performing classification model changes with the dataset and the document embedding employed.
Our DocEmb solutions were compared to the results we obtained when employing MisRoBÆRTa [23], a more complex state-of-the-art model that employs fine-tune BART and RoBERTa embeddings. We note that we did not use fine-tuning for our dataset as the authors did in the original work [23]. Thus, we used the pre-trained BART (facebook/bart-large) and RoBERTa (roberta-base) from HuggingFage [44]. Furthermore, we also compared the results we obtained on each dataset with the results obtained with other state-of-the-art models presented in the current literature.
In our experiments, we obtained results that lead to the following observation: feature selection is more important than the Deep Learning Architecture used for classification. To put it bluntly, the need to stack layers upon layers of neural cells, just to claim a novel architecture, does not solve real-world problems, it just exacerbates out of proportion our understanding of how to use Machine Learning/Deep Learning for Natural Language Processing tasks.
To conclude our findings:
(1)
A simpler neural architecture offers similar if not better results as complex deep learning architectures that employ multiple layers, i.e., in our comparison, we obtained similar results as the complex MisRoBÆRTa [23] architecture, better than state-of-the-art results, i.e., FakeBERT [18], and Poppy [70];
(2)
The embeddings used to vectorize the textual data makes all the difference in performance, i.e., the right embedding must be selected to obtain good results with a given model;
(3)
We need a data-driven approach to select the best model and the best embedding for our dataset;
(4)
The way the word embedding manages to encapsulate the semantic, syntactic, and context features improves the performance of the classification models.

6. Conclusions

In this article, we presented a new approach for fake news detection using document embeddings (DocEmbs). We also proposed a benchmark to establish the most efficient ways for finding misleading information. To detect fake news, we used multiple machine learning algorithms together with DocEmbs built using either TFIDF, or word and transformer embeddings: Word2Vec SG and CBOW, FastText SG and CBOW, GloVe, BERT, RoBERTa, and BART.
Our approach emphasizes the importance of an overall document representation when dealing with the task of fake news detection and shows state-of-the-art performance results. Depending on the dataset, the results show that BiGRU/BiLSTM with DocEmb BART outperform the other models. In the experiments, we obtained better results than state-of-the-art Deep Neural Network models, even though we used a simpler Deep Neural Network Architecture. Additionally, we obtained similar results as MisRoBÆRTa [23] when using pre-trained BART (facebook/bart-large) and RoBERTa (roberta-base) from HuggingFace [44]. These are significant results, not because of the evaluation scores but because of the complexity of the models. The main takeaway of this work is that a simpler neural architecture offers similar if not better results as complex architectures that employ multiple layers. We observe that the most relevant factor is the embedding employed for classification, as it can really make a difference.
In future research, we plan to use sentiment analysis with fake news detection to determine if there is a correlation between polarity and veracity. We also aim to use ensemble models that combine our proposed method with existing methods to determine if the performance of fake news detection is improved.

Author Contributions

Conceptualization, C.-O.T. and E.-S.A.; methodology, C.-O.T. and E.-S.A.; software, C.-O.T. and E.-S.A.; validation, C.-O.T. and E.-S.A.; formal analysis, C.-O.T. and E.-S.A.; investigation, C.-O.T. and E.-S.A.; resources, C.-O.T. and E.-S.A.; data curation, C.-O.T. and E.-S.A.; writing—original draft preparation, C.-O.T. and E.-S.A.; writing—review and editing, C.-O.T. and E.-S.A.; visualization, C.-O.T. and E.-S.A.; supervision, C.-O.T. and E.-S.A.; project administration, C.-O.T. and E.-S.A.; funding acquisition, C.-O.T. and E.-S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially funded by the “AI-based conversational agent for misinformation fact-checking” project financed through the OPTIM Research framework (POCU grant No. 62461/03.06.2022, SMIS code 153735) and partially funded by the University Politehnica of Bucharest through the PubArt program.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original dataset used in this study is publicly available at https://github.com/several27/FakeNewsCorpus (accessed on 27 December 2022).

Acknowledgments

This work was done within the “AI-based conversational agent for misinformation fact-checking” project financed through the OPTIM Research framework (POCU grant no. 62461/03.06.2022, SMIS code 153735) and partially funded by the University Politehnica of Bucharest through the PubArt program.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
TFTerm Frequency
IDFInverse Document Frequency
TFIDFTerm Frequency-Inverse Document Frequency
CLDFClass Label Frequency Distance
SGSkip-Gram
CBOWCommon Bag Of Words
GloVeGlobal Vectors
DocEmbDocument Embedding
BERTBidirectional Encoder Representations from Transformers
RoBERTaRobustly Optimized BERT pre-training Approach
XLM-RoBERTaCross-Lingual RoBERTa
BARTBidirectional and Auto-Regressive Transformers
NBNaïve Bayes
MNBMultinomial Naïve Bayes
GNBGaussian Naïve Bayes
SVMSupport Vector Machine
LogRegLogistic Regression
UFDUnsupervised Fake News Detection Framework
MLPMulti-layer Perceptron
RNNRecurrent Neural Network
CNNConvolutional Neural Networks
C-CNNConcatenated CNN
LSTMLong Short-Term Memory
BiLSTMBidirectional Long Short-Term Memory
GRUGated Recurrent Unit
BiGRUBidirectional Gated Recurrent Unit

References

  1. Truică, C.O.; Apostol, E.S.; Ștefu, T.; Karras, P. A Deep Learning Architecture for Audience Interest Prediction of News Topic on Social Media. In Proceedings of the International Conference on Extending Database Technology (EDBT2021), Nicosia, Cyprus, 23–26 March 2021; pp. 588–599. [Google Scholar] [CrossRef]
  2. Mustafaraj, E.; Metaxas, P.T. The Fake News Spreading Plague. In Proceedings of the ACM on Web Science Conference, Troy, NY, USA, 25–28 June 2017; pp. 235–239. [Google Scholar] [CrossRef]
  3. Ruths, D. The misinformation machine. Science 2019, 363, 348. [Google Scholar] [CrossRef] [PubMed]
  4. Bastos, M.T.; Mercea, D. The Brexit Botnet and User-Generated Hyperpartisan News. Soc. Sci. Comput. Rev. 2017, 37, 38–54. [Google Scholar] [CrossRef] [Green Version]
  5. Bovet, A.; Makse, H.A. Influence of fake news in Twitter during the 2016 US presidential election. Nat. Commun. 2019, 10, 7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Rzymski, P.; Borkowski, L.; Drąg, M.; Flisiak, R.; Jemielity, J.; Krajewski, J.; Mastalerz-Migas, A.; Matyja, A.; Pyrć, K.; Simon, K.; et al. The Strategies to Support the COVID-19 Vaccination with Evidence-Based Communication and Tackling Misinformation. Vaccines 2021, 9, 109. [Google Scholar] [CrossRef]
  7. Truică, C.O.; Apostol, E.S.; Paschke, A. Awakened at CheckThat! 2022: Fake news detection using BiLSTM and sentence transformer. In Proceedings of the Working Notes of the Conference and Labs of the Evaluation Forum (CLEF2022), Bologna, Italy, 5–8 September 2022; pp. 749–757. [Google Scholar]
  8. European Commission. Fighting Disinformation; European Commission: Brussels, Belgium, 2020. [Google Scholar]
  9. Chen, Y.; Conroy, N.K.; Rubin, V.L. News in an online world: The need for an “automatic crap detector”. Proc. Assoc. Inf. Sci. Technol. 2015, 52, 1–4. [Google Scholar] [CrossRef]
  10. Wang, W.Y. “Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; pp. 422–426. [Google Scholar] [CrossRef]
  11. Conroy, N.K.; Rubin, V.L.; Chen, Y. Automatic deception detection: Methods for finding fake news. Proc. Assoc. Inf. Sci. Technol. 2015, 52, 1–4. [Google Scholar] [CrossRef] [Green Version]
  12. Kaliyar, R.K.; Goswami, A.; Narang, P.; Sinha, S. FNDNet—A deep convolutional neural network for fake news detection. Cogn. Syst. Res. 2020, 61, 32–44. [Google Scholar] [CrossRef]
  13. Goldani, M.H.; Safabakhsh, R.; Momtazi, S. Convolutional neural network with margin loss for fake news detection. Inf. Process. Manag. 2021, 58, 102418. [Google Scholar] [CrossRef]
  14. Saleh, H.; Alharbi, A.; Alsamhi, S.H. OPCNN-FAKE: Optimized convolutional neural network for fake news detection. IEEE Access 2021, 9, 129471–129489. [Google Scholar] [CrossRef]
  15. Samantaray, S.; Kumar, A. Bi-directional Long Short-Term Memory Network for Fake News Detection from Social Media. In Intelligent and Cloud Computing; Springer: Berlin/Heidelberg, Germany, 2022; pp. 463–470. [Google Scholar] [CrossRef]
  16. Ilie, V.I.; Truică, C.O.; Apostol, E.S.; Paschke, A. Context-Aware Misinformation Detection: A Benchmark of Deep Learning Architectures Using Word Embeddings. IEEE Access 2021, 9, 162122–162146. [Google Scholar] [CrossRef]
  17. Jwa, H.; Oh, D.; Park, K.; Kang, J.; Lim, H. exBAKE: Automatic Fake News Detection Model Based on Bidirectional Encoder Representations from Transformers (BERT). Appl. Sci. 2019, 9, 4062. [Google Scholar] [CrossRef] [Green Version]
  18. Kaliyar, R.K.; Goswami, A.; Narang, P. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimed. Tools Appl. 2021, 80, 11765–11788. [Google Scholar] [CrossRef] [PubMed]
  19. Kula, S.; Choraś, M.; Kozik, R. Application of the BERT-Based Architecture in Fake News Detection. In Conference on Complex, Intelligent, and Software Intensive Systems; Springer: Berlin/Heidelberg, Germany, 2020; pp. 239–249. [Google Scholar] [CrossRef]
  20. Mersinias, M.; Afantenos, S.; Chalkiadakis, G. CLFD: A Novel Vectorization Technique and Its Application in Fake News Detection. In Proceedings of the Language Resources and Evaluation Conference, Marseille, France, 11–16 May 2020; pp. 3475–3483. [Google Scholar]
  21. Mondal, S.K.; Sahoo, J.P.; Wang, J.; Mondal, K.; Rahman, M.M. Fake News Detection Exploiting TF-IDF Vectorization with Ensemble Learning Models. In Advances in Distributed Computing and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2022; pp. 261–270. [Google Scholar] [CrossRef]
  22. Aslam, N.; Khan, I.U.; Alotaibi, F.S.; Aldaej, L.A.; Aldubaikil, A.K. Fake Detect: A Deep Learning Ensemble Model for Fake News Detection. Complexity 2021, 2021, 5557784. [Google Scholar] [CrossRef]
  23. Truică, C.O.; Apostol, E.S. MisRoBÆRTa: Transformers versus Misinformation. Mathematics 2022, 10, 569. [Google Scholar] [CrossRef]
  24. Sedik, A.; Abohany, A.A.; Sallam, K.M.; Munasinghe, K.; Medhat, T. Deep fake news detection system based on concatenated and recurrent modalities. Expert Syst. Appl. 2022, 208, 117953. [Google Scholar] [CrossRef]
  25. Verma, P.K.; Agrawal, P.; Amorim, I.; Prodan, R. WELFake: Word Embedding Over Linguistic Features for Fake News Detection. IEEE Trans. Comput. Soc. Syst. 2021, 8, 881–893. [Google Scholar] [CrossRef]
  26. Shu, K.; Cui, L.; Wang, S.; Lee, D.; Liu, H. dEFEND: Explainable Fake News Detection. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 395–405. [Google Scholar] [CrossRef]
  27. Khattar, D.; Goud, J.S.; Gupta, M.; Varma, V. MVAE: Multimodal Variational Autoencoder for Fake News Detection. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2915–2921. [Google Scholar] [CrossRef]
  28. Zhang, J.; Dong, B.; Yu, P.S. FakeDetector: Effective Fake News Detection with Deep Diffusive Neural Network. In Proceedings of the 2020 IEEE 36th International Conference on Data Engineering (ICDE), Dallas, TX, USA, 20–24 April 2020; pp. 1826–1829. [Google Scholar] [CrossRef]
  29. Yang, S.; Shu, K.; Wang, S.; Gu, R.; Wu, F.; Liu, H. Unsupervised Fake News Detection on Social Media: A Generative Approach. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 5644–5651. [Google Scholar] [CrossRef] [Green Version]
  30. Wang, Y.; Qian, S.; Hu, J.; Fang, Q.; Xu, C. Fake News Detection via Knowledge-driven Multimodal Graph Convolutional Networks. In Proceedings of the 2020 International Conference on Multimedia Retrieval, Dublin, Ireland, 8–11 June 2020; pp. 540–547. [Google Scholar] [CrossRef]
  31. Le, Q.; Mikolov, T. Distributed representations of sentences and documents. In Proceedings of the International Conference on Machine Learning PMLR, Bejing, China, 22–24 June 2014; pp. 1188–1196. [Google Scholar]
  32. Cui, J.; Kim, K.; Na, S.H.; Shin, S. Meta-Path-based Fake News Detection Leveraging Multi-level Social Context Information. In Proceedings of the Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, GA, USA, 17–21 October 2022; pp. 325–334. [Google Scholar]
  33. Singh, L. Fake news detection: A comparison between available Deep Learning techniques in vector space. In Proceedings of the 2020 IEEE 4th Conference on Information & Communication Technology (CICT), Chennai, India, 3–5 December 2020; pp. 1–4. [Google Scholar]
  34. Truică, C.O.; Apostol, E.S.; Darmont, J.; Assent, I. TextBenDS: A Generic Textual Data Benchmark for Distributed Systems. Inf. Syst. Front. 2021, 23, 81–100. [Google Scholar] [CrossRef]
  35. Paltoglou, G.; Thelwall, M. A Study of Information Retrieval Weighting Schemes for Sentiment Analysis. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, Stroudsburg, PA, USA, 11–16 July 2010; pp. 1386–1395. [Google Scholar]
  36. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. In Proceedings of the Workshop Proceedings of the International Conference on Learning Representations 2013, Scottsdale, AZ, USA, 2–4 May 2013; pp. 1–12. [Google Scholar]
  37. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed Representations of Words and Phrases and their Compositionality. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; Volume 26, pp. 1–9. [Google Scholar]
  38. Bojanowski, P.; Grave, E.; Joulin, A.; Mikolov, T. Enriching word vectors with subword information. Trans. Assoc. Comput. Linguist. 2017, 5, 135–146. [Google Scholar] [CrossRef] [Green Version]
  39. Pennington, J.; Socher, R.; Manning, C.D. GloVe: Global vectors for word representation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar] [CrossRef]
  40. Mikolov, T.; Grave, E.; Bojanowski, P.; Puhrsch, C.; Joulin, A. Advances in Pre-Training Distributed Word Representations. In Proceedings of the International Conference on Language Resources and Evaluation, Miyazaki, Japan, 7–12 May 2018; pp. 52–55. [Google Scholar]
  41. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the Conference of the North Association for Computational Linguistics, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  42. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  43. Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7871–7880. [Google Scholar] [CrossRef]
  44. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Honolulu, HI, USA, 16–20 November 2020; pp. 38–45. [Google Scholar] [CrossRef]
  45. Rennie, J.D.M.; Shih, L.; Teevan, J.; Karger, D.R. Tackling the Poor Assumptions of Naive Bayes Text Classifiers. In Proceedings of the International Conference on International Conference on Machine Learning, Los Angeles, CA, USA, 23–24 June 2003; pp. 616–623. [Google Scholar]
  46. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  47. Cho, K.; van Merrienboer, B.; Bahdanau, D.; Bengio, Y. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches. In Proceedings of the SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation, Doha, Qatar, 25 October 2014; pp. 103–111. [Google Scholar] [CrossRef]
  48. Hewamalage, H.; Bergmeir, C.; Bandara, K. Recurrent Neural Networks for Time Series Forecasting: Current status and future directions. Int. J. Forecast. 2021, 37, 388–427. [Google Scholar] [CrossRef]
  49. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  50. Szpakowski, M. FakeNewsCorpus. 2020. Available online: https://github.com/several27/FakeNewsCorpus (accessed on 27 December 2022).
  51. Upadhayay, B.; Behzadan, V. Sentimental LIAR: Extended Corpus and Deep Learning Models for Fake Claim Classification. In Proceedings of the 2020 IEEE International Conference on Intelligence and Security Informatics (ISI), Virtual Event, 9–10 November 2020; pp. 1–6. [Google Scholar] [CrossRef]
  52. Horne, B.; Adali, S. This Just In: Fake News Packs A Lot In Title, Uses Simpler, Repetitive Content in Text Body, More Similar to Satire Than Real News. In Proceedings of the International AAAI Conference on Web and Social Media, Montreal, QC, Canada, 15–18 May 2017; pp. 759–766. [Google Scholar]
  53. Rashkin, H.; Choi, E.; Jang, J.Y.; Volkova, S.; Choi, Y. Truth of Varying Shades: Analyzing Language in Fake News and Political Fact-Checking. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 2931–2937. [Google Scholar] [CrossRef] [Green Version]
  54. Barrón-Cedeño, A.; Jaradat, I.; Da San Martino, G.; Nakov, P. Proppy: Organizing the news based on their propagandistic content. Inf. Process. Manag. 2019, 56, 1849–1864. [Google Scholar] [CrossRef]
  55. Kurasinski, L.; Mihailescu, R.C. Towards Machine Learning Explainability in Text Classification for Fake News Detection. In Proceedings of the 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 14–17 December 2020; pp. 775–781. [Google Scholar] [CrossRef]
  56. Nørregaard, J.; Horne, B.D.; Adalı, S. NELA-GT-2018: A Large Multi-Labelled News Dataset for the Study of Misinformation in News Articles. In Proceedings of the International AAAI Conference on Web and Social Media, Münich, Germany, 11–14 June 2019; pp. 630–638. [Google Scholar]
  57. Kwak, H.; An, J.; Ahn, Y.Y. A Systematic Media Frame Analysis of 1. In 5 Million New York Times Articles from 2000 to 2017. In Proceedings of the ACM Conference on Web Science, Southampton, UK, 6–10 July 2020; pp. 305–314. [Google Scholar] [CrossRef]
  58. Reed, R.D.; Marks II, R.J. Neural Smithing: Supervised Learning in Feedforward Artificial Neural Networks; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  59. Arora, S.; Ge, R.; Moitra, A. Learning Topic Models – Going beyond SVD. In Proceedings of the Annual Symposium on Foundations of Computer Science, Washington, DC, USA, 20–23 October 2012; pp. 1–10. [Google Scholar] [CrossRef] [Green Version]
  60. Bird, S.; Loper, E.; Klein, E. Natural Language Processing with Python; O’Reilly: Sebastopol, CA, USA, 2009. [Google Scholar]
  61. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  62. Grootendorst, M. PolyFuzz. 2020. Available online: https://maartengr.github.io/PolyFuzz/ (accessed on 27 December 2022).
  63. Řehůřek, R.; Sojka, P. Software Framework for Topic Modelling with Large Corpora. In Proceedings of the Workshop on New Challenges for NLP Frameworks, Valletta, Malta, 22 May 2010; pp. 45–50. [Google Scholar]
  64. Kula, M. Python-Glove. 2020. Available online: https://github.com/maciejkula/glove-python (accessed on 27 December 2022).
  65. Rajapakse, T. SimpleTransformers. 2021. Available online: https://simpletransformers.ai/ (accessed on 27 December 2022).
  66. Reimers, N.; Gurevych, I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Hong Kong, China, 3–7 November 2019; pp. 3982–3992. [Google Scholar] [CrossRef] [Green Version]
  67. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  68. D’Ulizia, A.; Caschera, M.C.; Ferri, F.; Grifoni, P. Fake news detection: A survey of evaluation datasets. PeerJ Comput. Sci. 2021, 7, e518. [Google Scholar] [CrossRef]
  69. Alhindi, T.; Petridis, S.; Muresan, S. Where is Your Evidence: Improving Fact-checking by Justification Modeling. In Proceedings of the First Workshop on Fact Extraction and VERification (FEVER), Brussels, Belgium, 24 July 2018; pp. 85–90. [Google Scholar] [CrossRef]
  70. Barrón-Cedeño, A.; Martino, G.D.S.; Jaradat, I.; Nakov, P. Proppy: A System to Unmask Propaganda in Online News. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 9847–9848. [Google Scholar] [CrossRef] [Green Version]
  71. Truică, C.O.; Apostol, E.S.; Șerban, M.L.; Paschke, A. Topic-Based Document-Level Sentiment Analysis Using Contextual Cues. Mathematics 2021, 9, 2722. [Google Scholar] [CrossRef]
Figure 1. Proposed Pipeline.
Figure 1. Proposed Pipeline.
Mathematics 11 00508 g001
Table 1. Dataset description, statistics, and information.
Table 1. Dataset description, statistics, and information.
Statistics before Preprocessing#Tokens per Document#Tokens per Class
ClassEncodingDescriptionMeanMinMaxStdDevUniqueAll
Fake News1Fabricated or distorted information517.8368 812883.35119 2835 178 300
Reliable0Reliable information575.66710 541602.1682 2035 756 643
Entire Dataset Statistics546.75610 541618.63159 11310 934 943
Textual Information after PreprocessingSim (FT)Sim (BERT)
Top-10 UnigramsFake Newspeople time government world year story market American God day0.830.93
Reliablepeople God Christian government American time world war America political
Top-1 TopicFake Newspeople Trump year day government time state world market war0.840.94
Reliablechurch Trump people God president war state year Bush government
Table 2. Fake news detection results on the sample extracted from the Fake News Corpus.
Table 2. Fake news detection results on the sample extracted from the Fake News Corpus.
Naïve BayesGradient Boosted Trees
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF92.69 ± 0.2591.72 ± 0.3993.86 ± 0.4198.76 ± 0.2199.79 ± 0.0897.72 ± 0.43
DocEmbWord2VecCBOW66.29 ± 0.5360.78 ± 0.4291.85 ± 0.2795.67 ± 0.1896.17 ± 0.3795.13 ± 0.34
DocEmbWord2VecSG53.10 ± 0.3751.74 ± 0.2092.46 ± 0.7797.10 ± 0.2197.61 ± 0.3196.56 ± 0.16
DocEmbFastTextCBOW56.13 ± 0.5353.59 ± 0.3391.45 ± 0.5894.90 ± 0.2495.49 ± 0.3994.24 ± 0.46
DocEmbFastTextSG54.00 ± 0.6952.23 ± 0.3893.66 ± 0.9897.05 ± 0.2697.45 ± 0.3296.64 ± 0.49
DocEmbGloVe53.43 ± 0.4251.98 ± 0.2489.94 ± 0.7896.02 ± 0.3096.73 ± 0.3595.26 ± 0.43
DocEmbBERT80.90 ± 0.6474.94 ± 0.6792.87 ± 0.5897.43 ± 0.2197.75 ± 0.2997.10 ± 0.30
DocEmbRoBERTa91.98 ± 0.3194.05 ± 0.4489.63 ± 0.6097.38 ± 0.2298.72 ± 0.3196.02 ± 0.42
DocEmbBART89.13 ± 0.3283.71 ± 0.4697.19 ± 0.3798.26 ± 0.1998.18 ± 0.3198.35 ± 0.21
PerceptronMulti-Layer Perceptron
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF95.79 ± 0.3396.55 ± 0.4694.98 ± 0.4298.04 ± 0.1998.37 ± 0.2897.70 ± 0.33
DocEmbWord2VecCBOW93.61 ± 0.2794.22 ± 0.5392.93 ± 0.4994.96 ± 0.3195.26 ± 0.3694.63 ± 0.64
DocEmbWord2VecSG92.04 ± 0.3494.65 ± 0.4989.12 ± 0.9195.88 ± 0.3096.34 ± 0.6595.40 ± 1.09
DocEmbFastTextCBOW93.46 ± 0.3094.48 ± 0.6492.33 ± 0.6794.92 ± 0.2395.12 ± 0.7294.71 ± 0.67
DocEmbFastTextSG91.60 ± 0.4994.06 ± 0.5488.81 ± 0.7696.00 ± 0.3096.48 ± 0.3595.48 ± 0.74
DocEmbGloVe89.57 ± 0.5092.57 ± 0.6086.04 ± 1.1894.05 ± 0.3894.29 ± 0.8993.79 ± 0.60
DocEmbBERT97.09 ± 0.2197.50 ± 0.6296.66 ± 0.6298.34 ± 0.1998.56 ± 0.6298.11 ± 0.56
DocEmbRoBERTa96.19 ± 0.5096.89 ± 1.6295.49 ± 0.9997.28 ± 0.5598.51 ± 1.3296.04 ± 0.47
DocEmbBART98.57 ± 0.1598.71 ± 0.2998.43 ± 0.3398.93 ± 0.1699.07 ± 0.5598.80 ± 0.39
LSTMBidirectional LSTM
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF97.88 ± 0.2398.20 ± 0.1997.55 ± 0.4097.84 ± 0.2498.21 ± 0.3097.46 ± 0.49
DocEmbWord2VecCBOW96.59 ± 0.3696.38 ± 1.0596.84 ± 1.1796.89 ± 0.2696.89 ± 0.5396.89 ± 0.38
DocEmbWord2VecSG96.23 ± 0.3196.70 ± 0.9795.76 ± 1.3296.39 ± 0.3796.81 ± 1.3095.98 ± 1.32
DocEmbFastTextCBOW96.16 ± 0.2696.22 ± 0.5596.11 ± 0.8596.30 ± 0.3296.63 ± 1.1995.98 ± 1.14
DocEmbFastTextSG96.61 ± 0.2796.52 ± 0.9196.72 ± 0.9096.79 ± 0.2296.78 ± 0.9196.82 ± 0.88
DocEmbGloVe94.66 ± 0.4894.92 ± 2.0294.45 ± 1.7294.86 ± 0.3795.24 ± 1.6794.51 ± 1.63
DocEmbBERT98.57 ± 0.3498.59 ± 0.7698.57 ± 1.1098.72 ± 0.4098.90 ± 0.8198.55 ± 0.87
DocEmbRoBERTa96.88 ± 1.5798.02 ± 2.9595.80 ± 1.5696.97 ± 1.3397.78 ± 2.8596.22 ± 0.74
DocEmbBART99.29 ± 0.1099.46 ± 0.1399.13 ± 0.1499.34 ± 0.0899.48 ± 0.1299.20 ± 0.11
GRUBidirectional GRU
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF97.88 ± 0.2498.28 ± 0.2797.47 ± 0.4597.84 ± 0.3098.04 ± 0.6097.64 ± 0.58
DocEmbWord2VecCBOW96.56 ± 0.2996.43 ± 0.7596.71 ± 1.0196.57 ± 0.2396.37 ± 0.9096.81 ± 0.69
DocEmbWord2VecSG96.21 ± 0.4495.91 ± 1.4496.58 ± 0.9796.35 ± 0.3896.45 ± 1.0996.26 ± 1.47
DocEmbFastTextCBOW96.12 ± 0.1696.54 ± 1.0495.70 ± 1.0096.20 ± 0.3396.47 ± 0.8895.91 ± 1.00
DocEmbFastTextSG96.40 ± 0.3596.11 ± 1.2296.74 ± 1.3296.76 ± 0.2296.76 ± 0.5996.77 ± 0.50
DocEmbGloVe94.62 ± 0.6495.60 ± 2.3093.66 ± 1.8194.84 ± 0.6095.80 ± 1.7493.86 ± 2.01
DocEmbBERT98.82 ± 0.1198.71 ± 0.5298.92 ± 0.4898.61 ± 0.4199.17 ± 0.4798.05 ± 1.14
DocEmbRoBERTa97.37 ± 0.6099.17 ± 0.4395.56 ± 1.4797.31 ± 0.4498.34 ± 1.2496.26 ± 0.72
DocEmbBART99.31 ± 0.1099.39 ± 0.2399.22 ± 0.1499.36 ± 0.0899.50 ± 0.0999.22 ± 0.09
AccuracyPrecisionRecall
MisRoBÆRTa [23]99.34 ± 0.0399.34 ± 0.0399.34 ± 0.02
Table 3. Fake news detection results on Liar dataset with 6 labels as presented in Wang [10].
Table 3. Fake news detection results on Liar dataset with 6 labels as presented in Wang [10].
Naïve BayesGradient Boosted Trees
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF23.07 ± 0.7024.16 ± 2.0323.07 ± 0.7023.00 ± 0.9322.86 ± 0.8923.00 ± 0.93
DocEmbWord2VecCBOW18.32 ± 1.0121.62 ± 1.4218.32 ± 1.0122.40 ± 0.5922.37 ± 0.6822.40 ± 0.59
DocEmbWord2VecSG20.42 ± 0.9621.74 ± 0.8420.42 ± 0.9623.12 ± 0.6923.29 ± 0.6923.12 ± 0.69
DocEmbFastTextCBOW17.19 ± 0.6321.89 ± 1.4517.19 ± 0.6322.68 ± 0.5522.63 ± 0.7122.68 ± 0.55
DocEmbFastTextSG19.85 ± 1.1021.57 ± 1.2219.85 ± 1.1022.93 ± 0.8323.00 ± 0.8022.93 ± 0.83
DocEmbGloVe17.60 ± 0.7721.31 ± 1.1817.60 ± 0.7721.99 ± 0.6321.72 ± 0.7121.99 ± 0.63
DocEmbBERT20.58 ± 0.7122.40 ± 1.1720.58 ± 0.7123.78 ± 0.8224.03 ± 0.9723.78 ± 0.82
DocEmbRoBERTa15.91 ± 1.0220.31 ± 1.3815.91 ± 1.0221.09 ± 1.0820.51 ± 0.8521.09 ± 1.08
DocEmbBART21.79 ± 0.9024.07 ± 1.0921.79 ± 0.9024.93 ± 0.7425.26 ± 0.8324.93 ± 0.74
PerceptronMulti-Layer Perceptron
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF23.71 ± 0.6824.79 ± 1.2623.71 ± 0.6823.04 ± 0.6623.08 ± 0.7723.04 ± 0.66
DocEmbWord2VecCBOW22.83 ± 0.6522.37 ± 0.6622.83 ± 0.6522.70 ± 0.9122.56 ± 0.8722.70 ± 0.91
DocEmbWord2VecSG23.46 ± 0.7323.19 ± 0.6823.46 ± 0.7323.26 ± 0.6523.15 ± 1.0923.26 ± 0.65
DocEmbFastTextCBOW22.34 ± 0.4921.63 ± 0.8122.34 ± 0.4922.72 ± 0.8922.16 ± 1.1922.72 ± 0.89
DocEmbFastTextSG23.62 ± 0.8223.29 ± 1.0823.62 ± 0.8223.48 ± 0.9223.47 ± 1.1323.48 ± 0.92
DocEmbGloVe23.64 ± 0.5522.54 ± 0.9723.64 ± 0.5523.24 ± 0.7121.94 ± 1.3123.24 ± 0.71
DocEmbBERT24.06 ± 1.0324.33 ± 1.0024.06 ± 1.0323.58 ± 0.6623.88 ± 0.8423.58 ± 0.66
DocEmbRoBERTa21.66 ± 1.5921.01 ± 1.6021.66 ± 1.5922.82 ± 0.6120.83 ± 1.2222.82 ± 0.61
DocEmbBART25.60 ± 0.4125.84 ± 0.1825.60 ± 0.4125.89 ± 0.7526.15 ± 0.7225.89 ± 0.75
LSTMBidirectional LSTM
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF21.77 ± 0.5021.77 ± 0.4821.77 ± 0.5021.62 ± 0.4921.60 ± 0.4821.62 ± 0.49
DocEmbWord2VecCBOW22.70 ± 0.9522.53 ± 0.9422.70 ± 0.9522.65 ± 0.6022.57 ± 0.6222.65 ± 0.60
DocEmbWord2VecSG23.66 ± 0.9923.46 ± 1.0223.66 ± 0.9923.51 ± 0.6923.14 ± 0.8423.51 ± 0.69
DocEmbFastTextCBOW22.40 ± 1.0622.23 ± 1.0722.40 ± 1.0622.53 ± 0.8522.49 ± 0.8622.53 ± 0.85
DocEmbFastTextSG23.50 ± 0.7623.24 ± 0.9023.50 ± 0.7623.45 ± 0.6123.41 ± 0.8923.45 ± 0.61
DocEmbGloVe23.59 ± 0.4622.90 ± 1.4023.59 ± 0.4623.08 ± 0.4222.77 ± 0.8823.08 ± 0.42
DocEmbBERT23.21 ± 0.5523.37 ± 0.5223.21 ± 0.5523.31 ± 0.7023.25 ± 0.7623.31 ± 0.70
DocEmbRoBERTa22.94 ± 1.0021.65 ± 1.0822.94 ± 1.0022.96 ± 0.6119.77 ± 1.6622.96 ± 0.61
DocEmbBART25.02 ± 0.5725.08 ± 0.5925.02 ± 0.5725.75 ± 0.6125.78 ± 0.6225.75 ± 0.61
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF21.62 ± 0.5021.63 ± 0.4721.62 ± 0.5021.43 ± 0.4421.44 ± 0.4621.43 ± 0.44
DocEmbWord2VecCBOW22.36 ± 1.0022.14 ± 0.9422.36 ± 1.0022.51 ± 0.7622.46 ± 0.7922.51 ± 0.76
DocEmbWord2VecSG23.47 ± 0.5323.02 ± 0.7223.47 ± 0.5323.47 ± 0.7123.16 ± 0.6623.47 ± 0.71
DocEmbFastTextCBOW22.62 ± 0.7722.48 ± 0.7222.62 ± 0.7722.51 ± 0.5322.42 ± 0.4422.51 ± 0.53
DocEmbFastTextSG23.34 ± 0.5523.08 ± 0.7623.34 ± 0.5523.54 ± 0.7123.28 ± 0.5823.54 ± 0.71
DocEmbGloVe23.47 ± 0.6422.84 ± 1.7623.47 ± 0.6423.21 ± 0.5622.93 ± 1.2723.21 ± 0.56
DocEmbBERT23.64 ± 0.3623.90 ± 0.5323.64 ± 0.3623.00 ± 0.7223.21 ± 0.8823.00 ± 0.72
DocEmbRoBERTa22.69 ± 0.6819.84 ± 1.9522.69 ± 0.6822.73 ± 0.7221.55 ± 2.2622.73 ± 0.72
DocEmbBART24.99 ± 0.6625.00 ± 0.6924.99 ± 0.6625.20 ± 0.8825.26 ± 0.8625.20 ± 0.88
AccuracyPrecisionRecall
MisRoBÆRTa [23]24.62 ± 0.3925.87 ± 0.6724.61 ± 0.39
F1-Score
Hybrid CNNs [10]27.70
BiLSTM [69]26.00
Table 4. Fake news detection results on Liar dataset with 2 labels as presented in Upadhayay and Behzadan [51].
Table 4. Fake news detection results on Liar dataset with 2 labels as presented in Upadhayay and Behzadan [51].
Naïve BayesGradient Boosted Trees
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF83.92 ± 0.0483.96 ± 0.0199.94 ± 0.0483.64 ± 0.2084.09 ± 0.0899.31 ± 0.19
DocEmbWord2VecCBOW67.02 ± 1.2186.09 ± 0.4372.43 ± 1.7683.28 ± 0.1984.15 ± 0.0798.68 ± 0.25
DocEmbWord2VecSG64.57 ± 3.5486.07 ± 0.3168.96 ± 5.0183.30 ± 0.3384.15 ± 0.1198.70 ± 0.35
DocEmbFastTextCBOW67.89 ± 2.1985.60 ± 0.3274.26 ± 3.2083.25 ± 0.2884.16 ± 0.1298.61 ± 0.29
DocEmbFastTextSG65.19 ± 4.3286.17 ± 0.4369.75 ± 6.1783.28 ± 0.2384.16 ± 0.1098.65 ± 0.32
DocEmbGloVe59.71 ± 1.8085.89 ± 0.5262.24 ± 2.5483.05 ± 0.2684.10 ± 0.0898.42 ± 0.29
DocEmbBERT61.21 ± 0.9186.76 ± 0.5863.49 ± 1.1483.38 ± 0.1984.16 ± 0.0998.80 ± 0.19
DocEmbRoBERTa60.04 ± 3.9885.30 ± 0.5363.35 ± 6.0783.36 ± 0.2084.02 ± 0.0899.01 ± 0.25
DocEmbBART62.11 ± 1.0387.65 ± 0.5263.87 ± 1.2583.46 ± 0.1984.36 ± 0.1298.58 ± 0.32
PerceptronMulti-Layer Perceptron
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF83.97 ± 0.0183.97 ± 0.0199.99 ± 0.0180.87 ± 0.6984.50 ± 0.1594.58 ± 1.13
DocEmbWord2VecCBOW83.88 ± 0.0683.96 ± 0.0299.88 ± 0.0683.96 ± 0.0483.98 ± 0.0299.97 ± 0.03
DocEmbWord2VecSG83.94 ± 0.0483.97 ± 0.0199.95 ± 0.0683.90 ± 0.1083.99 ± 0.0499.86 ± 0.10
DocEmbFastTextCBOW83.87 ± 0.0683.98 ± 0.0399.82 ± 0.0883.95 ± 0.0683.99 ± 0.0399.94 ± 0.07
DocEmbFastTextSG83.97 ± 0.0283.97 ± 0.0199.99 ± 0.0183.93 ± 0.0883.99 ± 0.0299.91 ± 0.10
DocEmbGloVe83.97 ± 0.0183.97 ± 0.0199.99 ± 0.0183.96 ± 0.0183.97 ± 0.0199.99 ± 0.01
DocEmbBERT83.81 ± 0.1183.98 ± 0.0599.74 ± 0.1483.18 ± 0.5284.25 ± 0.2598.37 ± 1.11
DocEmbRoBERTa83.96 ± 0.0383.97 ± 0.0199.99 ± 0.0383.97 ± 0.0183.97 ± 0.0199.99 ± 0.01
DocEmbBART83.44 ± 0.5584.37 ± 0.2898.53 ± 1.1781.63 ± 1.0184.97 ± 0.3394.91 ± 1.52
LSTMBidirectional LSTM
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF77.05 ± 0.8984.75 ± 0.2088.62 ± 1.1376.96 ± 0.8684.75 ± 0.1788.49 ± 1.08
DocEmbWord2VecCBOW81.43 ± 0.5784.44 ± 0.1395.47 ± 0.8880.51 ± 1.0284.63 ± 0.1893.83 ± 1.65
DocEmbWord2VecSG83.86 ± 0.1384.04 ± 0.0599.72 ± 0.1383.79 ± 0.1684.07 ± 0.0599.57 ± 0.17
DocEmbFastTextCBOW81.20 ± 0.7484.45 ± 0.2395.13 ± 1.0380.77 ± 1.1284.49 ± 0.3494.44 ± 1.63
DocEmbFastTextSG83.88 ± 0.1484.01 ± 0.0499.79 ± 0.1683.78 ± 0.1984.00 ± 0.0699.66 ± 0.19
DocEmbGloVe83.99 ± 0.0283.98 ± 0.0299.99 ± 0.0183.98 ± 0.0483.98 ± 0.0299.99 ± 0.01
DocEmbBERT80.37 ± 1.5384.84 ± 0.4993.31 ± 2.7479.69 ± 1.7685.05 ± 0.3592.00 ± 2.95
DocEmbRoBERTa83.97 ± 0.0183.97 ± 0.0199.99 ± 0.0183.97 ± 0.0183.97 ± 0.0199.99 ± 0.01
DocEmbBART81.41 ± 0.8584.93 ± 0.2294.66 ± 1.2181.43 ± 0.9485.14 ± 0.2294.34 ± 1.35
GRUBidirectional GRU
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF76.99 ± 0.5984.71 ± 0.1988.58 ± 0.6576.87 ± 0.6684.72 ± 0.2288.39 ± 0.94
DocEmbWord2VecCBOW81.52 ± 0.6784.49 ± 0.2495.54 ± 0.9580.52 ± 1.0184.58 ± 0.1593.92 ± 1.53
DocEmbWord2VecSG83.91 ± 0.1384.04 ± 0.0499.80 ± 0.1483.81 ± 0.1684.06 ± 0.0599.60 ± 0.18
DocEmbFastTextCBOW80.93 ± 1.1484.44 ± 0.2494.74 ± 1.9380.20 ± 0.9684.67 ± 0.2793.31 ± 1.55
DocEmbFastTextSG83.89 ± 0.1284.01 ± 0.0499.80 ± 0.1383.82 ± 0.1684.02 ± 0.0599.70 ± 0.16
DocEmbGloVe83.98 ± 0.0383.98 ± 0.0299.99 ± 0.0183.98 ± 0.0383.99 ± 0.0299.99 ± 0.01
DocEmbBERT79.86 ± 2.3084.73 ± 0.4092.75 ± 3.8379.94 ± 1.2584.93 ± 0.3692.54 ± 2.08
DocEmbRoBERTa83.97 ± 0.0183.97 ± 0.0199.99 ± 0.0183.97 ± 0.0183.97 ± 0.0199.99 ± 0.01
DocEmbBART81.22 ± 0.7585.13 ± 0.2794.06 ± 1.1980.60 ± 0.9085.24 ± 0.2293.01 ± 1.34
AccuracyPrecisionRecall
MisRoBÆRTa [23]81.15 ± 0.0781.15 ± 0.0781.16 ± 0.07
Accuracy
CNN with BERT-base embeddings [51]70.00
UFD [29]75.90
Table 5. Fake news detection results on the Kaggle dataset as presented in Kaliyar et al. [12], Kaliyar et al. [18].
Table 5. Fake news detection results on the Kaggle dataset as presented in Kaliyar et al. [12], Kaliyar et al. [18].
Naïve BayesGradient Boosted Trees
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF89.54 ± 0.2092.94 ± 0.3384.93 ± 0.6796.74 ± 0.1596.22 ± 0.3697.10 ± 0.29
DocEmbWord2VecCBOW71.14 ± 0.4178.89 ± 0.6355.41 ± 0.6291.90 ± 0.2392.01 ± 0.4691.24 ± 0.40
DocEmbWord2VecSG60.91 ± 0.5385.19 ± 1.6923.66 ± 1.1693.31 ± 0.3793.81 ± 0.3292.33 ± 0.55
DocEmbFastTextCBOW67.45 ± 0.6777.26 ± 1.0546.75 ± 1.0691.45 ± 0.2991.32 ± 0.6891.06 ± 0.49
DocEmbFastTextSG60.22 ± 0.2288.13 ± 1.0820.93 ± 0.4193.41 ± 0.3194.05 ± 0.3492.28 ± 0.58
DocEmbGloVe62.22 ± 0.4981.02 ± 1.2729.04 ± 0.9190.63 ± 0.3089.98 ± 0.4890.83 ± 0.36
DocEmbBERT70.52 ± 0.4782.03 ± 0.7050.34 ± 1.0692.91 ± 0.3493.58 ± 0.2291.69 ± 0.65
DocEmbRoBERTa81.86 ± 0.5687.45 ± 0.7873.18 ± 1.2792.42 ± 0.2292.51 ± 0.3391.82 ± 0.45
DocEmbBART90.14 ± 0.4291.69 ± 0.6387.64 ± 0.5299.06 ± 0.1198.82 ± 0.2499.24 ± 0.19
PerceptronMulti-Layer Perceptron
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF93.76 ± 0.2794.45 ± 0.5192.59 ± 0.5695.36 ± 0.2395.24 ± 0.4295.20 ± 0.44
DocEmbWord2VecCBOW90.15 ± 0.3590.84 ± 0.6388.67 ± 0.7191.54 ± 0.4091.67 ± 1.4290.89 ± 1.45
DocEmbWord2VecSG88.98 ± 0.4690.48 ± 0.5186.41 ± 0.8692.45 ± 0.3292.59 ± 1.1991.83 ± 1.29
DocEmbFastTextCBOW90.10 ± 0.4490.23 ± 0.9589.30 ± 0.6292.00 ± 0.4491.99 ± 1.0391.54 ± 1.68
DocEmbFastTextSG88.45 ± 0.5990.65 ± 0.8284.99 ± 1.0792.15 ± 0.4392.78 ± 1.0390.94 ± 0.81
DocEmbGloVe83.90 ± 0.5285.26 ± 1.1580.89 ± 1.9787.57 ± 0.5788.79 ± 2.1685.29 ± 2.64
DocEmbBERT92.09 ± 0.5092.84 ± 1.0190.74 ± 1.3694.80 ± 0.4794.51 ± 1.9594.89 ± 1.73
DocEmbRoBERTa91.62 ± 0.4091.02 ± 2.0391.92 ± 1.9492.74 ± 0.9693.03 ± 3.7692.31 ± 4.31
DocEmbBART99.73 ± 0.0999.66 ± 0.1199.78 ± 0.1399.77 ± 0.0799.70 ± 0.1499.82 ± 0.08
LSTMBidirectional LSTM
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF95.05 ± 0.2594.83 ± 0.4495.00 ± 0.5095.01 ± 0.2594.75 ± 0.3995.00 ± 0.34
DocEmbWord2VecCBOW93.70 ± 0.3294.11 ± 1.1192.87 ± 1.5093.81 ± 0.3993.60 ± 0.8293.68 ± 0.87
DocEmbWord2VecSG93.03 ± 0.3293.22 ± 1.3592.40 ± 1.2493.14 ± 0.5094.08 ± 1.7491.71 ± 1.80
DocEmbFastTextCBOW93.25 ± 0.6592.83 ± 2.2193.44 ± 2.1693.52 ± 0.3093.04 ± 1.6693.73 ± 2.08
DocEmbFastTextSG92.90 ± 0.4193.13 ± 1.1792.21 ± 1.1592.67 ± 0.5694.61 ± 1.6590.12 ± 2.34
DocEmbGloVe88.84 ± 0.8288.36 ± 2.9688.98 ± 2.9888.83 ± 0.7989.23 ± 3.2787.93 ± 4.70
DocEmbBERT96.31 ± 0.7296.15 ± 2.4196.36 ± 1.8496.36 ± 1.3296.08 ± 3.1096.59 ± 1.13
DocEmbRoBERTa93.52 ± 0.7894.41 ± 2.0292.24 ± 2.9493.12 ± 1.2994.87 ± 2.5390.93 ± 4.62
DocEmbBART99.79 ± 0.0699.74 ± 0.1199.84 ± 0.0899.80 ± 0.1299.72 ± 0.1899.87 ± 0.11
GRUBidirectional GRU
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF94.99 ± 0.2294.64 ± 0.3495.08 ± 0.4594.98 ± 0.2794.71 ± 0.3894.98 ± 0.38
DocEmbWord2VecCBOW93.67 ± 0.4193.66 ± 1.0293.32 ± 1.1993.68 ± 0.3093.48 ± 0.6793.54 ± 1.20
DocEmbWord2VecSG93.05 ± 0.3693.28 ± 1.4892.40 ± 1.3592.91 ± 0.7793.55 ± 2.4991.87 ± 2.32
DocEmbFastTextCBOW93.18 ± 0.5492.94 ± 2.2193.14 ± 1.8593.34 ± 0.5992.43 ± 1.2694.02 ± 1.01
DocEmbFastTextSG92.73 ± 0.7692.94 ± 2.5492.16 ± 2.1692.86 ± 0.4594.07 ± 1.1491.09 ± 2.10
DocEmbGloVe88.83 ± 0.7689.73 ± 3.2587.26 ± 3.5489.19 ± 0.6490.81 ± 1.4786.59 ± 2.85
DocEmbBERT96.25 ± 0.5296.81 ± 1.9595.50 ± 1.8796.37 ± 0.8197.71 ± 0.9094.77 ± 2.28
DocEmbRoBERTa92.32 ± 1.9392.10 ± 5.3192.80 ± 5.3592.93 ± 1.2795.36 ± 2.4890.01 ± 4.51
DocEmbBART99.77 ± 0.0799.72 ± 0.1499.82 ± 0.0999.78 ± 0.0799.72 ± 0.1299.83 ± 0.09
AccuracyPrecisionRecall
MisRoBÆRTa [23]97.57 ± 0.2997.58 ± 0.2897.57 ± 0.31
C-CNN [24]99.9099.9099.90
Accuracy
FNDNet [12]98.36
FakeBERT [18]98.90
Table 6. Fake news detection results on the Buzz Feed News dataset as presented in Horne and Adali [52].
Table 6. Fake news detection results on the Buzz Feed News dataset as presented in Horne and Adali [52].
Naïve BayesGradient Boosted Trees
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0178.29 ± 0.9970.70 ± 2.6278.29 ± 0.99
DocEmbWord2VecCBOW55.26 ± 1.5272.29 ± 2.1655.26 ± 1.5278.72 ± 0.8872.97 ± 2.7178.72 ± 0.88
DocEmbWord2VecSG54.42 ± 2.9972.24 ± 2.0554.42 ± 2.9978.97 ± 0.9173.30 ± 2.6678.97 ± 0.91
DocEmbFastTextCBOW49.13 ± 2.2671.76 ± 1.8649.13 ± 2.2678.54 ± 0.5772.26 ± 1.5078.54 ± 0.57
DocEmbFastTextSG55.58 ± 2.3573.95 ± 1.7655.58 ± 2.3578.94 ± 0.6673.46 ± 2.9778.94 ± 0.66
DocEmbGloVe48.63 ± 8.4069.79 ± 3.1348.63 ± 8.4078.16 ± 1.3571.89 ± 4.0178.16 ± 1.35
DocEmbBERT59.91 ± 1.7976.22 ± 0.8559.91 ± 1.7978.44 ± 0.7371.31 ± 1.9578.44 ± 0.73
DocEmbRoBERTa62.02 ± 7.6570.54 ± 1.5262.02 ± 7.6577.98 ± 0.7069.83 ± 4.1777.98 ± 0.70
DocEmbBART61.56 ± 1.2980.57 ± 1.1761.56 ± 1.2979.28 ± 1.1873.92 ± 2.5079.28 ± 1.18
PerceptronMulti-Layer Perceptron
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0178.29 ± 0.5070.91 ± 6.3278.29 ± 0.50
DocEmbWord2VecCBOW77.98 ± 0.3163.63 ± 3.2677.98 ± 0.3178.10 ± 0.5065.94 ± 5.5278.10 ± 0.50
DocEmbWord2VecSG77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0177.88 ± 0.0160.66 ± 0.0177.88 ± 0.01
DocEmbFastTextCBOW77.79 ± 0.6266.38 ± 5.9177.79 ± 0.6277.91 ± 0.6566.63 ± 3.3677.91 ± 0.65
DocEmbFastTextSG77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0177.88 ± 0.0160.66 ± 0.0177.88 ± 0.01
DocEmbGloVe77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0177.88 ± 0.0160.66 ± 0.0177.88 ± 0.01
DocEmbBERT77.76 ± 1.0268.20 ± 3.3977.76 ± 1.0277.60 ± 1.6570.76 ± 3.7077.60 ± 1.65
DocEmbRoBERTa77.85 ± 0.3363.79 ± 5.2377.85 ± 0.3377.88 ± 0.0160.66 ± 0.0177.88 ± 0.01
DocEmbBART79.78 ± 0.8475.84 ± 1.3079.78 ± 0.8479.75 ± 1.7575.40 ± 2.3079.75 ± 1.75
LSTMBidirectional LSTM
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF78.91 ± 0.9873.90 ± 1.4078.91 ± 0.9878.63 ± 0.8073.77 ± 1.6378.63 ± 0.80
DocEmbWord2VecCBOW77.88 ± 1.4070.88 ± 3.5377.88 ± 1.4077.23 ± 1.5570.58 ± 2.3077.23 ± 1.55
DocEmbWord2VecSG78.04 ± 0.1663.35 ± 1.9778.04 ± 0.1677.88 ± 0.5967.35 ± 2.4377.88 ± 0.59
DocEmbFastTextCBOW78.07 ± 0.5971.74 ± 2.2378.07 ± 0.5978.10 ± 0.7973.05 ± 1.5878.10 ± 0.79
DocEmbFastTextSG77.98 ± 0.2061.64 ± 1.6977.98 ± 0.2077.85 ± 0.7467.51 ± 5.9477.85 ± 0.74
DocEmbGloVe77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0177.85 ± 0.1762.74 ± 3.4477.85 ± 0.17
DocEmbBERT77.57 ± 2.9573.20 ± 2.1377.57 ± 2.9577.51 ± 2.2274.46 ± 2.1977.51 ± 2.22
DocEmbRoBERTa77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0177.88 ± 0.0160.66 ± 0.0177.88 ± 0.01
DocEmbBART78.07 ± 2.1676.31 ± 3.0178.07 ± 2.1677.35 ± 2.3975.97 ± 1.8977.35 ± 2.39
GRUBidirectional GRU
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF78.94 ± 1.2173.99 ± 2.1978.94 ± 1.2178.60 ± 1.1373.69 ± 1.4078.60 ± 1.13
DocEmbWord2VecCBOW77.57 ± 1.2670.72 ± 2.8077.57 ± 1.2677.32 ± 1.3171.04 ± 2.3077.32 ± 1.31
DocEmbWord2VecSG78.29 ± 0.3766.63 ± 2.5278.29 ± 0.3778.22 ± 1.0068.24 ± 2.5678.22 ± 1.00
DocEmbFastTextCBOW78.19 ± 0.7872.06 ± 2.3778.19 ± 0.7877.73 ± 1.5174.22 ± 1.0677.73 ± 1.51
DocEmbFastTextSG77.88 ± 0.3764.47 ± 3.6877.88 ± 0.3777.76 ± 0.7967.54 ± 2.8277.76 ± 0.79
DocEmbGloVe77.82 ± 0.3161.81 ± 2.4377.82 ± 0.3177.66 ± 0.5264.45 ± 3.1777.66 ± 0.52
DocEmbBERT78.54 ± 1.3972.23 ± 3.3978.54 ± 1.3974.86 ± 4.2872.99 ± 2.2174.86 ± 4.28
DocEmbRoBERTa77.88 ± 0.0160.66 ± 0.0177.88 ± 0.0177.88 ± 0.0160.66 ± 0.0177.88 ± 0.01
DocEmbBART77.48 ± 2.0875.96 ± 2.3877.48 ± 2.0876.45 ± 4.5677.31 ± 3.1476.45 ± 4.56
AccuracyPrecisionRecall
MisRoBÆRTa [23]77.39 ± 0.8377.39 ± 0.8377.39 ± 0.83
Accuracy
SVM [52]78.00
UFD [29]67.90
Table 7. Fake news detection results on TSHP-17 dataset as presented in Rashkin et al. [53] and Barrón-Cedeño et al. [54].
Table 7. Fake news detection results on TSHP-17 dataset as presented in Rashkin et al. [53] and Barrón-Cedeño et al. [54].
Naïve BayesGradient Boosted Trees
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF92.08 ± 0.1592.11 ± 0.1592.08 ± 0.1598.05 ± 0.1298.05 ± 0.1298.05 ± 0.12
DocEmbWord2VecCBOW70.06 ± 0.5073.01 ± 0.3670.06 ± 0.5095.64 ± 0.2695.63 ± 0.2695.64 ± 0.26
DocEmbWord2VecSG55.33 ± 0.6368.18 ± 0.3355.33 ± 0.6395.76 ± 0.2195.75 ± 0.2195.76 ± 0.21
DocEmbFastTextCBOW62.83 ± 0.4770.17 ± 0.6662.83 ± 0.4794.36 ± 0.2794.34 ± 0.2794.36 ± 0.27
DocEmbFastTextSG59.72 ± 0.5269.49 ± 0.6359.72 ± 0.5295.66 ± 0.2895.65 ± 0.2895.66 ± 0.28
DocEmbGloVe52.99 ± 0.5665.84 ± 0.4352.99 ± 0.5696.19 ± 0.2896.19 ± 0.2896.19 ± 0.28
DocEmbBERT84.65 ± 0.6387.60 ± 0.4284.65 ± 0.6398.18 ± 0.1098.18 ± 0.1098.18 ± 0.10
DocEmbRoBERTa52.15 ± 0.3365.77 ± 0.6452.15 ± 0.3379.15 ± 0.3679.11 ± 0.3879.15 ± 0.36
DocEmbBART94.08 ± 0.2894.66 ± 0.2694.08 ± 0.2899.01 ± 0.1099.01 ± 0.1099.01 ± 0.10
PerceptronMulti-Layer Perceptron
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF97.64 ± 0.1397.64 ± 0.1397.64 ± 0.1397.54 ± 0.1697.54 ± 0.1697.54 ± 0.16
DocEmbWord2VecCBOW94.24 ± 0.2394.22 ± 0.2494.24 ± 0.2396.11 ± 0.2296.11 ± 0.2296.11 ± 0.22
DocEmbWord2VecSG90.14 ± 0.2990.09 ± 0.2990.14 ± 0.2993.91 ± 0.2493.89 ± 0.2493.91 ± 0.24
DocEmbFastTextCBOW93.14 ± 0.3093.14 ± 0.3093.14 ± 0.3095.63 ± 0.2495.64 ± 0.2495.63 ± 0.24
DocEmbFastTextSG90.00 ± 0.2089.94 ± 0.2190.00 ± 0.2093.64 ± 0.3093.64 ± 0.2993.64 ± 0.30
DocEmbGloVe90.97 ± 0.2790.95 ± 0.2790.97 ± 0.2794.04 ± 0.3094.05 ± 0.2994.04 ± 0.30
DocEmbBERT98.44 ± 0.1598.44 ± 0.1598.44 ± 0.1598.78 ± 0.1498.78 ± 0.1398.78 ± 0.14
DocEmbRoBERTa77.79 ± 1.8578.89 ± 0.6577.79 ± 1.8580.30 ± 1.5581.29 ± 0.6180.30 ± 1.55
DocEmbBART99.54 ± 0.0599.54 ± 0.0599.54 ± 0.0599.55 ± 0.0699.55 ± 0.0699.55 ± 0.06
LSTMBidirectional LSTM
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF97.10 ± 0.1797.10 ± 0.1797.10 ± 0.1797.03 ± 0.2297.03 ± 0.2297.03 ± 0.22
DocEmbWord2VecCBOW96.93 ± 0.1696.94 ± 0.1596.93 ± 0.1696.88 ± 0.0996.88 ± 0.0996.88 ± 0.09
DocEmbWord2VecSG95.37 ± 0.1795.40 ± 0.1795.37 ± 0.1795.55 ± 0.3295.56 ± 0.3095.55 ± 0.32
DocEmbFastTextCBOW96.24 ± 0.3096.26 ± 0.2996.24 ± 0.3096.37 ± 0.2396.38 ± 0.2396.37 ± 0.23
DocEmbFastTextSG95.06 ± 0.1395.07 ± 0.1395.06 ± 0.1395.10 ± 0.3795.11 ± 0.3295.10 ± 0.37
DocEmbGloVe95.17 ± 0.3495.22 ± 0.2895.17 ± 0.3495.31 ± 0.4295.35 ± 0.3695.31 ± 0.42
DocEmbBERT98.86 ± 0.2498.87 ± 0.2298.86 ± 0.2498.89 ± 0.1798.89 ± 0.1698.89 ± 0.17
DocEmbRoBERTa80.25 ± 1.3881.31 ± 0.7380.25 ± 1.3880.17 ± 1.6681.30 ± 0.8680.17 ± 1.66
DocEmbBART99.62 ± 0.0599.62 ± 0.0599.62 ± 0.0599.65 ± 0.0599.65 ± 0.0599.65 ± 0.05
GRUBidirectional GRU
VectorizationAccuracyPrecisionRecallAccuracyPrecisionRecall
TFIDF97.00 ± 0.1397.00 ± 0.1397.00 ± 0.1396.85 ± 0.2096.86 ± 0.2096.85 ± 0.20
DocEmbWord2VecCBOW96.85 ± 0.1696.86 ± 0.1696.85 ± 0.1696.86 ± 0.1496.87 ± 0.1496.86 ± 0.14
DocEmbWord2VecSG95.29 ± 0.2895.31 ± 0.2695.29 ± 0.2895.53 ± 0.1495.56 ± 0.1495.53 ± 0.14
DocEmbFastTextCBOW96.28 ± 0.1796.29 ± 0.1796.28 ± 0.1796.25 ± 0.2896.26 ± 0.2696.25 ± 0.28
DocEmbFastTextSG94.72 ± 0.3194.75 ± 0.2894.72 ± 0.3194.77 ± 0.5294.85 ± 0.4494.77 ± 0.52
DocEmbGloVe95.06 ± 0.3495.10 ± 0.3195.06 ± 0.3495.09 ± 0.2995.15 ± 0.2595.09 ± 0.29
DocEmbBERT98.91 ± 0.2798.92 ± 0.2598.91 ± 0.2798.99 ± 0.1098.99 ± 0.1098.99 ± 0.10
DocEmbRoBERTa80.44 ± 1.2681.44 ± 0.5580.44 ± 1.2680.04 ± 1.6981.36 ± 0.7680.04 ± 1.69
DocEmbBART99.62 ± 0.0999.62 ± 0.0999.62 ± 0.0999.64 ± 0.0799.64 ± 0.0799.64 ± 0.07
AccuracyPrecisionRecall
MisRoBÆRTa [23]99.52 ± 0.1299.52 ± 0.1299.52 ± 0.12
Accuracy
Proppy [70]98.36
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Truică, C.-O.; Apostol, E.-S. It’s All in the Embedding! Fake News Detection Using Document Embeddings. Mathematics 2023, 11, 508. https://doi.org/10.3390/math11030508

AMA Style

Truică C-O, Apostol E-S. It’s All in the Embedding! Fake News Detection Using Document Embeddings. Mathematics. 2023; 11(3):508. https://doi.org/10.3390/math11030508

Chicago/Turabian Style

Truică, Ciprian-Octavian, and Elena-Simona Apostol. 2023. "It’s All in the Embedding! Fake News Detection Using Document Embeddings" Mathematics 11, no. 3: 508. https://doi.org/10.3390/math11030508

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop