Next Article in Journal
A Theoretical Analysis of a Fractional Multi-Dimensional System of Boundary Value Problems on the Methylpropane Graph via Fixed Point Technique
Next Article in Special Issue
Incorporating Phrases in Latent Query Reformulation for Multi-Hop Question Answering
Previous Article in Journal
Robust Stability Analysis of Unstable Second Order Plus Time-Delay (SOPTD) Plant by Fractional-Order Proportional Integral (FOPI) Controllers
Previous Article in Special Issue
Gulf Countries’ Citizens’ Acceptance of COVID-19 Vaccines—A Machine Learning Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MisRoBÆRTa: Transformers versus Misinformation

by
Ciprian-Octavian Truică
1,2,*,† and
Elena-Simona Apostol
2,*,†
1
InfoLab, Department of Information Technology, Uppsala University, SE-751 05 Uppsala, Sweden
2
Computer Science and Engineering Department, Faculty of Automatic Control and Computers, University Politehnica of Bucharest, RO-060042 Bucharest, Romania
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Submission received: 17 January 2022 / Revised: 5 February 2022 / Accepted: 8 February 2022 / Published: 12 February 2022

Abstract

:
Misinformation is considered a threat to our democratic values and principles. The spread of such content on social media polarizes society and undermines public discourse by distorting public perceptions and generating social unrest while lacking the rigor of traditional journalism. Transformers and transfer learning proved to be state-of-the-art methods for multiple well-known natural language processing tasks. In this paper, we propose MisRoBÆRTa, a novel transformer-based deep neural ensemble architecture for misinformation detection. MisRoBÆRTa takes advantage of two state-of-the art transformers, i.e., BART and RoBERTa, to improve the performance of discriminating between real news and different types of fake news. We also benchmarked and evaluated the performances of multiple transformers on the task of misinformation detection. For training and testing, we used a large real-world news articles dataset (i.e., 100,000 records) labeled with 10 classes, thus addressing two shortcomings in the current research: (1) increasing the size of the dataset from small to large, and (2) moving the focus of fake news detection from binary classification to multi-class classification. For this dataset, we manually verified the content of the news articles to ensure that they were correctly labeled. The experimental results show that the accuracy of transformers on the misinformation detection problem was significantly influenced by the method employed to learn the context, dataset size, and vocabulary dimension. We observe empirically that the best accuracy performance among the classification models that use only one transformer is obtained by BART, while DistilRoBERTa obtains the best accuracy in the least amount of time required for fine-tuning and training. However, the proposed MisRoBÆRTa outperforms the other transformer models in the task of misinformation detection. To arrive at this conclusion, we performed ample ablation and sensitivity testing with MisRoBÆRTa on two datasets.

1. Introduction

The digital age has seen the emergence of new mass media paradigms for information distribution, substantially different from classical mass media. With new digital-enabled mass media, the communication process is centered around the user, while multimedia content is the new identity of news. Thus, the media landscape has shifted from mass media to personalized social media with these new paradigms. Along with the advantages this progress brings, it aggravates the risk of fake news [1], potentially with detrimental consequences to society, by facilitating the spread of misinformation in the form of propaganda, conspiracy theories, political bias, etc.
Misinformation consists of news articles that are intentionally and verifiably false or inaccurate and which could mislead readers by presenting alleged, imaginary, or false facts about social, economic, and political subjects of interest [2,3]. However, current trends and the accessibility of technology render this proneness even more potentially damaging than it has been historically, having dire consequences to the community and society at large [4]. Its spread influences communities and creates public polarization regarding elections [5], medical practices [6], etc. Furthermore, the practices of misinforming society are been condemned by politicians and businessmen alike, being considered real threats that undermine democracy and public health [7]; we believe that tools and methods for accurately detecting misinformation are of high relevance to the research community and society at large.
The main motivation for this paper is to evaluate the performance of different transformers for the task of misinformation detection in order to propose a novel deep learning transformer-based model, i.e., MisRoBÆRTa. We used a large real-world news dataset consisting of 100,000 news articles, labeled using 10 different classes, i.e., fake news, satire, extreme bias, hate news, etc. By employing such a dataset, we also address two shortcomings in the current research:
(1
Increasing the size of the dataset from small to large; and
(2
Moving the focus of fake news detection from binary to multi-class classification.
The experimental results show that the accuracy of transformers on the multi-class classification task is not influenced by the size of the network, but rather by the method employed to learn the context, and the size of both the dataset and vocabulary. We also include in our benchmark the state-of-the-art FakeBERT model [8].
The research questions for our benchmark are as follows:
(Q1
Which transformer obtains the overall best accuracy for the task of multi-class classification of misinformation?
(Q2
Which transformer obtains the overall best runtime performance without the model having a significant decrease in accuracy?
(Q3
Can there be a balance between accuracy and runtime required for fine-tuning and training?
Our contributions are as follows:
(1
We propose MisRoBÆRTa, a new transformer-based deep learning architecture for misinformation detection;
(2
We conducted an in-depth analysis of the current state-of-the-art deep learning, word embeddings, and transformer-based models for the task of fake news and misinformation detection;
(3
We propose a new balanced dataset for misinformation containing 100,000 articles extracted from the FakeNewsCorpus, where, for each news article, we (1) manually verified its content to make sure that it was correctly labeled; (2)verified the URLs to point to the correct article by matching the titles and authors;
(4
We performed an in-depth analysis of the proposed dataset;
(5
We extended the SimpleTransformers package with a multi-class implementation for BART;
(6
We performed a detailed benchmark analysis using multiple transformers and transfer learning on the task of misinformation and compared the results with the state-of-the-art model FakeBERT [8].
This paper is structured as follows. Section 2 discusses the current research and methods related to misinformation and fake news detection. Section 3 introduces the transformers employed in this study as well as the neural network used for classification. Section 5 presents the datasets and analyzes the experimental results. Section 6 discusses our findings and hints at current challenges that we identify for the task of misinformation detection. Section 7 presents the conclusions and outlines future directions.

2. Related Work

The misinformation and fake news detection problem has been tackled in the literature using different perspectives, from the linguistic approach to more sophisticated natural language processing (NLP) models. Whereas, the linguistic approach attempts to learn language cues that are hard to monitor (e.g., frequencies and patterns of pronouns, conjunctions, and negative emotion word usage [9]), the NLP approaches build models using content-based features for fake news detection [10].
In article [11], a deep convolutional-based GloVe-enabled model (FNDNet) is used for fake news detection. Although the experimental setup for the tested networks use small epoch numbers, FNDNet shows promising results compared to other deep learning architectures, i.e., classical convolutional neural network (CNN) and long short-term memory (LSTM), both trained using GloVe [12]. Another CNN-based architecture that show promising results is MCNN-TFW (multiple-level convolutional neural network-based) [13]. MCNN-TFW uses semantics features that are determined using a pre-trained Word2Vec [14].
Attention-based models are also good candidates for fake news detection. In [15], a new hierarchical attention network is used for learning GloVe embeddings for different latent aspects of news articles. The solution uses attention weights to support or refute the analyzed claims. The authors emphasize the importance of aspect embeddings for false news prediction. In Trueman et al. [16], the authors present an attention-based convolutional bidirectional LSTM approach to detect fake news. The news articles are classified using six multilevel categories ranging from “true” to “pants on fire”. The authors use the LIAR dataset consisting of 12,836 statements for training, validation, and testing.
Other current approaches use advanced pre-trained model-based detection, i.e., transformers [17,18,19,20], or new vectorization techniques, i.e., CLDF [21]. Among the transformers, bidirectional encoder representations from transformers (BERT) (Section 3) has been used the most for fake news detection.
Liu et al. [18] proposed a two-stage BERT-based model for fake news detection. Unlike simple BERT, the authors utilized all hidden states and applied attention mechanisms to calculate weights. The experiments were done using the LIAR dataset containing 12,836 short news statements. Being a relatively small dataset with little information, the proposed model and the compared ones have trouble learning correctly. Furthermore, current research proves that BERT does not provide accurate models if it is fine-tuned with a small dataset or short statements [22].
A comparison in performance between BERT and a deep-learning architecture using CNN and BiLSTM layers is presented in [19]. This article also offers an analysis of the attention coverage of texts for the two models. Although two small (∼1000 articles) subsets from the FakeNewsCorpus dataset are used, the classification is not multi-class, as the authors transformed the 11 categories into 2 classes, i.e., real and fake. The results show that, on the two small datasets, BiLSTM-CNN architecture outperforms or has the same accuracy as BERT; thus, strengthening the statement from [22].
In [23], a BERT-RNN architecture uses a relatively big balanced dataset and obtains very good results on a dataset labeled with different levels of veracity i.e., different degrees of truth and falsity. This solution has a hybrid architecture connecting BERT word embeddings with a recurrent neural network (RNN). The RNN network was used to obtain document embeddings. The major issue with this solution is that the results are not explained or discussed, and there is no comparison with other embedding models or deep neural networks.
Another hybrid BERT-based solution is FakeBERT [8]. FakeBERT uses the vectors generated after word-embedding from BERT as input to several 1D-CNN layers. The authors compared their solution with other deep learning solutions, considering a dataset from Kaggle with binary labels. The same dataset is used in FNDNet [11]. By analyzing these two articles written by the same main authors, we can conclude that FakeBERT has slightly better accuracy than FNDNet. Thus, we decided to add FakeBERT to our benchmark.
Other transformers did not get as much attention as BERT, but still, there are some articles that tackle this subject. A weak social supervision RoBERTa-based solution to detect fake news from limited annotated data is proposed in [24]. RoBERTa is an optimized method that improves BERT (Section 3). Their model outperforms other CNN or adversarial neural network-based solutions. In other papers, we found fake news detection solutions based on XLNet, an extension of BERT. In Gautam et al. [25], the authors combined topic information with an XLNet model. The authors compared their solution (XLNet + LDA) with basic XLNet, and also with BERT on a relatively small annotated dataset (∼10 K). To achieve an increase in time performance, some solutions use compact language models, such as a lite BERT (ALBERT) [26]. FakeFinder [27] is an ALBERT-based solution for mobile devices that offers real-time fake news detection from Twitter. This small ALBERT-based model achieves comparable F1 scores with a BERT model, although slightly smaller, when tested on a dataset with ∼2500 tweets and ∼117,000 comments.
To the best of our knowledge, there are almost no benchmarks on misinformation detection using transformers, a multi-class task. The only (found) article that presents a more in-depth benchmark on this subject is [28]. The authors only considered binary classification for five pre-trained models (i.e., BERT, RoBERTa, DistilBERT, ELECTRA, ELMo [29]). The best performance was obtained by RoBERTa, but it is not clear if they used fine-tuning.
After the state-of-the-art analysis, we noticed the following: none of these methods attempted to detect different types of misinformation, e.g., clickbait, extreme bias, but rather focused on the veracity of the articles, binary or multilevel [23], i.e., different degrees of truth and falsity. Moreover, the majority of the solutions used relatively small datasets.

3. Methodology

For our performance evaluation, we used multiple, state-of-the-art transformers that are on multiple natural language processing tasks. For classification, we used a neural network containing the embedding layer and one hidden layer with a dropout layer between them.

3.1. Transformers

For our performance evaluation, we used multiple transformers that employed transfer learning to achieve state-of-the-art performance on multiple natural language processing tasks.

3.1.1. BERT

Bidirectional encoder representations from transformers (BERT) [30], is a deep bidirectional transformer architecture that employs transfer learning to generalize language understanding and can be fined-tuned on the dataset at hand. Classic language models treat textual data as unidirectional or bidirectional sequences of words. Instead, BERT use the deep bidirectional transformer architecture to learns contextual relations between the words (or sub-words). BERT reads the entire sequence of words at once using the transformer encoder. Thus, the model manages to learn the context of a word based on all of the surrounding words. The experimental results show that the language models built with BERT manage to better determine the language context than the models that treat textual data as sequences of words.

3.1.2. DistilBERT

DistilBERT (Distilled BERT) [31] is a method used to pre-train a smaller general-purpose language representation model by reducing the size of BERT. As BERT, DistilBERT can be fine-tuned with good performances to solve different natural language processing tasks. DistilBERT uses a triple loss combining language modeling, distillation, and cosine-distance losses. The experimental results prove that DistilBERT is a smaller, faster, and lighter model than BERT, which manages to retain the language understanding capabilities.

3.1.3. RoBERTa

A robustly optimized BERT pre-training approach (RoBERTa) [17] is a robustly optimized method that improves BERT’s language masking strategy by modifying key hyperparameters:
(1)
It removes the next-sentence pre-training objective; and
(2)
It increases both the mini-batches and learning rates.
Thus, these modifications improve RoBERTa’s masked language modeling objective, leading to better downstream task performance.

3.1.4. DistilRoBERTa

DistilRoBERTa (Distilled RoBERTa) [31,32], similar to DistilBERT, is a method used to pre-train a smaller general-purpose language representation model by reducing the size of RoBERTa. It uses the same methodology employed to distill BERT on the RoBERTa transformer.

3.1.5. ALBERT

A lite BERT (ALBERT) [26] is another optimized method that improves on BERT. It reduces the number of parameters in order to lower memory consumption and increase training speed. The authors of this method show through experimental validation that ALBERT leads to models that scale much better compared to the BERT. Furthermore, ALBERT uses a self-supervised loss that focuses on modeling inter-sentence coherence, which consistently improves downstream tasks with multi-sentence inputs.

3.1.6. DeBERTa

Decoding-enhanced BERT with disentangled attention (DeBERTa) [33] is an improvement of BERT and RoBERTa models. It uses the disentangled attention mechanism, which employs two vectors to represent each word. The first vectors encodes the word’s content, while the second it’s position. Thus, the attention weights among words are computed using disentangled matrices on their contents and relative positions. Moreover, DeBERTa uses an enhanced mask decoder to replace the output softmax layer to predict the masked tokens for model pre-training.

3.1.7. XLNet

XLNet [34] is an extension of BERT that integrates ideas from the transformer-XL model [35]. It uses an autoregressive method to learn bidirectional contexts by maximizing the expected likelihood over all permutations of the input sequence factorization order. Thus, XLNet overcomes the limitations of BERT through the autoregressive formulation.

3.1.8. ELECTRA

Efficiently learning an encoder that classifies token replacements accurately (ELECTRA) [36] is a pre-training approach used to build two transformer models: the generator and the discriminator. The generator replaces tokens in a sequence and trains as a masked language model. The discriminator predicts which tokens are replaced by the generator in the sequence. We used ELECTRA’s discriminator for classification.

3.1.9. XLM

The cross-lingual language model (XLM) [37] is a transformer architecture that employs, during pre-training, either an unsupervised modeling technique that only relies on monolingual data, i.e., casual language modeling (CLM) or masked language modeling (MLM), or a supervised modeling technique that leverages parallel data with a new cross-lingual language model objective, i.e., MLM used in combination with translation language modeling (TLM). CLM models the probability of a word given the previous words in a sentence, while MLM uses the masked language modeling of BERT. TLM is an extension of MLM that concatenates parallel sentences instead of considering monolingual text streams. This model manages to obtain better results than the baseline on multiple tasks, such as cross-lingual classification, unsupervised machine translation, supervised machine translation, low-resource language modeling, and unsupervised cross-lingual word embeddings.

3.1.10. XLM-RoBERTa

XLM-RoBERTa [38] is a transformer-based pre-trained multilingual masked language model. The novelty of this model is that it does not require language tensors to understand which language is used. Thus, the model is able to determine the correct language just from the input IDs.

3.1.11. BART

The bidirectional and autoregressive transformer (BART) [39] uses a standard transformer-based neural machine translation architecture, which can be seen as generalizing BERT. Thus, BART uses a standard sequence-to-sequence for machine translation architecture with a bidirectional encoder, similar to BERT, and a left-to-right decoder, similar to GPT. BART is trained by corrupting text with an arbitrary noising function, and learning a model to reconstruct the original text. During the pre-training task, the model randomly shuffles the order of the original sentences. Moreover, the spans of text are replaced with a single mask token using a novel in-filling scheme. BART is particularly effective when fine-tuned for text generation but it also works well for comprehension tasks [39].

3.2. Classification Network

The network for classification extracts the last hidden layer of the transformer, except for Electra, which uses the discriminator hidden states. All networks use a final dense layer containing perceptrons for classification. Between the transformer model and the classification layer, we used a dropout layer to prevent overfitting. The number of perceptrons in the classification layer is equal to the number of classes. This final hidden layer, added after the transformer model and dropout layer, outputs the label for the multi-class classification. DistilBERT and DistilRoBERTa also use a pre-classification, a ReLU, and a dropout layer between the output of the transformer model and the classification layer. The pre-classification layer size is equal to the dimension of the encoder layers and the pooler layer. The pre-classification layer is used to stabilize the output of the transformer and archive an improved classification performance [40]. For XLM, XLM-RoBERTa, XLNet models, we used a sequence summary layer, which takes a sequence’s hidden states and computes a single vector summary, followed by a dense layer for classification.

4. MisRoBÆRTa: Misinformation RoBERTa-BART Ensemble Detection Model

In this section, we propose a new deep learning architecture that develops hierarchical representations of learning using multiple bidirectional long short-term memory (BiLSTM) and convolutional layers (CNN). Figure 1 presents the computational workflow of our novel architecture MisRoBÆRTa: misinformation using RoBERTa-Bart sentence embeddings. In our model, we have two inputs:
(1
The sentence embeddings constructed with RoBERTa base of size 768; and
(2
The sentence embeddings extracted with BART, a large of size 1024.
We chose these two transformers for MisRoBÆRTa because they had the best overall results among the pretrained transformers (Section 5) when classification was done using the simple neural network model (Section 3). Each embedding was then the input to a new block (branch) formed using the following layers: BiLSTM, reshape, CNN, max pooling, and flatten. During the ablation testing presented in Section 5.5 we obtained better results when employing two BiLSTM layers for the BART branch. The output of the flatten layers was then concatenated, reshaped, and sent to another block (ensemble branch) containing the same layers that had, as input, the RoBERTa embeddings (RoBERTa branch). Finally, there was a dense layer for the output of the classification.

4.1. MisRoBÆRTa Components

In this subsection, we present MisRoBÆRTa’s main deep learning layers: dense, BiLSTM, and CNN. We will not present again RoBERTa and BART, as both transformers are presented in Section 3.1.3 and Section 3.1.11, respectively.

4.1.1. Long Short-Term Memory Networks

The long short-term memory network (LSTM) [41] was developed to mitigate the vanishing or the exploding gradient problems that the classical recurrent neural network encountered. By introducing gates within the cell, the LSTM manages to better capture dependencies. The LSTM model uses the input gate  i t for deciding what relevant information can be added from the current step t, using the previous hidden state h t and the input x t , forget gate f t for deciding which information is used and which is ignored in the current step t, and output gate o t determines the value of the next hidden state h t . Equation (1) presents of each gate (i.e., f t , i t , o t ) the current hidden state ( h t 1 ) and current memory state ( c t ), where:
(1)
x t is the input at time step t;
(2)
h t is the output, or next hidden state;
(3)
h t 1 is the previous hidden state;
(4)
c ˜ t is the cell input activation vector;
(5)
c t is the current memory state;
(6)
c t 1 is the previous memory state;
(7)
W f , W c , W i , and W o are the weights for each gate’s current input;
(8)
U f , U c , U i , and U o are the weights for each gate’s previous hidden state;
(9)
b f , b c , b i , and b o are the bias vectors;
(10)
σ s is the sigmoid activation function;
(11)
σ h is the hyperbolic tangent activation function;
(12)
⊙ operator is the Hadamard product, i.e., the element-wise multiplication function.
i t = σ s ( W i x t + U i h t 1 + b i ) f t = σ s ( W f x t + U f h t 1 + b f ) o t = σ s ( W o x t + U o h t 1 + b t ) c ˜ t = σ h ( W c x t + U c h t 1 + b c ) c t = f t c t 1 + i t c ˜ t h t = o t tanh ( c t )
The LSTM processes sequences and it is able to capture past information. To also consider future information when building the model, the bidirectional LSTM (BiLSTM) uses two LSTM with two hidden states h t and h t (Equation (2)). The first hidden state h t processes the input in a forward manner using the past information provided by the forward LSTM L S T M F ). The second hidden state h t processes the input in a backwards manner using the future information provided by the backward LSTM ( L S T M B ).
h t = L S T M F ( x t ) h t = L S T M B ( x t )
Finally, the hidden states h t and h t are concatenated at every time-step to enable encoding the information from past and future contexts into one hidden state h t (Equation (3)).
h t = [ h t | | h t ]

4.1.2. Convolutional Neural Networks

Unlike recurrent neural networks, which are used to detect patterns in sequences, convolutional neural networks (CNNs) were developed to detect patterns in hyperspaces. In text classification, one-dimensional CNNs are used to apply a window of convolution that can extract n-grams. Equation (4) presents a convolution unit r i , where x i is a vector of filtered words, W is a weight matrix, b is the bias, and σ is a non-linear activation function.
r i = σ ( W · x i + b )

4.1.3. Max Pooling

The max pooling mechanism captures the most useful features produced by the previous layer. The mechanism involves sliding a two-dimensional filter over each channel of the feature map, summarizing the features lying within the region covered by the filter. This layer applies an affine transformation to the feature vector to produce logits d. Equation (5) presents the logits d outputted when using a feature map with the dimensions n h × n w × n c , where n h , n w , n c are the height, the width, and the the number of channels of the feature map, respectively, f is the size of the filter, and s is the stride length.
d = ( n h f + 1 ) s × ( n w f + 1 ) s × n c

4.1.4. Dense Layer

The dense layer uses perceptron units to classify the output from the previous layer. This layer received the information from the previous layer. Equation (6) presents the representation of the perceptron predicted label y ^ , where x i is the input, w is a weight vector, b represents the bias, and δ ( · ) is the activation function. The activation function takes the total input and produces an output for the unit given some threshold. In our model, we use the softmax activation function δ ( x ) = e x i = 1 m e x i with m the dimension of vector x.
y i ^ = δ ( w · x i + b )
A loss function (Equation (7)) is used at each iteration to adjust the weights to align the predicted label y ^ to the true label y.
L ( y ^ , y ) = ( y log y ^ + ( 1 y ) log ( 1 y ^ ) )

4.2. MisRoBÆRTa Description

The BiLSTM layers contain 256 units each, 128 units for each long short-term memory (LSTM) and a dropout rate of 0.2. We used the hyperbolic tangent as activation functions and the sigmoid for the recurrent activation function. The CNN layers contain 64 filters, a kernel size of 128, and ReLU as the activation function. After each CNN, we used a max pooling layer to down-sample the output and reduce the number. Flatten was used to transform the two-dimensional output of the max pooling layer into a one-dimensional tensor. The reshape layers were used to reshape the tensors to be accepted as input for the next layers. Between the BiLSTM and the CNN, we needed a two-dimensional tensor, while between the concatenate layer and the BiLSTM, we needed a three-dimensional tensor. The dense layer was a fully connected layer of perceptrons used for classification. All parameters were determined using hyperparameter tuning.

4.3. MisRoBÆRTa Architectural Choices

Unidirectional LSTM preserves input information that has passed through its hidden state. Thus, it only uses passed information to make assessments and correct predictions. A BiLSTM instead preserves input information in two ways. As LSTM, the BiLSTM will preserve the information that passes through the hidden layer, i.e., past to future. Unlike LSTM, the BiLSTM also preserves information using the output of the hidden layer, and running backwards the preserved information, i.e., future to past. Thus, BiLSTM combines two hidden states to preserve information from both past and future. BiLSTM is better suited than LSTM to understand context and preserved context. This is also supported by the ablation testing we present in Section 5.5 Furthermore, we observed experimentally that we obtained better results when using two BiLSTM layers instead of one for the BART branch.
After the last BiLSTM layer for each branch, we employed a CNN layer. CNNs have been used successfully for different text classifications and natural language tasks [42]. When using a CNN, the output of each convolution detects patterns of multiple sizes by varying the kernel shapes. These outputs are then concatenated to permit the detection of multiple size patterns of adjacent words. These patterns could be seen as multi-word expressions. Therefore, CNNs can accurately identify and group multi-word expressions in the sentence regardless of their position. These multi-word expressions better represent the context and the representative features of the input text.

5. Experimental Results

5.1. Dataset

The dataset used for experiments was the open-source FakeNewsCorpus (accessed on 6 August 2021) dataset available on GitHub [43]. It is comprised of more than 9.4 million labeled English news articles that contain both textual content (e.g., title, content) and metadata (e.g., publisher, authors). The original dataset is not manually filtered; therefore, some of the labels might not be correct. However, the authors argue that because the corpus is intended to be used in training machine learning algorithms, this shortcoming should not pose a practical issue. Moreover, this issue should help the models better generalize and remove overfitting. For our experiments, we used only the news content and their labels. Moreover, we made sure that the news sources contained articles that belonged to the type given by the class and we manually annotated each news article content and compared it with its original label. The annotation of the textual content was done by computer science students. We used two annotators for each article, introducing a third if there was no consensus between the first two. The annotation task had two main limitations: author bias [44,45] and incomplete annotations [46]. Annotator bias refers to the habits and backgrounds of annotators that label the data and how these can impact the quality of the annotation and, as a result, the model. To mitigate this, we presented the annotators with 10 classes from which to pick only one. Thus the annotator, cannot bring any of his habits within the annotation process. Incomplete annotations refer to the fact that the annotators may not agree among them. In this situation, we added a third annotator who could choose only one of the two labels selected by the previous two annotators. In this way, we mitigated disagreements among the annotators. At the end of this task, we did not find discrepancies between the original labels and the ones determined by the annotators.
Furthermore, we checked that the selected article URLs were pointing to the correct article by matching the titles and authors. We did not apply preprocessing. The final dataset contained 10 classes (currently the 11th class contains 0 articles), with each class represented by 10,000 documents selected at random (Table 1). We chose this dataset because it is large enough to use with transformers. We also presented the ablation tests for the Kaggle Fake News dataset (accessed on 9 October 2021) in Section 5.5 used in [8,11].
Table 2 presents the dataset’s description and token statistics (#Tokens). We observe that the distribution of tokens among the different classes is almost equal. The satire class has the shortest textual content on average, while Hate News has the largest mean number of tokens. We also note that the entire dataset contains 1,337,708 unique tokens and a total of 59,726,182 tokens. By analyzing the dataset, we can affirm that there is no real bias added to the classification problem by the length of the documents.
To better understand the token distribution among the different classes, we analyzed documents using unigrams. Table 3 presents the top-10 unigrams extracted using the NLTK (accessed on 10 October 2021) python package [47]. We extracted the top-10 unigrams for the entire corpus as well as the top-10 unigrams for each class. We used the cosine similarity, based on BERT and FastText, to measure the similarity between the unigrams extracted for the entire corpus and each class. In our implementation, we used PolyFuzz (accessed on 14 October 2021) [48]. We observed that the smallest similarity was obtained when comparing the unigrams for the entire corpus with the ones extracted for the satire class, i.e., 0.63 for BERT similarity and 0.44 for FastText similarity. The rest of results are above 0.76 for BERT similarity and 0.67 for FastText similarity. These results show that the terms appearing in each class are well represented within the entire corpus, although for the satire class, are underrepresented. Based on the unigram analysis, we can conclude the there is no real bias added to classes by specific terms, as the distribution of the most frequent terms was similar.
To better understand the context, we used a non-negative matrix factorization (NMF) [49] to extract the top-one topic for the entire dataset and each class. To extract the topics, we cleaned the corpus removing stopwords and punctuation, then we vectorized the documents using the TFIDF. For our implementation, we used the TFIDF and NMF models from the scikit-learn (accessed on 6 October 2021) [50] library. This analysis was performed in order to determine if the underline hidden context present for each class would influence the behavior of the transformers during fine-tuning. We observed that the best similarity was obtained by the topic extracted for the satire class (Table 4). The other classes had BERT-based similarities over 0.63. These results show that the majority of articles contain context that is present among all classes, which reduces the bias added by the frequent topic and hidden context, while preserving some features that can differentiate between the articles.
Based on the analysis we performed using token statistics, unigram, and topic modeling, we can conclude the following for the dataset. Firstly, the length of the documents and the number of tokens for each class did not add any bias for the models to discriminate based on these two dimensions. Secondly, the unique tokens per class versus the unique tokens for the entire dataset shows that some tokens were representative for some classes. Thirdly, the most frequent unigrams did not influence the models, as the similarity between the unigrams extracted for each class and for the entire corpus was high. Finally, the global context for each class had a moderate similarity with the context for the entire corpus. These results show that, among the different classes, there were some clear discriminative features.

5.2. Fine-Tuning

We used the SimpleTransformers (accessed on 12 October 2021) [51] library and the pre-trained transformers from HuggingFace (accessed on 9 October 2021) [52]. We extended SimpleTransformers with our own implementation for a BART-based multi-class classifier. Table 5 presents the transformers employed for the experiments. The ALBERT v2 models did not have dropout. We used the ELECTRA’s discriminator for classification.
The employed transformers were fine-tuned by training them on our dataset for 10 epochs with an early stopping mechanism that evaluated the loss over 5 epochs. We split the dataset using a 64%–16%–20% training–validation–testing ratio with random seeding. For each split, we maintained the stratification of the labels. We applied this method for 10 rounds of tests.

5.3. Classification Results

We used an NVIDIA® DGX Station containing 4 NVIDIA® V100 Tensor Core GPUs for our experiments. Table 6 presents the classification results. We observed that all transformer models had an accuracy of over 85%, with the lowest accuracy obtained by ALBERT base v2. For comparison reasons, we also tested the state-of-the-art model FakeBERT [8]. As the authors of this model did not provide their implementation, we implemented FakeBERT using BERT base together with the TensorFlow (accessed on 8 October 2021) [53] and Keras (accessed on 8 October 2021) [54] libraries. The code for the MisRoBÆRTa, transformer implementation, and FakeBERT are publicly available on GitHub at https://github.com/cipriantruica/MisRoBAERTa_Transformers-vs-misinformation (accessed on 8 October 2021).
The BERT, RoBERTa, DeBERTa, XLNet, ELECTRA, and XLM-RoBERTa base models obtained better results than their counterpart large models. For these models, we can conclude that a smaller number of encoder layers with smaller sizes better generalized and managed to correctly extract the hidden context within the texts.
The reverse happened for ALBERT and BART, where the large models outperformed the base models. Although the difference in performance between BART base and BART large was small (∼0.60%) for ALBERT, the difference was between ∼1.30% and ∼2.3%. For BART, we can conclude that the autoregressive method, the sequence-to-sequence architecture, and the left–right decoder played an important factor and improved performance as the model’s number of layers and size increased. For ALBERT, the self-supervised loss that focused on modeling inter-sentence coherence improved the model’s performance as the network size increased. Furthermore, the use of dropout improved ALBERT performance significantly, as ALBERT v2 versions had the overall worst performance.
We observes that DeBERTa obtained slightly better results than BERT, regardless if it was the base or the large model. DeBERTa base obtained an accuracy slightly worst than the RoBERTa base, while DeBERTa large outperformed RoBERTa large. XLM obtained the second worst accuracy among the tested transformers.
Both DistilBERT and DistilRoBERTa were outperformed by BERT, respectively, RoBERTa base models. The difference in accuracy between the base and distilled models was very small: ∼0.04% between DistilBERT and BERT base and ∼0.70% between DistilRoBERTa and RoBERTa base. Although DistilBERT outperformed BERT large with a small increase in accuracy (∼0.24%), DistilRoBERTa outperformed RoBERTa large with a large increase accuracy (∼2.72%). We can conclude that, for the use case of misinformation classification, distillation is not a good approach if the computational and memory resources needed for the base approaches are available.
The micro and macro recall metrics are the most relevant for misinformation classification, as they show the correctly classified news articles relative to all the news articles, regardless of the predicted label. The micro and macro precisions manage to show the percentage of the articles, which are relevant. All transformers obtained results over 85% for these four metrics, meaning that they managed to generalize well for the task of misinformation detection.
The overall best performing model was the novel MisRoBÆRTa with an accuracy of 92.50%, while the overall best performing transformer model was BART large, with an accuracy of 91.94%. The next three transformers that obtained accuracies over 91% were RoBERTa base (91.36%), BART base (91.35%), and DistilRoBERTa 91.32%.
We observed that the state-of-the-art-model FakeBERT had very low performance on our dataset. Although the original paper shows that FakeBERT obtained an accuracy of 98.9% on a binary dataset from Kaggle, in our multi-class dataset, it obtained only ∼70%. These results might be a consequence of the following:
(1
FakeBERT is developed for binary classification, while we used it for multi-class classification;
(2
To change FakeBERT from a binary classifier to a multi-class classifier, we modified the number of units of its final DENSE layer from 2 to the number of classes, i.e., 10;
(3
We only trained FakeBERT for 10 epochs, to be consistent in our performance experiments (note: this is also the number of epochs used in the original paper).
MisRoBÆRTa had the overall best classification results with an accuracy of 92.50%. This accuracy was a direct result of the way context was preserved through the use of:
(1
BART large and RoBERTa base;
(2
BiLSTM layers that combined two hidden states to preserve information from both past and future;
(3
CNNs to generate multi-word expressions, which better represent the feature space.

5.4. Runtime Evaluation

To evaluate the runtime, we ran each transformer for 10 epochs without early stopping. As already mentioned, we used an NVIDIA® DGX Station containing 4 NVIDIA® V100 Tensor Core GPUs for our evaluation. Figure 2 presents the mean execution time and the standard deviation for 10 rounds of tests. As expected, the large models are much slower than their base counterparts. Moreover, we note that the distilled versions DistilRoBERTa, ALBERT base, and DistilBERT outperformed both BERT and RoBERTa. MisRoBÆRTa had the fastest runtime among the tested models managing to converge in 1.83 h (this included the maximum time taken to create the sentence embeddings in parallel using the two transformers: 1.53 h for BART and 0.35 h for RoBERTa).
For the base models, the fastest transformer model was DistilRoBERTa, which required, on average, 1.93 h for fine-tuning and training, while the slowest was XLM-RoBERTa, which finished training and fine-tuning in a little less than 7 h on average. For the large models, RoBERTa large was the fastest, requiring on average 6.55 h for fine-tuning and training, while the slowest was ALBERT large v2, which finished in 11.14 h on average. FakeBERT converged in 2.59 h on average. This result was directly impacted by the CNN and max pooling layers it employed to minimize the input space. XLM runtime performance at 5.16 h on average was between BERT base and XLNet base. In conclusion, the fastest transformer model that also achieved a high accuracy was DistilRoBERTa.

5.5. Ablation Testing

In Table 7 and Table 8, we present ablation and sensitivity testing for MisRoBÆRTa using the publicly available Kaggle Fake News Dataset and the FakeNewsCorpus dataset presented in Section 5. The tests include (1) testing with LSTM and BiLSTM units, (2) variations on the number of layers for each branch, e.g., 1, 2, and 3 layers, and (3) variation in the number of units for each LSTM layer, e.g., 32, 64, 128. The results are aggregated after 10 runs, for each execution, the dataset was split at random without replacement, but keeping the class ratio into 64% training, 16% validation, and 20% testing. We present only the network’s training time, the sentence embedding creation time was not taken into account. We should note that the results we obtained during ablation and sensitivity testing on the publicly available Kaggle Fake News dataset are similar to the current state-of-the-art results obtained on this corpus by [8,11].

6. Discussion

The misinformation detection problem is a complex task that requires large datasets to accurately determine news content veracity. We found that in the current literature, many authors utilize relatively small datasets for this task, e.g., datasets with ∼1000 articles in [19], ∼10,000 in [25], which is negatively reflected in the accuracy obtained by the tested models, especially when it comes to BERT-based models. Surely, there are authors who test their solutions using larger corpora, e.g., ∼80,000 articles in [28]. In some cases, the dataset consists of short news statements, which in turn affect the accuracy considerably [18]. Furthermore, the current research articles only address the misinformation detection problem using either a binary approach (the news article’s veracity is either true or false [8,19,28]) or levels of veracity [23]. Thus, none of the current approaches uses a large dataset that contains fake news labeled with different categories. We address this shortcoming by using a large dataset containing 100,000 news articles labeled with 10 classes.
The size of the corpus, number of tokens, the used vocabulary, the frequent unigrams, and topics play an important role in the overall and per class detection task. Through fine-tuning, we include these dimensions in our models, thus improving the classification performance.
In the misinformation detection task, contextual information plays an important role that impacts the accuracy of determining hidden patterns within the text. Thus, using transformers and transfer learning, the models better discriminate between different types of misinformation as shown by the recall score of over 85%. By adding autoregression, sequence-to-sequence with a bidirectional encoder, and a left-right decoder, BART large achieves the overall best performance among the basic architecture with transformers.
For BERT, RoBERTa, DeBERTa, XLNet, ELECTRA, and XLM-RoBERTa, the base models manage to better generalize than the large ones, while for BART and ALBERT, the reverse happens. Thus, we can conclude that BERT, RoBERTa, DeBERTa, XLNet, ELECTRA, and XLM-RoBERTa manage to extract correctly the hidden context within the misinformation texts and better discriminate between the labels with a smaller network size and number of parameters. The autoregressive method to learn bidirectional contexts employed by XLNet and BART improves the misinformation detection performance as the network and number of parameters increases. We can conclude that, for XLNet base, although extracting less information from context than large, it achieves better results through the autoregressive method. Thus, both XLNet base and BART (base and large) extract more information from context using the autoregressive method.
XLM, a multilingual model, performs rather poorly. This is somehow expected as the best performing multilingual system scored below the worst performing native system [55].
Although distillation is a good technique to maintain classification accuracy by lowering resource consumption, DistilBERT and ALBERT models do not provide better results than the base models. As ALBERT also reduces BERT’s number of parameters, it still needs a large network to achieve good results. Furthermore, ALBERT performance significantly improves when dropout is used, which also lowers execution performance. These models can still be used in low resource environments, offering an adequate level of accuracy. We can conclude that by reducing both the number of parameters and the network size, the classification performance is decreased for these models.
DistilRoBERTa achieves the best accuracy (91.32%) among the distilled versions and the overall smallest runtime (1.93 h for fine-tuning and training). Its accuracy falls short with only 0.63% than the model with the highest accuracy, BART large. In conclusion, DistilRoBERTa model can be a viable option when the number of resources is limited, and both fine-tuning and training should be conducted with the least amount of time.
MisRoBÆRTa is a BiLSTM-CNN deep learning architecture that utilizes as input both RoBERTa and BART and sentence embeddings. This novel architecture manages to better preserve context than the transformer models, as it employs context-aware and future selection layers besides context-aware embeddings. MisRoBÆRTa outperforms the BART large model w.r.t. classification performance and DistilRoBERTa model w.r.t. execution time.
Finally, we would like to conclude the discussion of our results by answering one final question: is misinformation detection a solved problem? In short, no. Although the presented transformer-based methods and the novel architecture MisRoBÆRTa show their effectiveness by obtaining high accuracy using linguistic features, there remain a lot of challenges that, if tackled, will improve performance and help users make informed decisions about the content of their news feeds. Some of the challenges we identified while surveying the related work for the task of misinformation detection that require a shift from a model-centric perspective to a data-centric perspective are:
(1
Multi-modal (i.e., visual, audio, network propagation, and immunization, etc.) techniques are rarely used; this is due to a lack of complete, high quality, and well labeled corpora and the tedious work that is required to correctly annotate such datasets;
(2
Source verification is rarely taken into consideration for misinformation detection [56], this is due to the instability of links on the internet;
(3
Author credibility should be an important factor that is not really discussed or weighted in the feature selection for the models;
(4
The perpetual need to retrain or fine-tune models to capture the subtle shifts that appear over time in news articles that spread misinformation.

7. Conclusions

In this paper, we present a performance evaluation of transformers on the task of misinformation detection. We employed a large real-world dataset for our experiments and addressed two shortcomings in the current research: (1) Increasing the size of the dataset from small to large; and (2) Moving the focus of fake news detection from binary to multi-class classification. We present and discuss our findings in detail. Furthermore, the results show that the accuracy of transformers on the misinformation detection task is significantly influenced by the method employed to learn the context, dataset size, and vocabulary.
We propose MisRoBÆRTa, a new transformer-based deep learning architecture for misinformation detection. The proposed architecture used, as input, BART and RoBERTa sentence embeddings and a BiLSTM-CNN architecture for each embedding. The output of each BiLSTM-CNN was concatenated and passed through another BiLSTM-CNN to obtain the results. We conclude that the proposed architecture outperforms the other employed classification models w.r.t. (1) classification performance, obtaining an average accuracy of 92.50%; and (2) runtime, with an average execution time for training of 1.83 h.
We respond to the research questions as follows. The overall best accuracy among the transformer models is obtained by BART (91.94%), while DistilRoBERTa obtains the best accuracy (91.32%) in the least amount of time required for fine-tuning and training, i.e., 1.93 h on average. MisRoBÆRTa outperforms both of these models’ w.r.t. accuracies (92.50%) and performance times (1.83 h). We empirically observed that between BART base and large, the difference in performance is insignificant, while the difference in runtime is of almost 2.62 h on average. Moreover, some improved models of BERT slightly outperform the base model, i.e., DeBERTa and RoBERTa, while most fall short. Furthermore, multilingual models, such as XLM, do not obtain better accuracy than BERT base.
In future research, our solution can be enhanced with sentiment analysis techniques that will determine the polarity and strength of the sentiments expressed in each news article. Another research direction is to add global context through topic modeling [57,58].

Author Contributions

Conceptualization, C.-O.T. and E.-S.A.; methodology, C.-O.T. and E.-S.A.; software, C.-O.T. and E.-S.A.; validation, C.-O.T. and E.-S.A.; formal analysis, C.-O.T. and E.-S.A.; investigation, C.-O.T. and E.-S.A.; resources, C.-O.T. and E.-S.A.; data curation, C.-O.T. and E.-S.A.; writing—original draft preparation, C.-O.T. and E.-S.A.; writing—review and editing, C.-O.T. and E.-S.A.; visualization, C.-O.T. and E.-S.A.; supervision, C.-O.T. and E.-S.A.; project administration, C.-O.T. and E.-S.A.; funding acquisition, C.-O.T. and E.-S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original datasets used in this study are publicly available as follows: FakeNewsCorpus https://github.com/several27/FakeNewsCorpus (accessed on 9 October 2021); Kaggle Fake News Dataset https://www.kaggle.com/c/fake-news/data (accessed on 9 October 2021).

Acknowledgments

The publication of this paper is supported by the University Politehnica of Bucharest through the PubArt program and the work has been partly funded by EU CEF grant number 2394203 (NORDIS-NORdic observatory for digital media and information DISorders).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LSTMlong short-term memory
BiLSTMbidirectional long short-term memory
CNNconvolutional neural network
BERTbidirectional encoder representations from transformers
DistilBERTdistilled BERT
RoBERTarobustly optimized BERT pre-training approach
DistilRoBERTadistilled RoBERTa
ALBERTa lite BERT
DeBERTadecoding-enhanced BERT with disentangled attention
ELECTRAefficiently learning an encoder that classifies token replacements Accurately
XLMcross-lingual language model
BARTbidirectional and autoregressive transformer

References

  1. Ruths, D. The misinformation machine. Science 2019, 363, 348. [Google Scholar] [CrossRef] [PubMed]
  2. Shu, K.; Wang, S.; Le, T.; Lee, D.; Liu, H. Deep Headline Generation for Clickbait Detection. In Proceedings of the IEEE International Conference on Data Mining, Singapore, 17–20 November 2018. [Google Scholar] [CrossRef]
  3. Ilie, V.I.; Truica, C.O.; Apostol, E.S.; Paschke, A. Context-Aware Misinformation Detection: A Benchmark of Deep Learning Architectures Using Word Embeddings. IEEE Access 2021, 9, 162122–162146. [Google Scholar] [CrossRef]
  4. Zannettou, S.; Sirivianos, M.; Blackburn, J.; Kourtellis, N. The web of false information: Rumors, fake news, hoaxes, clickbait, and various other shenanigans. J. Data Inf. Qual. 2019, 11, 1–37. [Google Scholar] [CrossRef] [Green Version]
  5. Bovet, A.; Makse, H.A. Influence of fake news in Twitter during the 2016 US presidential election. Nat. Commun. 2019, 10, 7. [Google Scholar] [CrossRef]
  6. Marco-Franco, J.E.; Pita-Barros, P.; Vivas-Orts, D.; González-de-Julián, S.; Vivas-Consuelo, D. COVID-19, Fake News, and Vaccines: Should Regulation Be Implemented? Int. J. Environ. Res. Public Health 2021, 18, 744. [Google Scholar] [CrossRef]
  7. Commision, E. Fighting Disinformation. 2021. Available online: https://ec.europa.eu/info/live-work-travel-eu/health/coronavirus-response/fighting-disinformation_en (accessed on 6 September 2021).
  8. Kaliyar, R.K.; Goswami, A.; Narang, P. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimed. Tools Appl. 2021, 80, 11765–11788. [Google Scholar] [CrossRef]
  9. Conroy, N.K.; Rubin, V.L.; Chen, Y. Automatic deception detection: Methods for finding fake news. Proc. Assoc. Inf. Sci. Technol. 2015, 52, 1–4. [Google Scholar] [CrossRef] [Green Version]
  10. Gravanis, G.; Vakali, A.; Diamantaras, K.; Karadais, P. Behind the cues: A benchmarking study for fake news detection. Expert Syst. Appl. 2019, 128, 201–213. [Google Scholar] [CrossRef]
  11. Kaliyar, R.K.; Goswami, A.; Narang, P.; Sinha, S. FNDNet—A deep convolutional neural network for fake news detection. Cogn. Syst. Res. 2020, 61, 32–44. [Google Scholar] [CrossRef]
  12. Pennington, J.; Socher, R.; Manning, C. GloVe: Global Vectors for Word Representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar] [CrossRef]
  13. Li, Q.; Hu, Q.; Lu, Y.; Yang, Y.; Cheng, J. Multi-level word features based on CNN for fake news detection in cultural communication. Pers. Ubiquitous Comput. 2020, 24, 259–272. [Google Scholar] [CrossRef]
  14. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. In Proceedings of the International Conference on Learning Representations, Scottsdale, AZ, USA, 2–4 May 2013; pp. 1–12. [Google Scholar]
  15. Mishra, R.; Setty, V. Sadhan: Hierarchical attention networks to learn latent aspect embeddings for fake news detection. In Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval, Santa Clara, CA, USA, 2–5 October 2019; pp. 197–204. [Google Scholar] [CrossRef]
  16. Trueman, T.E.; Kumar, A.; Narayanasamy, P.; Vidya, J. Attention-based C-BiLSTM for fake news detection. Appl. Soft Comput. 2021, 110, 107600. [Google Scholar] [CrossRef]
  17. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  18. Liu, C.; Wu, X.; Yu, M.; Li, G.; Jiang, J.; Huang, W.; Lu, X. A Two-Stage Model Based on BERT for Short Fake News Detection. In Proceedings of the International Conference on Knowledge Science, Engineering and Management, Athens, Greece, 28–30 August 2019; pp. 172–183. [Google Scholar] [CrossRef]
  19. Kurasinski, L.; Mihailescu, R.C. Towards Machine Learning Explainability in Text Classification for Fake News Detection. In Proceedings of the IEEE International Conference on Machine Learning and Applications, Miami, FL, USA, 14–17 December 2020; pp. 775–781. [Google Scholar] [CrossRef]
  20. Jwa, H.; Oh, D.; Park, K.; Kang, J.; Lim, H. exBAKE: Automatic Fake News Detection Model Based on Bidirectional Encoder Representations from Transformers (BERT). Appl. Sci. 2019, 9, 4062. [Google Scholar] [CrossRef] [Green Version]
  21. Mersinias, M.; Afantenos, S.; Chalkiadakis, G. CLFD: A Novel Vectorization Technique and Its Application in Fake News Detection. In Proceedings of the Language Resources and Evaluation Conference, Marseille, France, 11–16 May 2020; pp. 3475–3483. [Google Scholar]
  22. Shaar, S.; Babulkov, N.; Da San Martino, G.; Nakov, P. That is a Known Lie: Detecting Previously Fact-Checked Claims. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 3607–3618. [Google Scholar] [CrossRef]
  23. Kula, S.; Choraś, M.; Kozik, R. Application of the BERT-Based Architecture in Fake News Detection. In Proceedings of the Conference on Complex Intelligent, and Software Intensive Systems, Asan, Korea, 1–3 July 2020; pp. 239–249. [Google Scholar] [CrossRef]
  24. Shu, K.; Zheng, G.; Li, Y.; Mukherjee, S.; Awadallah, A.H.; Ruston, S.; Liu, H. Leveraging multi-source weak social supervision for early detection of fake news. arXiv 2020, arXiv:2004.01732. [Google Scholar]
  25. Gautam, A.; Venktesh, V.; Masud, S. Fake News Detection System using XLNet model with Topic Distributions: CONSTRAINT@AAAI2021 Shared Task. In Proceedings of the CONSTRAINT Shared Task in AAAI-2021, Virtual Event, 8 February 2021; pp. 1–12. [Google Scholar] [CrossRef]
  26. Lan, Z.; Chen, M.; Goodman, S.; Gimpel, K.; Sharma, P.; Soricut, R. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020; pp. 1–17. [Google Scholar]
  27. Tian, L.; Zhang, X.; Peng, M. FakeFinder: Twitter Fake News Detection on Mobile. In Proceedings of the Companion Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 79–80. [Google Scholar] [CrossRef]
  28. Khan, J.Y.; Khondaker, M.T.I.; Afroz, S.; Uddin, G.; Iqbal, A. A benchmark study of machine learning models for online fake news detection. Mach. Learn. Appl. 2021, 4, 100032. [Google Scholar] [CrossRef]
  29. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep Contextualized Word Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers); Association for Computational Linguistics: New Orleans, LA, USA, 2018; pp. 2227–2237. [Google Scholar] [CrossRef] [Green Version]
  30. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  31. Sanh, V.; Debut, L.; Chaumond, J.; Wolf, T. DistilBERT, a distilled version of BERT: Smaller, faster, cheaper and lighter. In Proceedings of the Workshop on Energy Efficient Machine Learning and Cognitive Computing, Vancouver, BC, Canada, 13 December 2019; pp. 1–5. [Google Scholar]
  32. Sajjad, H.; Dalvi, F.; Durrani, N.; Nakov, P. Poor Man’s BERT: Smaller and Faster Transformer Models. arXiv 2020, arXiv:2004.03844. [Google Scholar]
  33. He, P.; Liu, X.; Gao, J.; Chen, W. DeBERTa: Decoding-enhanced BERT with Disentangled Attention. arXiv 2021, arXiv:2006.03654. [Google Scholar]
  34. Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.; Le, Q.V. XLNet: Generalized Autoregressive Pretraining for Language Understanding. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 5753–5763. [Google Scholar]
  35. Dai, Z.; Yang, Z.; Yang, Y.; Carbonell, J.; Le, Q.; Salakhutdinov, R. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 2978–2988. [Google Scholar] [CrossRef] [Green Version]
  36. Clark, K.; Luong, M.T.; Le, Q.V.; Manning, C.D. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020; pp. 1–17. [Google Scholar]
  37. Conneau, A.; Lample, G. Cross-lingual Language Model Pretraining. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 1–11. [Google Scholar]
  38. Conneau, A.; Khandelwal, K.; Goyal, N.; Chaudhary, V.; Wenzek, G.; Guzmán, F.; Grave, E.; Ott, M.; Zettlemoyer, L.; Stoyanov, V. Unsupervised Cross-lingual Representation Learning at Scale. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 8440–8451. [Google Scholar] [CrossRef]
  39. Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 7871–7880. [Google Scholar] [CrossRef]
  40. Seide, F.; Li, G.; Chen, X.; Yu, D. Feature engineering in Context-Dependent Deep Neural Networks for conversational speech transcription. In Proceedings of the IEEE Workshop on Automatic Speech Recognition & Understanding, Waikoloa, HI, USA, 11–15 December 2011. [Google Scholar] [CrossRef]
  41. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  42. Guo, B.; Zhang, C.; Liu, J.; Ma, X. Improving text classification with weighted word embeddings via a multi-channel TextCNN model. Neurocomputing 2019, 363, 366–374. [Google Scholar] [CrossRef]
  43. Szpakowski, M. Fake News Corpus. 2021. Available online: https://github.com/several27/FakeNewsCorpus (accessed on 6 August 2021).
  44. Geva, M.; Goldberg, Y.; Berant, J. Are We Modeling the Task or the Annotator? An Investigation of Annotator Bias in Natural Language Understanding Datasets. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 1161–1166. [Google Scholar] [CrossRef]
  45. Kuwatly, H.A.; Wich, M.; Groh, G. Identifying and Measuring Annotator Bias Based on Annotators’ Demographic Characteristics. In Proceedings of the Fourth Workshop on Online Abuse and Harms. Association for Computational Linguistics, Online, 20 November 2020; pp. 184–190. [Google Scholar] [CrossRef]
  46. Jie, Z.; Xie, P.; Lu, W.; Ding, R.; Li, L. Better Modeling of Incomplete Annotations for Named Entity Recognition. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; pp. 729–734. [Google Scholar] [CrossRef] [Green Version]
  47. Bird, S.; Loper, E.; Klein, E. Natural Language Processing with Python; O’Reilly Media Inc.: Sebastopol, CA, USA, 2009. [Google Scholar]
  48. Grootendorst, M. PolyFuzz. 2021. Available online: https://maartengr.github.io/PolyFuzz/ (accessed on 14 October 2021).
  49. Arora, S.; Ge, R.; Moitra, A. Learning Topic Models—Going beyond SVD. In Proceedings of the Annual Symposium on Foundations of Computer Science, New Brunswick, NJ, USA, 20–23 October 2012; pp. 1–10. [Google Scholar]
  50. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  51. Rajapakse, T. SimpleTransformers. 2021. Available online: https://simpletransformers.ai (accessed on 12 October 2021).
  52. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 8–12 November 2020; pp. 38–45. [Google Scholar]
  53. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. Available online: https://www.tensorflow.org/ (accessed on 8 October 2021).
  54. Chollet, F.; Jin, H.; Nozawa, K. Keras. 2015. Available online: https://keras.io (accessed on 8 October 2021).
  55. Zampieri, M.; Nakov, P.; Rosenthal, S.; Atanasova, P.; Karadzhov, G.; Mubarak, H.; Derczynski, L.; Pitenis, Z.; Çağrı, Ç. SemEval-2020 Task 12: Multilingual Offensive Language Identification in Social Media (OffensEval 2020). In Proceedings of the Workshop on Semantic Evaluation, Online, 12–13 December 2020; pp. 1425–1447. [Google Scholar]
  56. Parikh, S.B.; Atrey, P.K. Media-rich fake news detection: A survey. In Proceedings of the 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), Miami, FL, USA, 10–12 April 2018; pp. 436–441. [Google Scholar] [CrossRef]
  57. Mitroi, M.; Truică, C.O.; Apostol, E.S.; Florea, A.M. Sentiment Analysis using Topic-Document Embeddings. In Proceedings of the IEEE 16th International Conference on Intelligent Computer Communication and Processing, Cluj-Napoca, Romania, 3–5 September 2020; pp. 75–82. [Google Scholar] [CrossRef]
  58. Truică, C.O.; Apostol, E.S.; Șerban, M.L.; Paschke, A. Topic-Based Document-Level Sentiment Analysis Using Contextual Cues. Mathematics 2021, 9, 2722. [Google Scholar] [CrossRef]
Figure 1. MisRoBÆRTa Architecture.
Figure 1. MisRoBÆRTa Architecture.
Mathematics 10 00569 g001
Figure 2. Performance time for the tested architectures.
Figure 2. Performance time for the tested architectures.
Mathematics 10 00569 g002
Table 1. News article classes.
Table 1. News article classes.
ClassDescription
Fake NewsFabricated or distorted information
SatireHumorous or ironic information
Extreme BiasPropaganda articles
Conspiracy TheoryPromote conspiracy theories
Junk ScienceScientifically dubious claims
Hate NewsPromote discrimination
ClickbaitCredible content, but misleading headline
Proceed With CautionMay be reliable or not
PoliticalPromote political orientations
CredibleReliable information
Table 2. Dataset description and statistics.
Table 2. Dataset description and statistics.
ClassNumber of Tokens per DocumentNumber of Tokens per Class
MeanMinMaxStdDevUniqueAll
Fake News660.86718,179899.56318,8086,808,099
Satire245.19114942232.01153,5092,502,701
Extreme Bias539.71517,4021192.27308,8275,549,908
Conspiracy Theory800.17717,448947.65315,1468,182,202
Junk Science511.05712,716732.69253,0045,208,868
Hate News930.92517,8712064.67348,2959,516,495
Clickbait361.2075544361.62193,8443,702,766
Unreliable Sources505.63717,232982.76236,4955,149,370
Political Bias547.17915,221795.81258,2695,581,168
Credible737.98615,049818.02257,4757,524,605
Entire dataset statistics583.99518,1791037.131,337,70859,726,182
Table 3. Top-10 unigram similarity between the entire dataset and each class.
Table 3. Top-10 unigram similarity between the entire dataset and each class.
ClassTop-10 UnigramsBert
Similarity
FastText
Similarity
Entire DatasetPeople Time Government American World System Year America State Public
Fake Newspeople time government world year story market American day God0.890.84
Satireorder close continue user policy agree send deny click advertising0.630.44
Extreme Biastalk people American government time Russia America state war country0.870.84
Conspiracypeople government time American America world report year country state0.940.91
Junk Sciencehealth people food free time world found body cancer study0.760.67
Hate Newspeople black white snip percent American whites time country race0.790.71
Clickbaitpeople Donald time state Clinton government America told country American0.870.84
Unreliable Sourcessystem Tor operating computer submission stick people public time order0.790.71
Political Biaspeople government time American percent year state president tax political0.890.83
Crediblepeople God Christian government American time world war told Iraq0.850.81
Table 4. Top-1 Topic similarity between the entire dataset and each class.
Table 4. Top-1 Topic similarity between the entire dataset and each class.
ClassTop-1 TopicBert
Similarity
FastText
Similarity
Entire DatasetBanner Preference Navigation Consent Technical Advertising Profile Element User Click
Fake Newsnot people Trump year day government time state world no0.630.41
Satirenavigation banner advertising consent technical preference element click profile access0.970.95
Extreme Biastalk hide link page category template user supply file previous0.710.51
Conspiracynot report government people president year world state no time0.650.41
Junk Sciencefood free health reference offer documentary herb nutrition list program0.680.42
Hate Newsnot black white snip people year percent school no student0.640.41
ClickbaitTrump not president people twitter video state woman year no0.640.41
Unreliable SourcesTor system tail browser submission computer stick communication GNU bundle0.700.45
Political BiasTrump not president year state people house government no white0.640.41
Crediblenot church people Trump God president no war Bush state0.630.39
Table 5. Transformer details.
Table 5. Transformer details.
ModelPretrained Model NameEncoder LayersHidden StateAttention HeadsParameters
BERT basebert-base-cased1276812110 M
BERT largebert-large-cased24102416335 M
DistilBERTdistilbert-base-cased67681265 M
RoBERTa baseroberta-base1276812125 M
RoBERTa largeroberta-large24102416355 M
DistilRoBERTadistilroberta-base67681282 M
XLNet basexlnet-base-cased1276812110 M
XLNet largexlnet-large-cased24102416340 M
ALBERT base v1albert-base-v1127681211 M
ALBERT base v2albert-base-v2127681211 M
ALBERT xxlarge v1albert-xxlarge-v112409664223 M
ALBERT xxlarge v2albert-xxlarge-v212409664223 M
DeBERTa basemicrosoft/deberta-base1276812140 M
DeBERTa largemicrosoft/deberta-large24102416400 M
ELECTRA basegoogle/electra-base-discriminator1276812110 M
ELECTRA largegoogle/electra-large-discriminator24102416335 M
XLMxlm-mlm-100-128016128016550 M
XLM-RoBERTa basexlm-roberta-base127688270 M
XLM-RoBERTa largelm-roberta-large24102416550 M
BART basefacebook/bart-base1276816139 M
BART largefacebook/bart-large12102416406 M
Table 6. Misinformation classification results (note: italic text marks similar results while bold text shows the overall best result).
Table 6. Misinformation classification results (note: italic text marks similar results while bold text shows the overall best result).
ModelAccuracyMicro PrecisionMacro PrecisionMicro RecallMacro RecallExecution Time (Hours)
MisRoBÆRTa92.50 ± 0.2692.50 ± 0.2692.69 ± 0.2192.50 ± 0.2692.50 ± 0.261.83 ± 0.01
BERT base90.03 ± 0.1990.03 ± 0.1990.05 ± 0.2190.03 ± 0.1990.03 ± 0.195.12 ± 0.22
BERT large89.12 ± 0.1489.12 ± 0.1489.11 ± 0.1589.12 ± 0.1489.12 ± 0.148.97 ± 0.05
DistilBERT89.36 ± 0.1589.36 ± 0.1589.40 ± 0.1689.36 ± 0.1589.36 ± 0.152.30 ± 0.01
RoBERTa base91.36 ± 0.1591.36 ± 0.1591.39 ± 0.1691.36 ± 0.1591.36 ± 0.153.95 ± 0.15
RoBERTa large88.60 ± 0.2388.61 ± 0.2288.85 ± 0.2388.59 ± 0.2488.62 ± 0.216.55 ± 0.32
DistilRoBERTa91.32 ± 0.1091.32 ± 0.1091.34 ± 0.0891.32 ± 0.1091.32 ± 0.101.93 ± 0.01
ALBERT base v186.92 ± 0.1786.92 ± 0.1786.95 ± 0.1886.92 ± 0.1786.92 ± 0.172.16 ± 0.19
ALBERT base v285.05 ± 0.2085.05 ± 0.2085.08 ± 0.1785.05 ± 0.2085.05 ± 0.202.76 ± 0.17
ALBERT xxlarge v189.20 ± 0.0489.20 ± 0.0489.36 ± 0.0289.20 ± 0.0489.20 ± 0.049.98 ± 0.09
ALBERT xxlarge v286.51 ± 2.2386.51 ± 2.2386.66 ± 2.2186.51 ± 2.2386.51 ± 2.2311.14 ± 0.04
DeBERTa base90.56 ± 0.0890.56 ± 0.0890.56 ± 0.1090.56 ± 0.0890.56 ± 0.086.90 ± 0.35
DeBERTa large89.93 ± 0.2789.93 ± 0.2789.96 ± 0.2989.93 ± 0.2789.93 ± 0.2711.06 ± 1.10
XLNet base89.94 ± 0.0989.94 ± 0.0989.94 ± 0.0989.94 ± 0.0989.94 ± 0.095.50 ± 0.06
XLNet large88.05 ± 0.3988.04 ± 0.3888.38 ± 0.3988.03 ± 0.3788.07 ± 0.4010.44 ± 1.54
ELECTRA base87.10 ± 0.2487.10 ± 0.2487.12 ± 0.2287.10 ± 0.2487.10 ± 0.244.49 ± 0.34
ELECTRA large86.92 ± 0.0986.92 ± 0.0986.92 ± 0.0886.92 ± 0.0986.92 ± 0.098.74 ± 0.43
XLM85.82 ± 0.1985.82 ± 0.1985.90 ± 0.2385.82 ± 0.1985.82 ± 0.195.16 ± 0.01
XLM-RoBERTa base89.78 ± 0.1789.78 ± 0.1789.77 ± 0.1789.78 ± 0.1789.78 ± 0.176.97 ± 0.38
XLM-RoBERTa large87.50 ± 0.6187.50 ± 0.6187.58 ± 0.5887.50 ± 0.6187.50 ± 0.619.46 ± 2.16
BART base91.35 ± 0.0491.35 ± 0.0491.35 ± 0.0491.35 ± 0.0491.35 ± 0.044.17 ± 0.16
BART large91.94 ± 0.1591.94 ± 0.1591.97 ± 0.1691.94 ± 0.1591.94 ± 0.156.79 ± 0.29
FakeBERT [8]70.18 ± 0.0170.18 ± 0.0170.21 ± 0.0370.18 ± 0.0170.18 ± 0.012.59 ± 0.01
Table 7. MisRoBÆRTa ablation and sensitivity testing using the Kaggle Fake News Dataset. (Note: bold text marks the overall best result).
Table 7. MisRoBÆRTa ablation and sensitivity testing using the Kaggle Fake News Dataset. (Note: bold text marks the overall best result).
32 Units/LayerMisRoBÆRTa with LSTM CellsMisRoBÆRTa with BiLSTM Cells
BART
Branch
RoBERTa
Branch
Ensemble
Branch
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
1 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.10 ± 0.4397.10 ± 0.4397.16 ± 0.4297.10 ± 0.4397.07 ± 0.430.01 ± 0.0097.35 ± 0.3797.35 ± 0.3797.40 ± 0.3697.35 ± 0.3797.33 ± 0.370.01 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.14 ± 0.3597.14 ± 0.3597.17 ± 0.3597.14 ± 0.3597.12 ± 0.350.02 ± 0.0097.30 ± 0.4297.30 ± 0.4297.33 ± 0.3897.30 ± 0.4297.29 ± 0.450.03 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.10 ± 0.5797.10 ± 0.5797.13 ± 0.5697.10 ± 0.5797.08 ± 0.570.03 ± 0.0097.32 ± 0.3797.32 ± 0.3797.34 ± 0.3597.32 ± 0.3797.34 ± 0.360.05 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.06 ± 0.2897.06 ± 0.2897.09 ± 0.2897.06 ± 0.2897.05 ± 0.270.04 ± 0.0097.02 ± 0.4897.02 ± 0.4897.10 ± 0.4497.02 ± 0.4896.99 ± 0.490.06 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.15 ± 0.3597.15 ± 0.3597.16 ± 0.3597.15 ± 0.3597.15 ± 0.350.05 ± 0.0097.43 ± 0.2897.43 ± 0.2897.44 ± 0.2697.43 ± 0.2897.43 ± 0.300.08 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.16 ± 0.3797.16 ± 0.3797.18 ± 0.3797.16 ± 0.3797.17 ± 0.350.07 ± 0.0097.36 ± 0.3297.36 ± 0.3297.38 ± 0.3097.36 ± 0.3297.38 ± 0.330.11 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.09 ± 0.4697.09 ± 0.4697.10 ± 0.4597.09 ± 0.4697.11 ± 0.440.08 ± 0.0097.33 ± 0.3297.33 ± 0.3297.35 ± 0.3097.33 ± 0.3297.32 ± 0.340.13 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.07 ± 0.3797.07 ± 0.3797.08 ± 0.3897.07 ± 0.3797.08 ± 0.360.09 ± 0.0097.14 ± 0.4197.14 ± 0.4197.16 ± 0.3997.14 ± 0.4197.17 ± 0.390.15 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.30 ± 0.3797.30 ± 0.3797.35 ± 0.3897.30 ± 0.3797.27 ± 0.360.11 ± 0.0097.24 ± 0.5397.24 ± 0.5397.31 ± 0.4897.24 ± 0.5397.22 ± 0.550.18 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.56 ± 0.2997.56 ± 0.2997.57 ± 0.2797.56 ± 0.2997.55 ± 0.300.12 ± 0.0097.57 ± 0.2997.57 ± 0.2997.58 ± 0.2897.57 ± 0.2997.57 ± 0.310.19 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.34 ± 0.3697.34 ± 0.3697.38 ± 0.3597.34 ± 0.3697.32 ± 0.360.13 ± 0.0097.44 ± 0.3897.44 ± 0.3897.44 ± 0.3897.44 ± 0.3897.44 ± 0.370.21 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.20 ± 0.3897.20 ± 0.3897.24 ± 0.3697.20 ± 0.3897.18 ± 0.390.15 ± 0.0097.31 ± 0.4497.31 ± 0.4497.34 ± 0.4297.31 ± 0.4497.30 ± 0.460.23 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.31 ± 0.4797.31 ± 0.4797.38 ± 0.4197.31 ± 0.4797.28 ± 0.500.16 ± 0.0097.41 ± 0.3797.41 ± 0.3797.42 ± 0.3897.41 ± 0.3797.41 ± 0.370.25 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.44 ± 0.2697.44 ± 0.2697.45 ± 0.2697.44 ± 0.2697.42 ± 0.270.18 ± 0.0097.26 ± 0.3797.26 ± 0.3797.30 ± 0.3697.26 ± 0.3797.25 ± 0.360.28 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.30 ± 0.3397.30 ± 0.3397.34 ± 0.2597.30 ± 0.3397.29 ± 0.350.19 ± 0.0097.40 ± 0.4497.40 ± 0.4497.44 ± 0.4297.40 ± 0.4497.37 ± 0.440.30 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.37 ± 0.4197.37 ± 0.4197.40 ± 0.4197.37 ± 0.4197.36 ± 0.410.21 ± 0.0097.40 ± 0.3797.40 ± 0.3797.42 ± 0.3697.40 ± 0.3797.39 ± 0.370.33 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.39 ± 0.3697.39 ± 0.3697.39 ± 0.3797.39 ± 0.3697.39 ± 0.360.22 ± 0.0097.46 ± 0.2297.46 ± 0.2297.47 ± 0.2097.46 ± 0.2297.47 ± 0.200.35 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.09 ± 0.3597.09 ± 0.3597.14 ± 0.3497.09 ± 0.3597.10 ± 0.330.24 ± 0.0097.16 ± 0.3797.16 ± 0.3797.17 ± 0.3897.16 ± 0.3797.16 ± 0.360.38 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.43 ± 0.3997.43 ± 0.3997.46 ± 0.3797.43 ± 0.3997.41 ± 0.410.26 ± 0.0097.45 ± 0.3997.45 ± 0.3997.48 ± 0.3997.45 ± 0.3997.44 ± 0.390.40 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.42 ± 0.3397.42 ± 0.3397.44 ± 0.3497.42 ± 0.3397.40 ± 0.320.27 ± 0.0197.36 ± 0.3597.36 ± 0.3597.40 ± 0.3597.36 ± 0.3597.35 ± 0.350.43 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.19 ± 0.2997.19 ± 0.2997.30 ± 0.2397.19 ± 0.2997.15 ± 0.320.29 ± 0.0097.41 ± 0.2797.41 ± 0.2797.42 ± 0.2997.41 ± 0.2797.40 ± 0.270.46 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.25 ± 0.4497.25 ± 0.4497.28 ± 0.4197.25 ± 0.4497.25 ± 0.420.31 ± 0.0197.33 ± 0.4797.33 ± 0.4797.38 ± 0.4497.33 ± 0.4797.33 ± 0.430.48 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.31 ± 0.3297.31 ± 0.3297.36 ± 0.2897.31 ± 0.3297.29 ± 0.340.32 ± 0.0197.30 ± 0.3997.30 ± 0.3997.34 ± 0.3297.30 ± 0.3997.29 ± 0.420.51 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.23 ± 0.4197.23 ± 0.4197.26 ± 0.4097.23 ± 0.4197.23 ± 0.380.34 ± 0.0197.48 ± 0.4297.48 ± 0.4297.48 ± 0.4297.48 ± 0.4297.50 ± 0.410.54 ± 0.01
3 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.41 ± 0.3497.41 ± 0.3497.43 ± 0.3597.41 ± 0.3497.39 ± 0.320.36 ± 0.0197.31 ± 0.3097.31 ± 0.3097.35 ± 0.3197.31 ± 0.3097.30 ± 0.300.57 ± 0.01
3 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.24 ± 0.2997.24 ± 0.2997.29 ± 0.3197.24 ± 0.2997.21 ± 0.280.38 ± 0.0197.42 ± 0.4597.42 ± 0.4597.44 ± 0.4397.42 ± 0.4597.42 ± 0.460.60 ± 0.02
3 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.15 ± 0.3197.15 ± 0.3197.17 ± 0.3097.15 ± 0.3197.13 ± 0.320.40 ± 0.0197.30 ± 0.4497.30 ± 0.4497.32 ± 0.4397.30 ± 0.4497.31 ± 0.430.63 ± 0.02
64 Units/LayerMisRoBÆRTa with LSTM CellsMisRoBÆRTa with BiLSTM Cells
BART
Branch
RoBERTa
Branch
Ensemble
Branch
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
1 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.24 ± 0.7897.24 ± 0.7897.29 ± 0.7297.24 ± 0.7897.25 ± 0.770.01 ± 0.0097.42 ± 0.4797.42 ± 0.4797.49 ± 0.4097.42 ± 0.4797.39 ± 0.480.01 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.36 ± 0.4297.36 ± 0.4297.41 ± 0.3997.36 ± 0.4297.34 ± 0.430.02 ± 0.0097.29 ± 0.4897.29 ± 0.4897.35 ± 0.4097.29 ± 0.4897.29 ± 0.500.03 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.42 ± 0.4797.42 ± 0.4797.44 ± 0.4197.42 ± 0.4797.41 ± 0.480.03 ± 0.0097.20 ± 0.5097.20 ± 0.5097.25 ± 0.4697.20 ± 0.5097.20 ± 0.480.04 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.48 ± 0.3797.48 ± 0.3797.48 ± 0.3797.48 ± 0.3797.48 ± 0.380.04 ± 0.0097.26 ± 0.2897.26 ± 0.2897.36 ± 0.1797.26 ± 0.2897.24 ± 0.330.06 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.33 ± 0.3997.33 ± 0.3997.34 ± 0.3997.33 ± 0.3997.33 ± 0.380.05 ± 0.0097.13 ± 0.3997.13 ± 0.3997.25 ± 0.3097.13 ± 0.3997.10 ± 0.430.08 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.36 ± 0.4297.36 ± 0.4297.38 ± 0.4097.36 ± 0.4297.35 ± 0.400.06 ± 0.0097.35 ± 0.3997.35 ± 0.3997.40 ± 0.3797.35 ± 0.3997.34 ± 0.390.10 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.32 ± 0.4697.32 ± 0.4697.35 ± 0.4597.32 ± 0.4697.31 ± 0.460.07 ± 0.0097.29 ± 0.7797.29 ± 0.7797.37 ± 0.6797.29 ± 0.7797.29 ± 0.730.12 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.53 ± 0.3497.53 ± 0.3497.56 ± 0.3397.53 ± 0.3497.50 ± 0.350.09 ± 0.0097.20 ± 0.7997.20 ± 0.7997.30 ± 0.6697.20 ± 0.7997.22 ± 0.730.14 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.32 ± 0.4097.32 ± 0.4097.32 ± 0.3997.32 ± 0.4097.33 ± 0.380.10 ± 0.0097.33 ± 0.5097.33 ± 0.5097.38 ± 0.4597.33 ± 0.5097.35 ± 0.470.16 ± 0.00
2 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.84 ± 0.5097.84 ± 0.5097.85 ± 0.5097.84 ± 0.5097.84 ± 0.480.11 ± 0.0097.85 ± 0.2397.85 ± 0.2397.88 ± 0.2497.85 ± 0.2397.83 ± 0.220.18 ± 0.00
2 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.56 ± 0.3897.56 ± 0.3897.58 ± 0.4097.56 ± 0.3897.55 ± 0.370.12 ± 0.0097.40 ± 0.3897.40 ± 0.3897.51 ± 0.3497.40 ± 0.3897.36 ± 0.380.20 ± 0.00
2 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.56 ± 0.4897.56 ± 0.4897.60 ± 0.4297.56 ± 0.4897.56 ± 0.500.14 ± 0.0097.39 ± 0.3697.39 ± 0.3697.45 ± 0.3797.39 ± 0.3697.37 ± 0.330.22 ± 0.00
2 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.53 ± 0.4897.53 ± 0.4897.57 ± 0.4897.53 ± 0.4897.50 ± 0.470.15 ± 0.0097.33 ± 0.4697.33 ± 0.4697.41 ± 0.3597.33 ± 0.4697.32 ± 0.500.24 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.56 ± 0.4597.56 ± 0.4597.61 ± 0.4297.56 ± 0.4597.55 ± 0.430.16 ± 0.0097.38 ± 0.6597.38 ± 0.6597.42 ± 0.6097.38 ± 0.6597.38 ± 0.630.26 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.53 ± 0.3897.53 ± 0.3897.56 ± 0.3697.53 ± 0.3897.52 ± 0.390.18 ± 0.0097.35 ± 0.4197.35 ± 0.4197.43 ± 0.3497.35 ± 0.4197.34 ± 0.410.28 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.52 ± 0.3597.52 ± 0.3597.55 ± 0.3497.52 ± 0.3597.52 ± 0.350.19 ± 0.0097.47 ± 0.2197.47 ± 0.2197.52 ± 0.2097.47 ± 0.2197.44 ± 0.220.30 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.52 ± 0.3797.52 ± 0.3797.57 ± 0.3497.52 ± 0.3797.50 ± 0.390.21 ± 0.0097.46 ± 0.3497.46 ± 0.3497.52 ± 0.2997.46 ± 0.3497.47 ± 0.310.33 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.63 ± 0.3197.63 ± 0.3197.65 ± 0.3097.63 ± 0.3197.60 ± 0.310.22 ± 0.0197.46 ± 0.2597.46 ± 0.2597.49 ± 0.1897.46 ± 0.2597.46 ± 0.280.36 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.72 ± 0.2597.72 ± 0.2597.73 ± 0.2697.72 ± 0.2597.70 ± 0.250.24 ± 0.0197.60 ± 0.3797.60 ± 0.3797.65 ± 0.4097.60 ± 0.3797.56 ± 0.370.38 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.67 ± 0.4397.67 ± 0.4397.72 ± 0.4397.67 ± 0.4397.65 ± 0.410.25 ± 0.0197.66 ± 0.3497.66 ± 0.3497.66 ± 0.3397.66 ± 0.3497.66 ± 0.330.40 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.58 ± 0.4397.58 ± 0.4397.64 ± 0.3597.58 ± 0.4397.55 ± 0.470.27 ± 0.0197.57 ± 0.3197.57 ± 0.3197.60 ± 0.3297.57 ± 0.3197.55 ± 0.310.42 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.59 ± 0.3097.59 ± 0.3097.62 ± 0.2997.59 ± 0.3097.58 ± 0.310.28 ± 0.0197.52 ± 0.2997.52 ± 0.2997.54 ± 0.2897.52 ± 0.2997.50 ± 0.300.44 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.57 ± 0.3797.57 ± 0.3797.60 ± 0.3897.57 ± 0.3797.56 ± 0.360.30 ± 0.0197.38 ± 0.4797.38 ± 0.4797.41 ± 0.4397.38 ± 0.4797.41 ± 0.460.47 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.63 ± 0.3897.63 ± 0.3897.65 ± 0.3597.63 ± 0.3897.63 ± 0.390.31 ± 0.0197.56 ± 0.3497.56 ± 0.3497.57 ± 0.3397.56 ± 0.3497.57 ± 0.340.50 ± 0.01
3 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.36 ± 0.4897.36 ± 0.4897.41 ± 0.4797.36 ± 0.4897.35 ± 0.460.33 ± 0.0197.30 ± 0.4897.30 ± 0.4897.41 ± 0.4297.30 ± 0.4897.24 ± 0.500.52 ± 0.01
3 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.62 ± 0.4897.62 ± 0.4897.65 ± 0.4597.62 ± 0.4897.61 ± 0.490.35 ± 0.0197.52 ± 0.2097.52 ± 0.2097.54 ± 0.2297.52 ± 0.2097.50 ± 0.200.55 ± 0.01
3 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.46 ± 0.3597.46 ± 0.3597.47 ± 0.3597.46 ± 0.3597.46 ± 0.360.37 ± 0.0197.50 ± 0.3997.50 ± 0.3997.51 ± 0.3897.50 ± 0.3997.51 ± 0.370.58 ± 0.01
128 Units/LayerMisRoBÆRTa with LSTM CellsMisRoBÆRTa with BiLSTM Cells
BART
Branch
RoBERTa
Branch
Ensemble
Branch
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
1 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.17 ± 0.5497.17 ± 0.5497.22 ± 0.5097.17 ± 0.5497.17 ± 0.540.01 ± 0.0097.23 ± 0.5297.23 ± 0.5297.26 ± 0.4997.23 ± 0.5297.25 ± 0.490.01 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM93.98 ± 0.9793.98 ± 0.9797.06 ± 0.8493.98 ± 0.9797.01 ± 0.920.02 ± 0.0097.32 ± 0.2797.32 ± 0.2797.36 ± 0.2497.32 ± 0.2797.31 ± 0.290.03 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.04 ± 0.5597.04 ± 0.5597.10 ± 0.4797.04 ± 0.5597.04 ± 0.590.03 ± 0.0097.19 ± 0.3297.19 ± 0.3297.21 ± 0.3097.19 ± 0.3297.20 ± 0.320.04 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.18 ± 0.6197.18 ± 0.6197.25 ± 0.6097.18 ± 0.6197.18 ± 0.580.03 ± 0.0097.33 ± 0.4297.33 ± 0.4297.38 ± 0.4497.33 ± 0.4297.33 ± 0.420.06 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.13 ± 0.8197.13 ± 0.8197.19 ± 0.7697.13 ± 0.8197.14 ± 0.790.05 ± 0.0097.37 ± 0.2697.37 ± 0.2697.37 ± 0.2597.37 ± 0.2697.38 ± 0.270.08 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.34 ± 0.4897.34 ± 0.4897.37 ± 0.4797.34 ± 0.4897.34 ± 0.470.06 ± 0.0097.03 ± 0.8697.03 ± 0.8697.11 ± 0.7297.03 ± 0.8697.06 ± 0.800.10 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.38 ± 0.6597.38 ± 0.6597.40 ± 0.6497.38 ± 0.6597.37 ± 0.640.07 ± 0.0097.04 ± 0.5397.04 ± 0.5397.14 ± 0.4497.04 ± 0.5397.03 ± 0.540.12 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.23 ± 0.4197.23 ± 0.4197.25 ± 0.3897.23 ± 0.4197.22 ± 0.430.08 ± 0.0097.15 ± 0.4297.15 ± 0.4297.17 ± 0.4097.15 ± 0.4297.17 ± 0.420.14 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.43 ± 0.4097.43 ± 0.4097.45 ± 0.4197.43 ± 0.4097.42 ± 0.380.09 ± 0.0097.24 ± 0.2897.24 ± 0.2897.24 ± 0.2897.24 ± 0.2897.28 ± 0.280.16 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.55 ± 0.3897.55 ± 0.3897.63 ± 0.3797.55 ± 0.3897.51 ± 0.380.10 ± 0.0097.58 ± 0.7397.58 ± 0.7397.55 ± 0.6597.58 ± 0.7397.58 ± 0.700.18 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.43 ± 0.3597.53 ± 0.3597.48 ± 0.3697.43 ± 0.3597.42 ± 0.340.11 ± 0.0097.18 ± 0.4697.18 ± 0.4697.24 ± 0.4097.18 ± 0.4697.16 ± 0.480.19 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.35 ± 0.4797.35 ± 0.4797.40 ± 0.4697.35 ± 0.4797.34 ± 0.450.13 ± 0.0097.36 ± 0.2797.36 ± 0.2797.37 ± 0.2897.36 ± 0.2797.35 ± 0.260.22 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.00 ± 0.9797.00 ± 0.9797.16 ± 0.7997.00 ± 0.9793.97 ± 0.970.14 ± 0.0097.04 ± 0.5497.04 ± 0.5497.13 ± 0.4497.04 ± 0.5497.05 ± 0.510.23 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.36 ± 0.5797.36 ± 0.5797.39 ± 0.5497.36 ± 0.5797.36 ± 0.570.15 ± 0.0097.25 ± 0.3997.25 ± 0.3997.29 ± 0.3597.25 ± 0.3997.26 ± 0.390.25 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.23 ± 0.4797.23 ± 0.4797.29 ± 0.4397.23 ± 0.4797.23 ± 0.500.16 ± 0.0097.23 ± 0.2197.23 ± 0.2197.26 ± 0.2197.23 ± 0.2197.24 ± 0.180.28 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.31 ± 0.6097.31 ± 0.6097.37 ± 0.5897.31 ± 0.6097.30 ± 0.590.18 ± 0.0197.24 ± 0.4497.24 ± 0.4497.32 ± 0.3797.24 ± 0.4497.21 ± 0.470.30 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.47 ± 0.5197.47 ± 0.5197.53 ± 0.4997.47 ± 0.5197.45 ± 0.500.19 ± 0.0197.26 ± 0.1797.26 ± 0.1797.31 ± 0.1797.26 ± 0.1797.24 ± 0.160.32 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.26 ± 0.4397.26 ± 0.4397.28 ± 0.4197.26 ± 0.4397.29 ± 0.430.21 ± 0.0197.23 ± 0.3197.23 ± 0.3197.28 ± 0.2597.23 ± 0.3197.23 ± 0.350.35 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM97.40 ± 0.5897.40 ± 0.5897.45 ± 0.5597.40 ± 0.5897.39 ± 0.570.22 ± 0.0197.40 ± 0.3597.40 ± 0.3597.40 ± 0.3497.40 ± 0.3597.43 ± 0.340.37 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM97.37 ± 0.4797.37 ± 0.4797.42 ± 0.4897.37 ± 0.4797.38 ± 0.450.23 ± 0.0197.33 ± 0.3497.33 ± 0.3497.38 ± 0.3297.33 ± 0.3497.32 ± 0.320.39 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM97.41 ± 0.6197.41 ± 0.6197.44 ± 0.6197.41 ± 0.6197.41 ± 0.600.25 ± 0.0197.35 ± 0.3097.35 ± 0.3097.37 ± 0.2997.35 ± 0.3097.35 ± 0.310.41 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM97.35 ± 0.4197.35 ± 0.4197.40 ± 0.4097.35 ± 0.4197.33 ± 0.400.26 ± 0.0197.06 ± 0.7997.06 ± 0.7997.19 ± 0.5797.06 ± 0.7997.03 ± 0.850.43 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM97.25 ± 0.5897.25 ± 0.5897.29 ± 0.5797.25 ± 0.5897.27 ± 0.580.27 ± 0.0197.35 ± 0.4197.35 ± 0.4197.37 ± 0.3897.35 ± 0.4197.33 ± 0.430.46 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM97.24 ± 0.6097.24 ± 0.6097.29 ± 0.5997.24 ± 0.6097.26 ± 0.560.29 ± 0.0197.13 ± 0.3297.13 ± 0.3297.15 ± 0.3097.13 ± 0.3297.15 ± 0.310.48 ± 0.01
3 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM97.47 ± 0.5997.47 ± 0.5997.52 ± 0.5797.47 ± 0.5997.47 ± 0.580.30 ± 0.0197.34 ± 0.2997.34 ± 0.2997.35 ± 0.2997.34 ± 0.2997.36 ± 0.280.51 ± 0.02
3 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM97.45 ± 0.5397.45 ± 0.5397.49 ± 0.5297.45 ± 0.5397.46 ± 0.510.32 ± 0.0197.23 ± 0.4497.23 ± 0.4497.24 ± 0.4397.23 ± 0.4497.24 ± 0.410.53 ± 0.02
3 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM97.48 ± 0.4797.48 ± 0.4797.49 ± 0.4797.48 ± 0.4797.48 ± 0.480.34 ± 0.0197.23 ± 0.5097.23 ± 0.5097.27 ± 0.4597.23 ± 0.5097.23 ± 0.480.57 ± 0.02
Table 8. MisRoBÆRTa ablation and sensitivity testing using the FakeNewsCorpus dataset presented in Section 5. (Note: bold text marks the overall best result).
Table 8. MisRoBÆRTa ablation and sensitivity testing using the FakeNewsCorpus dataset presented in Section 5. (Note: bold text marks the overall best result).
32 Units/LayerMisRoBÆRTa with LSTM CellsMisRoBÆRTa with BiLSTM Cells
BART
Branch
RoBERTa
Branch
Ensemble
Branch
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
1 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM91.38 ± 0.2491.38 ± 0.2491.55 ± 0.2691.38 ± 0.2491.38 ± 0.240.03 ± 0.0091.91 ± 0.3091.91 ± 0.3092.16 ± 0.2091.91 ± 0.3091.91 ± 0.300.03 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM91.35 ± 0.3691.35 ± 0.3691.57 ± 0.3291.35 ± 0.3691.35 ± 0.360.06 ± 0.0091.90 ± 0.2191.90 ± 0.2192.08 ± 0.2091.90 ± 0.2191.90 ± 0.210.06 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM90.87 ± 0.3490.87 ± 0.3491.16 ± 0.3590.87 ± 0.3490.87 ± 0.340.11 ± 0.0091.61 ± 0.2191.61 ± 0.2191.79 ± 0.1791.61 ± 0.2191.61 ± 0.210.10 ± 0.01
1 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM90.90 ± 0.3590.90 ± 0.3591.13 ± 0.3090.90 ± 0.3590.90 ± 0.350.14 ± 0.0191.62 ± 0.2991.62 ± 0.2991.84 ± 0.2891.62 ± 0.2991.62 ± 0.290.13 ± 0.01
1 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM90.87 ± 0.2590.87 ± 0.2591.13 ± 0.2590.87 ± 0.2590.87 ± 0.250.18 ± 0.0191.60 ± 0.1691.60 ± 0.1691.75 ± 0.1791.60 ± 0.1691.60 ± 0.160.17 ± 0.01
1 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM90.60 ± 0.3590.60 ± 0.3590.79 ± 0.3090.60 ± 0.3590.60 ± 0.350.23 ± 0.0191.45 ± 0.2191.45 ± 0.2191.60 ± 0.2491.45 ± 0.2191.45 ± 0.210.21 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM90.64 ± 0.4390.64 ± 0.4390.77 ± 0.4390.64 ± 0.4390.64 ± 0.430.26 ± 0.0191.51 ± 0.2291.51 ± 0.2291.70 ± 0.2491.51 ± 0.2291.51 ± 0.220.25 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM90.83 ± 0.2190.83 ± 0.2191.09 ± 0.1890.83 ± 0.2190.83 ± 0.210.31 ± 0.0191.60 ± 0.2191.60 ± 0.2191.78 ± 0.2491.60 ± 0.2191.59 ± 0.210.29 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM90.56 ± 0.3490.56 ± 0.3490.79 ± 0.3290.56 ± 0.3490.56 ± 0.340.36 ± 0.0291.32 ± 0.1891.32 ± 0.1891.53 ± 0.1691.32 ± 0.1891.32 ± 0.180.35 ± 0.02
2 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM91.98± 0.1991.98 ± 0.1992.17± 0.1591.98 ± 0.1991.98 ± 0.190.39 ± 0.0292.08± 0.1892.08 ± 0.1892.25 ± 0.1992.08 ± 0.1892.08 ± 0.180.38 ± 0.02
2 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM91.60 ± 0.1991.60 ± 0.1991.78 ± 0.2591.60 ± 0.1991.60 ± 0.190.43 ± 0.0291.80 ± 0.1791.80 ± 0.1791.94 ± 0.1891.80 ± 0.1791.80 ± 0.170.42 ± 0.02
2 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM91.19 ± 0.2891.19 ± 0.2891.37 ± 0.2791.19 ± 0.2891.19 ± 0.280.48 ± 0.0291.53 ± 0.3091.53 ± 0.3091.67 ± 0.3091.53 ± 0.3091.53 ± 0.300.47 ± 0.02
2 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM91.73 ± 0.1291.73 ± 0.1291.90 ± 0.0991.73 ± 0.1291.73 ± 0.120.51 ± 0.0291.88 ± 0.1791.88 ± 0.1792.07 ± 0.2091.88 ± 0.1791.88 ± 0.170.51 ± 0.02
2 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM91.35 ± 0.2491.35 ± 0.2491.51 ± 0.2291.35 ± 0.2491.35 ± 0.240.56 ± 0.0291.71 ± 0.2091.71 ± 0.2091.86 ± 0.1991.71 ± 0.2091.71 ± 0.200.56 ± 0.02
2 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM90.92 ± 0.3790.92 ± 0.3791.15 ± 0.3490.92 ± 0.3790.92 ± 0.370.61 ± 0.0291.20 ± 0.2591.20 ± 0.2591.32 ± 0.2491.20 ± 0.2591.20 ± 0.250.62 ± 0.03
2 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM91.55 ± 0.1891.55 ± 0.1891.68 ± 0.2191.55 ± 0.1891.55 ± 0.180.64 ± 0.0291.56 ± 0.2191.56 ± 0.2191.75 ± 0.1991.56 ± 0.2191.56 ± 0.210.67 ± 0.03
2 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM91.10 ± 0.4691.10 ± 0.4691.23 ± 0.4491.10 ± 0.4691.10 ± 0.460.69 ± 0.0391.49 ± 0.2191.49 ± 0.2191.58 ± 0.2491.49 ± 0.2191.49 ± 0.210.72 ± 0.03
2 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM90.69 ± 0.3790.69 ± 0.3790.92 ± 0.3490.69 ± 0.3790.69 ± 0.370.75 ± 0.0391.19 ± 0.2591.19 ± 0.2591.35 ± 0.3291.19 ± 0.2591.19 ± 0.250.79 ± 0.03
3 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM91.74 ± 0.2391.74 ± 0.2391.92 ± 0.2291.74 ± 0.2391.74 ± 0.230.80 ± 0.0391.92 ± 0.2291.92 ± 0.2292.06 ± 0.2391.92 ± 0.2291.92 ± 0.220.84 ± 0.04
3 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM91.39 ± 0.2091.39 ± 0.2091.58 ± 0.1791.39 ± 0.2091.39 ± 0.200.85 ± 0.0391.72 ± 0.3291.72 ± 0.3291.91 ± 0.3191.72 ± 0.3291.72 ± 0.320.92 ± 0.04
3 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM90.86 ± 0.2090.86 ± 0.2091.06 ± 0.1690.86 ± 0.2090.86 ± 0.200.90 ± 0.0491.48 ± 0.2691.48 ± 0.2691.64 ± 0.2391.48 ± 0.2691.48 ± 0.260.97 ± 0.04
3 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM91.62 ± 0.1991.62 ± 0.1991.82 ± 0.1791.62 ± 0.1991.62 ± 0.190.95 ± 0.0491.84 ± 0.2591.84 ± 0.2592.00 ± 0.2891.84 ± 0.2591.84 ± 0.251.02 ± 0.04
3 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM91.16 ± 0.2691.16 ± 0.2691.31 ± 0.2691.16 ± 0.2691.16 ± 0.261.01 ± 0.0491.51 ± 0.1991.51 ± 0.1991.66 ± 0.1791.51 ± 0.1991.51 ± 0.191.09 ± 0.05
3 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM90.36 ± 0.0090.36 ± 0.0090.79 ± 0.0090.36 ± 0.0090.36 ± 0.001.03 ± 0.0491.25 ± 0.2791.25 ± 0.2791.49 ± 0.2791.25 ± 0.2791.25 ± 0.271.17 ± 0.06
3 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM91.23 ± 0.2891.23 ± 0.2891.37 ± 0.2691.23 ± 0.2891.23 ± 0.281.09 ± 0.0491.50 ± 0.2391.50 ± 0.2391.63 ± 0.2191.50 ± 0.2391.50 ± 0.231.23 ± 0.06
3 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM90.77 ± 0.2890.77 ± 0.2890.92 ± 0.3090.77 ± 0.2890.77 ± 0.281.15 ± 0.0491.32 ± 0.1991.32 ± 0.1991.47 ± 0.1891.32 ± 0.1991.32 ± 0.191.30 ± 0.06
3 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM88.50 ± 0.3688.50 ± 0.3688.87 ± 0.2488.50 ± 0.3688.50 ± 0.361.18 ± 0.0591.22 ± 0.2591.22 ± 0.2591.41 ± 0.2091.22 ± 0.2591.22 ± 0.251.36 ± 0.08
64 Units/LayerMisRoBÆRTa with LSTM CellsMisRoBÆRTa with BiLSTM Cells
BART
Branch
RoBERTa
Branch
Ensemble
Branch
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
1 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM92.05 ± 0.1892.05 ± 0.1892.23 ± 0.1992.05 ± 0.1892.05 ± 0.180.02 ± 0.0092.26 ± 0.2392.26 ± 0.2392.40 ± 0.2192.26 ± 0.2392.26 ± 0.230.02 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM91.77 ± 0.2991.77 ± 0.2991.96 ± 0.2591.77 ± 0.2991.77 ± 0.290.04 ± 0.0092.12 ± 0.1992.12 ± 0.1992.22 ± 0.1992.12 ± 0.1992.12 ± 0.190.05 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM91.72 ± 0.2191.72 ± 0.2191.90 ± 0.2191.72 ± 0.2191.72 ± 0.210.07 ± 0.0091.90 ± 0.1791.90 ± 0.1792.07 ± 0.1891.90 ± 0.1791.90 ± 0.170.08 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM91.72 ± 0.2891.72 ± 0.2891.92 ± 0.2991.72 ± 0.2891.72 ± 0.280.09 ± 0.0092.02 ± 0.2092.02 ± 0.2092.16 ± 0.1892.02 ± 0.2092.02 ± 0.200.11 ± 0.01
1 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM91.63 ± 0.1391.63 ± 0.1391.83 ± 0.1391.63 ± 0.1391.63 ± 0.130.11 ± 0.0091.92 ± 0.2891.92 ± 0.2892.14 ± 0.2191.92 ± 0.2891.92 ± 0.280.14 ± 0.01
1 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM91.48 ± 0.3591.48 ± 0.3591.73 ± 0.2891.48 ± 0.3591.48 ± 0.350.15 ± 0.0191.97 ± 0.1591.97 ± 0.1592.06 ± 0.1591.97 ± 0.1591.97 ± 0.150.18 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM91.57 ± 0.2391.57 ± 0.2391.78 ± 0.2091.57 ± 0.2391.57 ± 0.230.17 ± 0.0192.01 ± 0.1692.01 ± 0.1692.16 ± 0.1992.01 ± 0.1692.01 ± 0.160.21 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM91.66 ± 0.1691.66 ± 0.1691.79 ± 0.1891.66 ± 0.1691.66 ± 0.160.19 ± 0.0191.90 ± 0.2191.90 ± 0.2192.08 ± 0.2291.90 ± 0.2191.90 ± 0.210.25 ± 0.01
1 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM91.37 ± 0.2691.37 ± 0.2691.55 ± 0.2791.37 ± 0.2691.37 ± 0.260.23 ± 0.0191.64 ± 0.2291.64 ± 0.2291.77 ± 0.2491.64 ± 0.2291.64 ± 0.220.29 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM92.30± 0.3092.30 ± 0.3092.47 ± 0.2492.30 ± 0.3092.30 ± 0.300.25 ± 0.0192.41± 0.1892.41 ± 0.1892.57 ± 0.1392.41 ± 0.1892.41 ± 0.180.32 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM92.00 ± 0.3592.00 ± 0.3592.14 ± 0.3292.00 ± 0.3592.00 ± 0.350.28 ± 0.0192.22 ± 0.1792.22 ± 0.1792.40 ± 0.1692.22 ± 0.1792.22 ± 0.170.35 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM91.94 ± 0.2991.94 ± 0.2992.06 ± 0.3191.94 ± 0.2991.94 ± 0.290.31 ± 0.0192.12 ± 0.2792.12 ± 0.2792.24 ± 0.2592.12 ± 0.2792.12 ± 0.270.40 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM92.06 ± 0.2192.06 ± 0.2192.18 ± 0.2392.06 ± 0.2192.06 ± 0.210.33 ± 0.0192.32 ± 0.2392.32 ± 0.2392.49 ± 0.2492.32 ± 0.2392.32 ± 0.230.43 ± 0.02
2 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM91.83 ± 0.2591.83 ± 0.2591.99 ± 0.2591.83 ± 0.2591.83 ± 0.250.36 ± 0.0192.16 ± 0.2592.16 ± 0.2592.31 ± 0.1592.16 ± 0.2592.16 ± 0.250.47 ± 0.02
2 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM91.61 ± 0.1991.61 ± 0.1991.77 ± 0.2191.61 ± 0.1991.61 ± 0.190.40 ± 0.0191.74 ± 0.2591.74 ± 0.2591.91 ± 0.2691.74 ± 0.2591.74 ± 0.250.52 ± 0.02
2 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM91.97 ± 0.2791.97 ± 0.2792.10 ± 0.2491.97 ± 0.2791.97 ± 0.270.43 ± 0.0192.01 ± 0.2592.01 ± 0.2592.19 ± 0.1692.01 ± 0.2592.01 ± 0.250.56 ± 0.02
2 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM91.81 ± 0.3191.81 ± 0.3191.97 ± 0.3191.81 ± 0.3191.81 ± 0.310.46 ± 0.0291.81 ± 0.1791.81 ± 0.1791.96 ± 0.2391.81 ± 0.1791.81 ± 0.170.60 ± 0.02
2 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM91.47 ± 0.2691.47 ± 0.2691.64 ± 0.2491.47 ± 0.2691.47 ± 0.260.50 ± 0.0291.51 ± 0.4391.51 ± 0.4391.70 ± 0.4191.51 ± 0.4391.51 ± 0.430.66 ± 0.02
3 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM92.29 ± 0.2692.29 ± 0.2692.43 ± 0.2192.29 ± 0.2692.29 ± 0.260.52 ± 0.0292.19 ± 0.3192.19 ± 0.3192.31 ± 0.3492.19 ± 0.3192.19 ± 0.310.70 ± 0.02
3 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM91.90 ± 0.2191.90 ± 0.2192.09 ± 0.1691.90 ± 0.2191.90 ± 0.210.56 ± 0.0291.97 ± 0.2191.97 ± 0.2192.09 ± 0.2091.97 ± 0.2191.97 ± 0.210.74 ± 0.03
3 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM91.85 ± 0.2891.85 ± 0.2891.98 ± 0.2691.85 ± 0.2891.85 ± 0.280.60 ± 0.0291.91 ± 0.1591.91 ± 0.1592.06 ± 0.1791.91 ± 0.1591.91 ± 0.150.80 ± 0.03
3 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM92.08 ± 0.1992.08 ± 0.1992.20 ± 0.1992.08 ± 0.1992.08 ± 0.190.63 ± 0.0292.16 ± 0.2392.16 ± 0.2392.31 ± 0.1992.16 ± 0.2392.16 ± 0.230.84 ± 0.03
3 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM91.73 ± 0.2791.73 ± 0.2791.93 ± 0.2791.73 ± 0.2791.73 ± 0.270.67 ± 0.0291.91 ± 0.1391.91 ± 0.1392.03 ± 0.1691.90 ± 0.1391.91 ± 0.130.92 ± 0.03
3 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM91.43 ± 0.1691.43 ± 0.1691.63 ± 0.1691.43 ± 0.1691.43 ± 0.160.70 ± 0.0291.59 ± 0.0791.59 ± 0.0791.71 ± 0.1191.59 ± 0.0791.59 ± 0.070.93 ± 0.04
3 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM91.92 ± 0.2291.92 ± 0.2292.03 ± 0.2291.92 ± 0.2291.92 ± 0.220.73 ± 0.0291.97 ± 0.2391.97 ± 0.2392.13 ± 0.2491.97 ± 0.2391.97 ± 0.230.98 ± 0.04
3 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM91.61 ± 0.2891.61 ± 0.2891.77 ± 0.2791.61 ± 0.2891.61 ± 0.280.77 ± 0.0391.80 ± 0.1691.80 ± 0.1691.93 ± 0.1891.81 ± 0.1691.80 ± 0.161.04 ± 0.04
3 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM89.81 ± 0.2189.81 ± 0.2189.91 ± 0.2789.81 ± 0.2189.81 ± 0.210.79 ± 0.0385.53 ± 0.1185.53 ± 0.1185.64 ± 0.1785.53 ± 0.1185.53 ± 0.111.08 ± 0.04
128 Units/LayerMisRoBÆRTa with LSTM CellsMisRoBÆRTa with BiLSTM Cells
BART
Branch
RoBERTa
Branch
Ensemble
Branch
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
AccuracyPrecision
Micro
Precision
Macro
Recall
Micro
Recall
Macro
Execution
Time (Hours)
1 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM92.16 ± 0.2792.16 ± 0.2792.35 ± 0.2292.16 ± 0.2792.16 ± 0.270.02 ± 0.0092.20 ± 0.3092.20 ± 0.3092.35 ± 0.2992.20 ± 0.3092.20 ± 0.300.03 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM92.11 ± 0.3792.11 ± 0.3792.26 ± 0.3592.11 ± 0.3792.11 ± 0.370.03 ± 0.0092.23 ± 0.1492.23 ± 0.1492.37 ± 0.1392.23 ± 0.1492.23 ± 0.140.05 ± 0.00
1 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM91.95 ± 0.2991.95 ± 0.2992.14 ± 0.2391.95 ± 0.2991.95 ± 0.290.05 ± 0.0092.04 ± 0.2592.04 ± 0.2592.17 ± 0.2492.04 ± 0.2592.04 ± 0.250.09 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM92.02 ± 0.2992.02 ± 0.2992.17 ± 0.2492.02 ± 0.2992.02 ± 0.290.07 ± 0.0092.17 ± 0.2692.17 ± 0.2692.36 ± 0.2392.17 ± 0.2692.17 ± 0.260.11 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM91.87 ± 0.2491.87 ± 0.2491.98 ± 0.2691.87 ± 0.2491.87 ± 0.240.09 ± 0.0091.94 ± 0.2391.94 ± 0.2392.03 ± 0.2591.94 ± 0.2391.94 ± 0.230.15 ± 0.00
1 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM91.56 ± 0.3291.56 ± 0.3291.72 ± 0.3291.56 ± 0.3291.56 ± 0.320.11 ± 0.0091.62 ± 0.2791.62 ± 0.2791.76 ± 0.2791.62 ± 0.2791.62 ± 0.270.18 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM91.86 ± 0.3091.86 ± 0.3092.01 ± 0.3191.86 ± 0.3091.86 ± 0.300.13 ± 0.0091.97 ± 0.1891.97 ± 0.1892.09 ± 0.1891.97 ± 0.1891.97 ± 0.180.21 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM91.58 ± 0.5291.58 ± 0.5291.74 ± 0.5191.58 ± 0.5291.58 ± 0.520.15 ± 0.0091.53 ± 0.2891.53 ± 0.2891.65 ± 0.3391.53 ± 0.2891.53 ± 0.280.25 ± 0.00
1 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM91.09 ± 0.4391.09 ± 0.4391.21 ± 0.4291.09 ± 0.4391.09 ± 0.430.17 ± 0.0091.07 ± 0.3291.07 ± 0.3291.17 ± 0.3491.07 ± 0.3291.07 ± 0.320.29 ± 0.00
2 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM92.50± 0.3892.50 ± 0.3892.67 ± 0.3592.50 ± 0.3892.50 ± 0.380.19 ± 0.0092.50± 0.2692.50 ± 0.2692.69 ± 0.2192.50 ± 0.2692.50 ± 0.260.32 ± 0.00
2 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM92.30 ± 0.2892.30 ± 0.2892.46 ± 0.2592.30 ± 0.2892.30 ± 0.280.21 ± 0.0192.35 ± 0.2692.35 ± 0.2692.52 ± 0.2292.35 ± 0.2692.35 ± 0.260.35 ± 0.01
2 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM92.04 ± 0.3792.04 ± 0.3792.20 ± 0.3892.04 ± 0.3792.04 ± 0.370.24 ± 0.0092.11 ± 0.2592.11 ± 0.2592.20 ± 0.2192.11 ± 0.2592.11 ± 0.250.39 ± 0.00
2 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM92.36 ± 0.3192.36 ± 0.3192.48 ± 0.2592.36 ± 0.3192.36 ± 0.310.26 ± 0.0092.40 ± 0.2092.40 ± 0.2092.52 ± 0.2192.40 ± 0.2092.40 ± 0.200.42 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM92.05 ± 0.2392.05 ± 0.2392.17 ± 0.2192.05 ± 0.2392.05 ± 0.230.28 ± 0.0092.19 ± 0.2592.19 ± 0.2592.33 ± 0.2292.19 ± 0.2592.19 ± 0.250.46 ± 0.01
2 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM91.70 ± 0.4791.70 ± 0.4791.87 ± 0.4591.70 ± 0.4791.70 ± 0.470.31 ± 0.0191.69 ± 0.3591.69 ± 0.3591.88 ± 0.4091.69 ± 0.3591.69 ± 0.350.51 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM92.13 ± 0.2892.13 ± 0.2892.36 ± 0.2792.13 ± 0.2892.13 ± 0.280.33 ± 0.0192.16 ± 0.2892.16 ± 0.2892.33 ± 0.2792.16 ± 0.2892.16 ± 0.280.54 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM91.94 ± 0.3291.94 ± 0.3292.06 ± 0.3391.94 ± 0.3291.94 ± 0.320.35 ± 0.0191.69 ± 0.3291.69 ± 0.3291.79 ± 0.3791.69 ± 0.3291.69 ± 0.320.58 ± 0.01
2 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM91.18 ± 0.2591.18 ± 0.2591.30 ± 0.2791.18 ± 0.2591.18 ± 0.250.38 ± 0.0191.18 ± 0.2391.18 ± 0.2391.29 ± 0.2491.18 ± 0.2391.18 ± 0.230.63 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM1 × [Bi]LSTM92.38 ± 0.2592.38 ± 0.2592.50 ± 0.2292.38 ± 0.2592.38 ± 0.250.40 ± 0.0192.34 ± 0.2192.34 ± 0.2192.45 ± 0.2392.34 ± 0.2192.34 ± 0.210.67 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM2 × [Bi]LSTM92.13 ± 0.3092.13 ± 0.3092.23 ± 0.3092.13 ± 0.3092.13 ± 0.300.43 ± 0.0192.16 ± 0.2792.16 ± 0.2792.26 ± 0.2692.16 ± 0.2792.16 ± 0.270.71 ± 0.01
3 × [Bi]LSTM1 × [Bi]LSTM3 × [Bi]LSTM91.87 ± 0.2691.87 ± 0.2692.13 ± 0.2491.87 ± 0.2691.87 ± 0.260.45 ± 0.0192.06 ± 0.2392.06 ± 0.2392.16 ± 0.2692.06 ± 0.2392.06 ± 0.230.76 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM1 × [Bi]LSTM92.28 ± 0.2892.28 ± 0.2892.42 ± 0.2692.28 ± 0.2892.28 ± 0.280.48 ± 0.0192.22 ± 0.1792.22 ± 0.1792.37 ± 0.1192.22 ± 0.1792.22 ± 0.170.80 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM2 × [Bi]LSTM92.04 ± 0.2592.04 ± 0.2592.20 ± 0.3192.04 ± 0.2592.04 ± 0.250.50 ± 0.0192.04 ± 0.2592.04 ± 0.2592.17 ± 0.2692.04 ± 0.2592.04 ± 0.250.84 ± 0.01
3 × [Bi]LSTM2 × [Bi]LSTM3 × [Bi]LSTM91.14 ± 0.2191.14 ± 0.2191.19 ± 0.2491.14 ± 0.2191.14 ± 0.210.52 ± 0.0191.30 ± 0.1191.30 ± 0.1191.43 ± 0.1291.30 ± 0.1191.30 ± 0.110.88 ± 0.01
3 × [Bi]LSTM3 × [Bi]LSTM1 × [Bi]LSTM92.19 ± 0.2392.19 ± 0.2392.31 ± 0.2592.19 ± 0.2392.19 ± 0.230.55 ± 0.0192.19 ± 0.1892.19 ± 0.1892.27 ± 0.2192.19 ± 0.1892.19 ± 0.180.92 ± 0.02
3 × [Bi]LSTM3 × [Bi]LSTM2 × [Bi]LSTM92.10 ± 0.1392.10 ± 0.1392.19 ± 0.1292.10 ± 0.1392.10 ± 0.130.58 ± 0.0291.57 ± 0.3291.57 ± 0.3291.70 ± 0.3591.57 ± 0.3291.57 ± 0.320.97 ± 0.02
3 × [Bi]LSTM3 × [Bi]LSTM3 × [Bi]LSTM91.01 ± 0.0991.01 ± 0.0991.09 ± 0.0891.01 ± 0.0991.01 ± 0.090.60 ± 0.0291.03 ± 0.1791.03 ± 0.1791.14 ± 0.2091.03 ± 0.1791.03 ± 0.171.02 ± 0.02
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Truică, C.-O.; Apostol, E.-S. MisRoBÆRTa: Transformers versus Misinformation. Mathematics 2022, 10, 569. https://0-doi-org.brum.beds.ac.uk/10.3390/math10040569

AMA Style

Truică C-O, Apostol E-S. MisRoBÆRTa: Transformers versus Misinformation. Mathematics. 2022; 10(4):569. https://0-doi-org.brum.beds.ac.uk/10.3390/math10040569

Chicago/Turabian Style

Truică, Ciprian-Octavian, and Elena-Simona Apostol. 2022. "MisRoBÆRTa: Transformers versus Misinformation" Mathematics 10, no. 4: 569. https://0-doi-org.brum.beds.ac.uk/10.3390/math10040569

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop