Next Article in Journal
Impact Load Localization Based on Multi-Scale Feature Fusion Convolutional Neural Network
Previous Article in Journal
Contrastive Learning-Based Personalized Tag Recommendation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Techniques for Robust Fake News Detection: Integrating Transformers, Natural Language Processing, and Machine Learning

Department of Electrical Engineering and Computer Science, Howard University, Washington, DC 20059, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(18), 6062; https://doi.org/10.3390/s24186062
Submission received: 23 July 2024 / Revised: 6 September 2024 / Accepted: 16 September 2024 / Published: 19 September 2024
(This article belongs to the Section Internet of Things)

Abstract

:
The proliferation of fake news across multiple modalities has emerged as a critical challenge in the modern information landscape, necessitating advanced detection methods. This study proposes a comprehensive framework for fake news detection integrating text, images, and videos using machine learning and deep learning techniques. The research employs a dual-phased methodology, first analyzing textual data using various classifiers, then developing a multimodal approach combining BERT for text analysis and a modified CNN for visual data. Experiments on the ISOT fake news dataset and MediaEval 2016 image verification corpus demonstrate the effectiveness of the proposed models. For textual data, the Random Forest classifier achieved 99% accuracy, outperforming other algorithms. The multimodal approach showed superior performance compared to baseline models, with a 3.1% accuracy improvement over existing multimodal techniques. This research contributes to the ongoing efforts to combat misinformation by providing a robust, adaptable framework for detecting fake news across different media formats, addressing the complexities of modern information dissemination and manipulation.

1. Introduction

The widespread challenge of detecting misinformation, particularly in the form of fake news, has emerged as an essential research priority in modern information distribution. Social media platforms and digital communication channels have heralded an era in which misleading narratives spread easily across several modalities, including text, photos, videos, and speech [1]. To meet this need, modern machine learning and deep learning algorithms play a critical role in ensuring the integrity of information sources. The significance of multimodal false news detection approaches extends beyond theoretical notions, with substantial societal consequences [2]. The consequences of false narratives are diverse and deep in this era of the rapid dissemination of information.
The concept of “fake news” is not a new phenomenon; it has deep roots in society and has risen to the level of a significant problem requiring attention from the research community [3,4]. Recently, the term has evolved, diverting from prior definitions that embraced a wide range of content, encompassing satires, scams, propaganda, and clickbait [5]. Understanding the causes of the spread of fake news is critical to resolving this global challenge. One key factor is the viewers’ lack of information about the source legitimacy and news authenticity [6]. This information void exposes the public to potentially dangerous misinformation. Another factor is the lack of effective automated fact-checking systems [7]. While systems show progress in detecting false news, the manual aspect of their methods makes them time-consuming and incapable of preventing the rapid spread of fake news [8]. Furthermore, multi-modal data, which might come from textual articles, videos, images, and speech, demand a complex analytical approach. When confronted with modified images and sophisticated false narratives, conventional techniques of news verification fail. As a result, deep learning techniques, such as natural language processing (NLP) for text analysis and computer vision for image and video authentication, appear to be a viable alternative [9]. The shortcomings in multi-modal data sets highlight the need for reliable methods for identifying them. Images and videos, which can be easily changed using video editing applications, can distort the facts and spread false narratives. Textual articles, despite their simplicity, might include complex nuances and language manipulations that challenge established verification methods. Speech data add complexity, with artificial voices possibly intensifying false messaging.
In response to these challenges, adopting machine learning models becomes critical, providing an advanced method of detecting detailed patterns and correlations within massive datasets. Spurious patterns in textual content, inappropriate visual features in photos and videos, and anomalies in voice patterns can all be recognized and flagged by combining machine learning techniques. Deep learning models, which use neural networks to imitate human-like learning and decision-making processes, improve detection across multi-modal data sources. Navigating the complexities of the age of technology requires developing and deploying cutting-edge machine learning and deep learning approaches for detecting fake news. These approaches not only deter the concealed transmission of misinformation but also highlight the expanding role of technology in ensuring the credibility of worldwide information dissemination [10].
The framework integrates text, images, and videos to thoroughly detect fake news. We utilize advanced machine learning (ML) and deep learning (DL) methodologies, specifically NLP for text and computer vision for images and videos. A detailed comparison between simpler textual data analysis using traditional machine learning algorithms and complex multi-modal data analysis using deep learning models demonstrates the superior capabilities of the proposed model. We apply BERT (Bidirectional Encoder Representations from Transformers) to integrate textual and visual data and combine BERT with sophisticated deep learning layers to enhance the detection capabilities. Performance assessments indicate the superior accuracy, recall capabilities, and F1-score of the suggested model. These data also suggest the effectiveness of the random forest model in unimodal textual data classification, achieving a 99% accuracy rate. We identifiy and address the specific challenges posed by text, images, videos, and speech data. This paper proposes robust solutions to detect and mitigate the spread of false narratives across these modalities.
In summary, our contributions are as follows:
  • The development of a framework that integrates text, images, and videos for comprehensive fake news detection, leveraging advanced machine learning and deep learning methodologies.
  • The application of BERT to integrate textual and visual data, combining it with sophisticated deep learning layers for improved detection accuracy.
  • The demonstration of the effectiveness of machine learning models in unimodal textual data classification, achieving a 99% accuracy rate. This shows the importance of machine learning models in terms of their complexity for unimodal data.
  • The identification of and attention to the specific challenges posed by text, images, and video data in the context of fake news detection.
Furthermore, the paper is organized into the following sections:
The literature review (Section 2) provides insights into the historical context and definitions of fake news, examines the causes of its spread, and discusses the current state of research in multi-modal data analysis and machine learning approaches for detecting fake news. The methodology (Section 3) details the proposed framework for detecting fake news using multi-modal data, including data collection, preprocessing steps, application of ML and DL techniques, and metrics used for performance evaluation. The results (Section 4) presents the performance outcomes of the proposed model, including a comparison with baseline models. This section highlights the effectiveness of the framework in handling both unimodal and multi-modal data. The conclusions (Section 5) present the most consequential findings and impacts of the study.

2. Literature Review

The authors of [11] proposed a unique approach to detect false news by integrating text and photos using a cultural algorithm that also utilizes data gained from situational and normative knowledge. Their model includes multiple components: a sentiment analysis-based textual feature extractor, a visual feature extractor, and a classifier-based false information detector. Extensive trials on real-world multi-modal datasets, such as Weibo and X (formerly Twitter), showed that their method outperformed state-of-the-art algorithms by 9% on average. Singh, Ghosh, and Sonagara [12] presented a multi-modal technique combining text and visual analytics for automated fake news identification. Using the Kaggle Fake News Dataset, their approach involves training classifiers on balanced subsets of fake and credible news articles across 100 iterations. They implemented numerous MLMs, including random forest, logistic regression, and SVM, achieving robust classifier performance through 10-fold cross validation.
Ying et al. [13] introduced the Multi-level pre-trainedodal Cross-attention Network (MMCN) to tackle the challenges of detecting fake news in the mobile internet era. MMCN leverages pre-trained BERT and ResNet models to generate high-quality approximations for text and image features, combined through a multi-modal cross-attention network. Their experiments on the WEIBO and PHEME datasets demonstrated the MMCN’s superior performance over existing models.
Song et al. [14] developed the Cross-modal Attention Residual Network (CARN) and Multichannel Convolutional Neural Network (MCN) within their Cross-modal Attention Residual framework (CARMN). This approach effectively extracts and fuses essential data from different modalities while mitigating noisy information. Their model outperformed state-of-the-art methods in extensive tests across four real-world datasets.
Chen et al. [15] proposed CAFE (Cross-modal Ambiguity-aware Fake News Detection), encompassing a fusion, cross-modal alignment, and ambiguity learning modules. This method adjusts its approach based on cross-modal ambiguity levels, and significantly improves fake news detection accuracy on Twitter and Weibo datasets.
Qian et al. [16] presented the Hierarchical pre-trained Multi-modal Contextual Attention Network (HMCAN), which employs ResNet and BERT for image and text representations, respectively. Their network considers both inter-modality and intra-modality interactions, with hierarchical encoding to capture extensive hierarchical semantics. The HMCAN showed effectiveness across the WEIBO, TWITTER, and PHEME datasets.
Raj and Meel [17] explored multi-modal online information credibility assessments using deep networks such as CNNs and RNNs. Their pre-trained Multi-modal Coupled ConvNet architecture effectively classified online news based on textual and visual information, demonstrating high accuracy across datasets like the TI-CNN, EMERGENT, and MICC-F220.
Choi and Ko [18] focused on detecting misleading videos by combining domain knowledge with multi-modal data fusion. By incorporating domain-specific information and using a linear combination of features, their approach improved detection performance, achieving a 3% gain in accuracy across the test datasets.
Chen, Chu, and Subbalakshmi [19] addressed COVID-19 misinformation with a novel multi-modal dataset and proposed a framework for classifying news as true or false. Their F-Score of this method was estimated at 0.919, and it also achieved 0.882 accuracy in identifying misleading information.
Xue et al. [20] introduced the Multi-modal Consistency Neural Network (MCNN) to detect fake news by extracting and fusing textual and visual features. Their approach demonstrated significant accuracy improvements on several datasets by effectively handling multi-modal data.
Danlei Chen et al. [21] introduced the relevance classifier method and integrated it into the multi-modal framework, with image-text similarity visualization using feature extraction.
Qi et al. [22] identified key textual–image relationships in multi-modal fake news and proposed an entity-enhanced multi-modal fusion method. Their model, which captures critical text-image correlations, was superior in detecting multi-modal fake news.
Singhal et al. [23] developed SpotFake, a multi-modal framework for detecting fake news without relying on extra subtasks, using BERT for text and VGG-19 for image features. When used on datasets obtained from X (formerly Twitter) and Weibo, the algorithm outperformed existing algorithms by an average of 3.27% on X (formerly Twitter) and 6.83% on Weibo. Table 1 shows the most closely related papers with the necessary parameters from the literature.

3. Methodology

This section outlines the dual-phased methodology of the research. The main aim of this methodology is to present a separate model for fake news detection based on the nature and dimensions of data. This methodology is not only helpful in identifying more accurate algorithms in terms of accuracy, but it also provides insight into the performance of the algorithms, which leads to better utilization of resources and efficiently identifying false news.

3.1. Datasets

The primary ISOT fake news dataset [24] contains textual data from various sources, including political statements, news articles, and press reports from world seminars. It comprises over 40,000 text articles, evenly balanced between true and false classes. The second dataset analyzed is an evolving collection of images shared on social media, notably Twitter, and is available on GitHub [25]. This free corpus evaluates online image verification techniques by leveraging user characteristics and tweeted text. It includes three essential files, serving as a comprehensive resource for confirmed fake and real images Table 2.
The set_images.txt file details the image_id, image_url, annotation (indicating the image’s legitimacy), and associated events. The tweets_images.txt file links each image_id with the tweet’s validity, the event’s origin, and the accompanying tweets. The tweets_images_update.txt file focuses on misleading tweets, specifically those lacking humor or containing false remarks, thereby improving the dataset by retaining tweets with erroneous information. The tweets_event.txt file filters out fabricated tweets that have been deleted or whose accounts have been deactivated. Researchers can use these files in conjunction with set_images.txt to maximize the dataset’s utility.
This resource is crucial for computational verification endeavors, offering a fundamental framework for researchers in the field. In addition to features based on user and tweet attributes and forensic features for related images, the dev set and test set files provide Twitter data for training and testing, respectively. This large-scale dataset and its well-structured arrangement support numerous research projects related to social media analysis and computational verification. Figure 1 and Figure 2 illustrate examples images from the MediaEval 2016 dataset.

3.2. Proposed Models on Textual Data

The work commenced with meticulous preparation of the textual data using the spaCy natural language processing toolset. The method involved tokenization and cleaning to prepare the text for examination. Afterward, the TF-IDF vectorizer was utilized to transform the preprocessed text data into a numerical representation, establishing the basis for subsequent analysis. To evaluate the chosen classifiers, such as random forest, multinomial naïve Bayes, support vector machine, logistic regression, and k-nearest neighbors, the dataset was carefully split into separate training and testing sets. After receiving training on the specified sets, the accuracy of each classifier was assessed using the testing set. The documentation of the outcomes facilitated the development of a thorough comparison summary table that demonstrates the relative performance of each classifier.
In the case of the random forest classifier, an extra stage was performed to optimize the hyperparameters using Grid Search. The objective of this approach was to enhance the performance of the classifier by identifying the most efficient hyperparameters. After being established, the optimized random forest classifier was trained and evaluated on the testing set. A comprehensive classification report was generated, which provided information on the accuracy, recall, and F1-score. To investigate possible improvements in the performance of the model, different methods of representing features were analyzed. The techniques utilized were TF-IDF, Word2Vec, N-grams, FastText, Doc2Vec, Bag of Words (BoW), and Hashing Vectorizer. The project sought to evaluate the influence of various feature extraction strategies on the overall effectiveness of the models. This methodological framework is thorough and follows a methodical and step-by-step plan. It starts with preparing the data and evaluating the classifier. Then, it moves on to modifying the hyperparameters and finally explores several techniques for representing features. The systematic approach of these processes enhances the rigorous and comprehensive research into the detection of false information within the dataset. Figure 3 represents the architecture diagram of textual data methodology.

3.3. Proposed Models on the Multi-Modal Dataset

This section presents an improved multi-modal approach that includes a modified Convolutional Neural Network (CNN) structure to accurately identify disinformation. The system consists of several essential elements, including a module that combines multiple features, a feature extractor that incorporates an attention mechanism, a textual feature extractor, and a visual feature extractor. The textual feature extractor begins by carefully preparing the textual data, which includes tokenization, word normalization using metaphor, replacing text-based emojis with sentiment terms, and shortening long sentences. This section presents a sophisticated approach for extracting features from text, utilizing a pre-trained BERT model that is specifically tailored for analyzing Tweet data. The combination of the last four hidden layers of BERT, which are known for their effectiveness in extracting features, produces contextual embeddings. Figure 4 shows the visual encoder used in the multi-modal method Algorithm 1.
Algorithm 1 Multi-modal Disinformation Detection
1:
Input: Raw text data T, Image data I
2:
Output: Comprehensive representation C
3:
Textual Feature Extraction:
4:
   Tokenize text: T t = tokenize ( T )
5:
   Normalize words: T n = normalize ( T t )
6:
   Replace emojis: T e = replace _ emojis ( T n )
7:
   Shorten sentences: T s = shorten ( T e )
8:
   Extract BERT embeddings:
E = BERT ( T s ) where E = [ h 4 , h 3 , h 2 , h 1 ]
9:
   Combine embeddings: T f = combine ( E )
10:
Visual Feature Extraction:
11:
   Pre-trained ResNet V2 model:
I 1 = ResNet ( I )
12:
   Fully connected layers:
I f = FC ( I 1 )
13:
   Process visual representation:
I m = process ( I f ) where d I = 16
14:
Attention Mechanism:
15:
   Apply attention:
A T I = Attention ( T f , I f )
A I T = Attention ( I f , T f )
A I I = Attention ( I f , I f )
16:
   Fully connected layers with normalization:
R T I = FC ( A T I ) + T f
R I T = FC ( A I T ) + I f
R I I = FC ( A I I ) + I f
17:
Final Processing:
18:
   Compress and combine features:
R I I = FC ( R I I )
19:
   Fully connected layer with 32 neural units:
C = FC 32 ( [ T f , I f , R T I , R I T , R I I ] )
The visual feature extraction process utilizes a pre-trained ResNet V2 model with an input size of 128 × 128 × 3. Two fully connected layers are used, with the output of the second-to-last layer reducing the dimension to a vector size of d I = 16 . This vector forms the final visual representation, denoted as I f . The output from the third final layer undergoes additional processing and manipulation to generate a visual feature representation with a dimension of d I m = 16 .
The common feature extractor with an attention mechanism presents an enhanced scaled dot-product attention approach ( I f , I m , T f , T m ) that may be used for both textual and visual components. This mechanism facilitates the establishment of linkages between the text and images of a post. It incorporates self-attention on images and bidirectional attention processes on text and visual elements. The matrices A t t T I , A t t I T , and A t t I I are processed using fully connected layers, which incorporate layer normalization and a residual connection. The outcome consists of three vectors: R T I , R I T , and R I I , which represent the combined features. The final step entails compressing the feature vector R I I and passing it through a fully connected layer to obtain R I I . Afterward, a fully connected layer consisting of 32 neural units combines and transmits the outputs ( T f , I f , R T I , R I T , and R I I ) to provide a comprehensive representation of both textual and visual features. This improved design seeks to enhance the extraction and integration of textual and visual information to increase the effectiveness of the model in recognizing disinformation. Figure 5 represents the architecture diagram for the proposed methodology.

4. Results

4.1. Comparison of Machine Leaning Algorithms on a Unimodal Textual Dataset

Different ML algorithms were used to assess various classification models and their performance in identifying fake news, followed by examining their performance. The table presents the essential parameters of these algorithms: F1-scores, accuracy, recall, precision. It also summarizes the results across several modalities. With a 99% accuracy rate and high precision, recall, and F1-score values at 0.99, the random forest classifier proved effective. With a 95% accuracy rate, the support vector machine (SVM) demonstrated impressive performance. Its precision, recall, and F1-score metrics were also good, at 0.96, 0.94, and 0.95, respectively. The k-nearest neighbors (KNN) and logistic regression demonstrated excellent accuracy rates of 92% and 91%, respectively, along with balanced precision, recall, and F1-score values. Multinomial naïve Bayes, while slightly lower in accuracy at 88%, demonstrated consistent precision, recall, and F1-score metrics, at 0.89, 0.87, and 0.88, respectively. These results underscore the effectiveness of the random forest model and the competitive performance of the other classifiers in discerning fake news within the dataset. Table 3 shows the results of the textual dataset.

4.2. Result of the Proposed Model on a Multi-Modal Dataset

To assess our suggested model’s effectiveness in the field of false news detection, we carried out a thorough and rigorous comparison study throughout our trials with some baseline unimodal and multi-modal models. TextLSTM was the first LSTM network in the lineup. It was unique in that it had three layers: a bidirectional LSTM, a softmax layer, and a feed-forward layer. Interestingly, this model used 32-dimensional pre-trained word embeddings from Google, concentrating only on textual aspects in its methodology [5]. Textual (BERTweet) [2] was then studied, which uses the pre-trained BERTweet model as the only source of textual features. On the contrary, the visual model focused only on the visual components taken from the VGG-19 architecture. Lastly, Spotfake was a convergence of multi-modal traits that came from VGG-19 and BERT [23]. In comparison to these models, our proposed model for multi-modal data has the capability of textual analysis using BERT and can also analyze Images with the utilization of CNN layers.
The results are depicted in Table 4, showing the comparison between our suggested technique and the baseline using the MediaEval 2016 dataset. Interestingly, our suggested model outperformed several multi-modal models and baseline unimodal models. Our model significantly outperformed Spotfake, a model that included multi-modal features, delivering a 3.1% gain in accuracy. Table 4 provides the results of the multi-modal dataset. Table 5 shows the normalized confusion matrix, and Figure 6 shows the accuracy and loss curves of the proposed method Figure 7.

5. Conclusions

This study confronts the intricate challenge of misinformation detection, with a particular focus on fabricated news in today’s digital age. By advocating for a robust, systematic approach leveraging advanced machine learning and deep learning techniques, the research introduces a multi-modal architecture that combines natural language processing (NLP) for text analysis with computer vision for image and video verification. This framework’s capacity to analyze diverse forms of communication—written content, images, and videos—significantly enhances its ability to discern genuine news from misleading information. The model, evaluated using the MediaEval 2016 dataset, demonstrates an improved accuracy, precision, recall, and F1-score, reflecting its effectiveness in tackling contemporary media challenges. Future research could explore evaluating multilingual models to include text data in various languages and developing lightweight models for real-time fake news detection. These advancements could further enhance the practical applications and the framework’s adaptability to diverse linguistic and operational environments.
The study highlights the exceptional performance of the random forest model, achieving a 99% accuracy rate. However, it is essential to consider the model’s limitations, such as its potential overfitting to specific datasets and the computational resources required for deployment. Random forest may not always be the best choice in scenarios involving high-dimensional or sparse data or real-time processing needs, where algorithms like support vector machine or neural networks could perform better. The MediaEval 2016 dataset, while valuable, may not fully represent the diversity and complexity of global misinformation. Future work should incorporate additional datasets to ensure the framework’s robustness across various types of misinformation. Additionally, addressing trade-offs in model choice and evaluating scalability with increasing data volume and complexity is critical for optimizing performance.
The social impact of this research is significant. By improving the detection of fake news, the framework can contribute to increasing social trust and reducing the societal divisions caused by misinformation. Incorporating user feedback into the framework can enhance its usability and effectiveness in real-world settings. Optimizing resource use without compromising performance is crucial, especially for deploying the framework in practical applications. Future research should also explore methods for efficient resource management and strategies for scaling the model effectively. These considerations will help ensure that the framework remains practical and impactful, addressing the global challenge of disinformation comprehensively.

Author Contributions

Conceptualization, M.A.-a. and D.B.R.; Methodology, M.A.-a. and D.B.R.; Software, M.A.-a.; Validation, C.L.; Formal Analysis, M.A.-a.; Investigation, M.A.-a.; D.B.R. and C.L.; Resources, D.B.R. and C.L.; Data Curation, M.A.-a.; Writing—Original Draft Preparation, M.A.-a.; Writing—Review & Editing, M.A.-a. and D.B.R.; Visualization, M.A.-a.; Supervision, D.B.R.; Project Administration, D.B.R.; Funding Acquisition, C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the NSF under grant agreement DMS-2022448.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We would like to acknowledge unanimous reviewers.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, S. Cross-lingual Transfer Learning for Fake News Detector in a Low-Resource Language. arXiv 2022, arXiv:2208.12482. [Google Scholar]
  2. Duc Tuan, N.M.; Minh, P. Multimodal Fusion with BERT and Attention Mechanism for Fake News Detection. In Proceedings of the 2021 RIVF International Conference on Computing and Communication Technologies, RIVF 2021, Hanoi, Vietnam, 2–4 December 2021. [Google Scholar]
  3. Kaliyar, R.K.; Goswami, A.; Narang, P. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimed. Tools Appl. 2021, 80, 11765–11788. [Google Scholar] [CrossRef] [PubMed]
  4. Al-Alshaqi, M.; Rawat, D.B.; Liu, C. Emotion-Aware Fake News Detection on Social Media with BERT Embeddings. In Proceedings of the 2023 International Conference on Modeling & E-Information Research, Artificial Learning and Digital Applications (ICMERALDA), Karawang, Indonesia,, 24 November 2023; pp. 1–7. [Google Scholar]
  5. Sarkar, S.; Yang, F.; Mukherjee, A. Attending sentences to detect satirical fake news. In Proceedings of the 27th International Conference on Computational Linguistics, COLING 2018, Santa Fe, NM, USA, 20–26 August 2018; pp. 3371–3380. [Google Scholar]
  6. Castillo, C.; Mendoza, M.; Poblete, B. Information credibility on Twitter. In Proceedings of the 20th International Conference Companion on World Wide Web, WWW 2011, Hyderabad, India, 28 March–1 April 2011; Association for Computing Machinery: New York, NY, USA, 2011; pp. 675–684. [Google Scholar]
  7. Al-Alshaqi, M.; Rawat, D.B. Using Attention-Based Models to Automate Fake News Detection. In Proceedings of the 2023 Tenth International Conference on Social Networks Analysis, Management and Security (SNAMS), Abu Dhabi, United Arab Emirates, 21–24 November 2023; pp. 1–7. [Google Scholar]
  8. Rajapaksha, P.; Farahbakhsh, R.; Crespi, N. BERT, XLNet or RoBERTa: The Best Transfer Learning Model to Detect Clickbaits. IEEE Access 2021, 9, 154704–154716. [Google Scholar] [CrossRef]
  9. Huy, N.Q.; Duc Tuan, N.M.; Duong, N.M.; Minh, P. AimeLaw at ALQAC 2021: Enriching Neural Network Models with Legal-Domain Knowledge. In Proceedings of the 2021 13th International Conference on Knowledge and Systems Engineering (KSE), Bangkok, Thailand, 10–12 November 2021; pp. 1–6. [Google Scholar]
  10. Choudhury, D.; Acharjee, T. A novel approach to fake news detection in social networks using genetic algorithm applying machine learning classifiers. Multimed. Tools Appl. 2022, 82, 9029–9045. [Google Scholar] [CrossRef]
  11. Shah, P.; Kobti, Z. Multimodal fake news detection using a Cultural Algorithm with situational and normative knowledge. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation, CEC 2020, Glasgow, UK, 19–24 July 2020; pp. 1–7. [Google Scholar]
  12. Singh, V.K.; Ghosh, I.; Sonagara, D. Detecting fake news stories via multimodal analysis. J. Assoc. Inf. Sci. Technol. 2021, 72, 3–17. [Google Scholar] [CrossRef]
  13. Ying, L.; Yu, H.; Wang, J.; Ji, Y.; Qian, S. Multi-Level Multi-Modal Cross-Attention Network for Fake News Detection. IEEE Access 2021, 9, 132363–132373. [Google Scholar] [CrossRef]
  14. Song, C.; Ning, N.; Zhang, Y.; Wu, B. A multimodal fake news detection model based on crossmodal attention residual and multichannel convolutional neural networks. Inf. Process. Manag. 2021, 58, 102437. [Google Scholar] [CrossRef]
  15. Chen, Y.; Li, D.; Zhang, P.; Sui, J.; Lv, Q.; Tun, L.; Shang, L. Cross-modal Ambiguity Learning for Multimodal Fake News Detection. In Proceedings of the ACM Web Conference 2022, Lyon, France, 25–29 April 2022; Association for Computing Machinery: New York, NY, USA, 2022; pp. 1286–1295. [Google Scholar]
  16. Qian, S.; Wang, J.; Hu, J.; Fang, Q.; Xu, C. Hierarchical Multi-modal Contextual Attention Network for Fake News Detection. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2021, Virtual, 11–15 July 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 2380–2389. [Google Scholar]
  17. Raj, C.; Meel, P. ConvNet frameworks for multi-modal fake news detection. Appl. Intell. 2021, 51, 8132–8148. [Google Scholar] [CrossRef]
  18. Choi, H.; Ko, Y. Effective fake news video detection using domain knowledge and multimodal data fusion on YouTube. Pattern Recognit. Lett. 2022, 154, 44–52. [Google Scholar] [CrossRef]
  19. Chen, M.; Chu, X.; Subbalakshmi, K.P. MMCoVaR: Multimodal COVID-19 vaccine focused data repository for fake news detection and a baseline architecture for classification. In Proceedings of the 2021 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, ASONAM 2021, Virtual, 8–11 November 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 897–904. [Google Scholar]
  20. Xue, J.; Wang, Y.; Tian, Y.; Li, Y.; Shi, L.; Wei, L. Detecting fake news by exploring the consistency of multimodal data. Inf. Process. Manag. 2021, 58, 102610. [Google Scholar] [CrossRef] [PubMed]
  21. Chen, D.; Su, W.; Wu, P.; Hua, B. Joint multimodal sentiment analysis based on information relevance. Inf. Process. Manag. 2023, 60, 103193. [Google Scholar] [CrossRef]
  22. Qi, P.; Cao, J.; Li, R.; Liu, H. Improving Fake News Detection by Using an Entity-enhanced Framework to Fuse Diverse Multimodal Clues. In Proceedings of the 29th ACM International Conference on Multimedia, MM 2021, Virtual, 20–24 October 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 3340–3349. [Google Scholar]
  23. Singhal, S.; Shah, R.R.; Chakraborty, T.; Kumaraguru, P.; Satoh, S. SpotFake: A multi-modal framework for fake news detection. In Proceedings of the 2019 IEEE 5th International Conference on Multimedia Big Data, BigMM 2019, Singapore, 1–13 September 2019; pp. 39–47. [Google Scholar]
  24. Prakash, K.B. Fake News Detection. In Data Science Handbook; Scrivener Publishing LLC: Beverly, MA, USA, 2022. [Google Scholar]
  25. MKLab-ITI. Available online: https://github.com/MKLab-ITI/image-verification-corpus/tree/master (accessed on 12 December 2023).
Figure 1. Example images data from MediaEval 2016 dataset.
Figure 1. Example images data from MediaEval 2016 dataset.
Sensors 24 06062 g001
Figure 2. Samples from a multi-modal dataset.
Figure 2. Samples from a multi-modal dataset.
Sensors 24 06062 g002
Figure 3. Stepwise architecture of unimodal methodology.
Figure 3. Stepwise architecture of unimodal methodology.
Sensors 24 06062 g003
Figure 4. Proposed BERT-based CNN model.
Figure 4. Proposed BERT-based CNN model.
Sensors 24 06062 g004
Figure 5. Architecture diagram for multi-modal methodology (proposed model).
Figure 5. Architecture diagram for multi-modal methodology (proposed model).
Sensors 24 06062 g005
Figure 6. Accuracy of proposed model.
Figure 6. Accuracy of proposed model.
Sensors 24 06062 g006
Figure 7. Loss of proposed model.
Figure 7. Loss of proposed model.
Sensors 24 06062 g007
Table 1. Summary of fake news detection methodologies.
Table 1. Summary of fake news detection methodologies.
RefMethodologyDatasets UsedEvaluation Metrics
Singhal et al., 2019 [23]SpotFake framework, BERT, and VGG-19Twitter, WeiboOutperformed state-of-the-art models by 3.27% and 6.83%
Shah and Kobti, 2020 [11]Cultural algorithm for fake news detection; independent detection from text and imagesWeibo, TwitterAccuracy improvement of 9% on average
Singh, Ghosh, and Sonagara, 2021 [12]Logistic regression, linear discrimination analysis, quadratic discriminant analysis, k-nearest neighbors, naïve Bayes, support vector machine, classification and regression tree, and random forestKaggle Fake News DatasetVarious classifiers, 70–30 train/test split
Ying et al., 2021 [13]MMCN model leveraging BERT and ResNetWeibo, PHEMENot specified
Song et al., 2021 [14]CARMN framework with Crossmodal Attention Residual Network (CARN) and Multichannel CNN (MCN).Four real-world datasetsNot specified
Raj and Meel, 2021 [17]Coupled ConvNet architecture; comparative analysis with various CNN models.TI-CNN, EMERGENT, MICC-F220High accuracies, outperformed state-of-the-art methods
Chen, Chu, and Subbalakshmi, 2021 [19]Introduction of the multi-modal dataset on COVID-19 vaccine news; proposal of a fake news detection architectureCOVID-19 vaccine news datasetF-Score of 0.919, accuracy of 0.882
Qian et al., 2021 [16]HMCAN model utilizing BERT and ResNetWeibo, Twitter, PHEMEOutperformed state-of-the-art baselines
Qi et al., 2021 [22]Entity-enhanced multi-modal fusion frameworkNot specifiedOutperformed current state of the art
Xue et al., 2021 [20]MCNN framework with five subnetworksFour commonly used datasetsClear improvement in detection accuracy
Chen et al., 2022 [15]CAFÉ (Cross-modal Ambiguity-aware Fake news detection)Twitter, WeiboAccuracy improvement by 2.2–18.9% and 1.7–11.4%
Choi and Ko, 2022 [18]Integration of domain knowledge and multi-modal data fusion, Linear combinationFVC 2.0, FVC 3.0, Volunteer Annotated Video Dataset, Misleading Youtube Video CorpusF1-score of 0.93, outperformed the comparison models
Danlei Chen et al., 2023 [21]Relevance classifier methodNot SpecifiedNot Specified
Table 2. Summary of datasets for fake news and image verification.
Table 2. Summary of datasets for fake news and image verification.
DatasetContentSizeBalancePurposeFiles Included
ISOT Fake News DatasetTextual data from political statements, news, and press reportsOver 40,000 articlesBalanced (true and false articles)To study and analyze fake news characteristics and patternsN/A
Image Verification CorpusFake and real images shared on social media, notably TwitterEvolving datasetN/ATo evaluate online image verification techniques using user and tweet data’set_images.txt’: image_id, image_url, annotation, events; ’tweets_images.txt’: image_id, tweet validity, event origin, accompanying tweets; ’tweets_images_update.txt’: misleading tweets, erroneous information; ’tweets_event.txt’: deleted or deactivated account tweets
Table 3. Results of textual dataset.
Table 3. Results of textual dataset.
ModelAccuracyPrecisionRecallF1-Score
Random forest0.990.990.990.99
Support vector machine0.950.960.940.95
Logistic regression0.920.930.910.92
Multinomial naïve Bayes0.880.890.870.88
K-nearest neighbors0.910.920.900.91
Table 4. Results for multi-modal dataset.
Table 4. Results for multi-modal dataset.
ModelAccuracyPrecisionRecallF1-Score
TextLSTM (de Sarkar, Yang, and Mukherjee, 2018) [5]0.5120.5760.5430.564
Textual (BERTweet) (Duc Tuan and Quang Nhat Minh, 2021) [2]0.6560.6570.8200.733
Att-RNN (Huy et al., 2021) [9]0.6540.7390.6050.667
MVAE (Choi and Ko, 2022) [18]0.7350.7910.7090.748
Spotfake (Singhal et al., 2019) [23]0.7680.7410.8900.802
Proposed model0.9440.9020.9110.892
The bold formatting in the table is used to highlight the highest results for Accuracy, Precision, Recall, and F1-score.
Table 5. Normalized confusion matrix for proposed model.
Table 5. Normalized confusion matrix for proposed model.
PositiveNeutralNegative
Positive0.950.020.03
Neutral0.040.940.02
Negative0.030.030.94
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-alshaqi, M.; Rawat, D.B.; Liu, C. Ensemble Techniques for Robust Fake News Detection: Integrating Transformers, Natural Language Processing, and Machine Learning. Sensors 2024, 24, 6062. https://doi.org/10.3390/s24186062

AMA Style

Al-alshaqi M, Rawat DB, Liu C. Ensemble Techniques for Robust Fake News Detection: Integrating Transformers, Natural Language Processing, and Machine Learning. Sensors. 2024; 24(18):6062. https://doi.org/10.3390/s24186062

Chicago/Turabian Style

Al-alshaqi, Mohammed, Danda B. Rawat, and Chunmei Liu. 2024. "Ensemble Techniques for Robust Fake News Detection: Integrating Transformers, Natural Language Processing, and Machine Learning" Sensors 24, no. 18: 6062. https://doi.org/10.3390/s24186062

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop