Open Access
ARTICLE
Multimodal Social Media Fake News Detection Based on Similarity Inference and Adversarial Networks
1 College of Computer, Zhongyuan University of Technology, Zhengzhou, 450007, China
2 Henan Key Laboratory of Cyberspace Situation Awareness, Zhengzhou 450001, China
* Corresponding Author: Fangfang Shan. Email:
Computers, Materials & Continua 2024, 79(1), 581-605. https://doi.org/10.32604/cmc.2024.046202
Received 22 September 2023; Accepted 23 February 2024; Issue published 25 April 2024
Abstract
As social networks become increasingly complex, contemporary fake news often includes textual descriptions of events accompanied by corresponding images or videos. Fake news in multiple modalities is more likely to create a misleading perception among users. While early research primarily focused on text-based features for fake news detection mechanisms, there has been relatively limited exploration of learning shared representations in multimodal (text and visual) contexts. To address these limitations, this paper introduces a multimodal model for detecting fake news, which relies on similarity reasoning and adversarial networks. The model employs Bidirectional Encoder Representation from Transformers (BERT) and Text Convolutional Neural Network (Text-CNN) for extracting textual features while utilizing the pre-trained Visual Geometry Group 19-layer (VGG-19) to extract visual features. Subsequently, the model establishes similarity representations between the textual features extracted by Text-CNN and visual features through similarity learning and reasoning. Finally, these features are fused to enhance the accuracy of fake news detection, and adversarial networks have been employed to investigate the relationship between fake news and events. This paper validates the proposed model using publicly available multimodal datasets from Weibo and Twitter. Experimental results demonstrate that our proposed approach achieves superior performance on Twitter, with an accuracy of 86%, surpassing traditional unimodal modal models and existing multimodal models. In contrast, the overall better performance of our model on the Weibo dataset surpasses the benchmark models across multiple metrics. The application of similarity reasoning and adversarial networks in multimodal fake news detection significantly enhances detection effectiveness in this paper. However, current research is limited to the fusion of only text and image modalities. Future research directions should aim to further integrate features from additional modalities to comprehensively represent the multifaceted information of fake news.Keywords
With the rapid development of mobile Internet technology, the primary platform for accessing news has shifted from traditional paper-based media, such as newspapers, to social media platforms represented by Twitter and Weibo [1]. The real-time and convenient nature of social media enables people to quickly access and disseminate information. However, in the absence of effective supervision, the openness and low entry barriers of social media also facilitate the simultaneous spread of fake news [2]. Fake news is characterized by its low cost, rapid dissemination, and easy accessibility, which can lead to serious social issues, spark public opinion storms, and even manipulate public events, thereby undermining the credibility of social media. The rapid expansion of social media has become a breeding ground for the dissemination of fake news, where various fake news spread widely. Fake news not only has the potential to mislead the public but can also cause harm to individuals, organizations, and society. For instance, in 2019, Reuters reported that Hong Kong’s Chief Executive, Carrie Lam, submitted a report to Beijing recommending the consideration of the ‘five major demands’ of Hong Kong’s protesters. However, the report was allegedly rejected by Beijing. This fabricated news aimed to incite radical demonstrators, escalating the turmoil in Hong Kong, undermining the government’s relationship with the people, and disrupting social harmony in China. Additionally, in 2020, amid the global COVID-19 pandemic, it became one of the primary sources of fake news [3].
To foster a harmonious online environment and mitigate the negative impact of fake news, there is an urgent need for reliable methods and technologies to address the issue of false information dissemination. Consequently, the detection of multimodal fake news on social media has emerged as a prominent research focus in recent years. Generally, there is no clear definition of fake news. The Merriam-Webster Online Dictionary defines fake news as “a news report that is intentionally false or misleading”. Shu et al. [4] defined fake news as “false information that is intentionally misleading to readers and can be verified”. Ajao et al. [5] defined fake news as “anything that is circulated, shared, or spread that cannot be authenticated”. However, in academic research, fake news is usually defined as unverified or unconfirmed news. In this study, fake news is defined as intentionally misleading information that has been confirmed as false [6,7].
At present, the detection of fake news is primarily divided into two main directions: Unimodal modal detection and multimodal detection. Unimodal modal methods rely solely on text or image features for fake news detection. However, a news story embodies falseness in both text and image aspects, limiting the effectiveness of this approach in capturing the diverse features of fake news. In contrast, multimodal detection integrates features from various modalities, such as text and images, allowing for a more comprehensive understanding of fake news content.
Current multimodal fake news detection methods typically connect textual and visual features to obtain a unified multimodal feature representation. Nevertheless, these methods have yet to fully explore the similarity relationships between multimodal information, which is crucial for accurate fake news detection. Some fake news stories, aiming to garner clicks and widespread dissemination, often employ provocative image information that deviates significantly from the actual news text. For instance, Fig. 1 illustrates a fake news story about the U.S. government purchasing 30,000 guillotines, where the accompanying image features a historical painting depicting the beheading of Queen Marie Antoinette, creating a discordance with the textual content.
Addressing the semantic bias between textual and visual content, this paper introduces a multimodal social media fake news detection approach grounded in similarity reasoning and adversarial networks. Specifically tailored to bridge the research gap in understanding the similarity relationships within multimodal information, our method aims to comprehensively and accurately unveil the characteristics of fake news. By delving deeply into the correlations between text and images, our approach provides a nuanced and precise perspective for fake news detection. The method comprises five modules: (1) multimodal feature extractor; (2) similarity representation learning and reasoning; (3) multimodal feature fusion; (4) fake news detector; (5) event classifier. Designed to discern the falseness of news articles in terms of text, image, or “mismatch”. Our main contributions are summarized as follows:
(1) We conduct a comprehensive consideration of both local and global features in text and images. To extract text features, we integrate BERT and Text-CNN, introducing an attention mechanism post-Text-CNN to capture global text features. Simultaneously, leveraging the pre-trained VGG-19 on ImageNet, we extract local features from images. Following VGG-19, an attention mechanism is incorporated to capture global image features.
(2) We employ similarity representation learning and inference to infer the similarity between images and text, thereby recognizing more intricate matching patterns.
(3) By integrating event-based adversarial networks with multimodal networks, we not only capture features specific to particular events but also learn the associations between modal features and events.
(4) Extensive experiments on publicly available multimodal datasets, the research results demonstrate the outstanding performance of the proposed model in fake news detection tasks, particularly on the Twitter dataset. In comparison to traditional detection models, this model consistently achieves superior results across multiple metrics, effectively enhancing the accuracy and performance of fake news identification.
In this section, we present the research related to the proposed model for detecting fake news. Fake news detection has been a widely researched area due to its significance in maintaining the accuracy and reliability of public information. Existing fake news detection methods can be classified into two categories: Unimodal-based and multimodal-based fake news detection.
2.1 Unimodal-Based Fake News Detection
In the domain of unimodal-based fake news detection, the prevailing strategy revolves around leveraging textual information to ascertain the authenticity of news articles. This approach involves evaluating text content, syntactic structures, themes, and other factors to establish the veracity of the news. In the study conducted by Liu [8], the TF-IDF algorithm is applied to extract text features, and these features are utilized as inputs for a Support Vector Machine (SVM) to distinguish the authenticity of news. In contrast to certain intricate deep learning models, the TF-IDF algorithm exhibits superior computational efficiency, particularly when handling extensive textual data. Nevertheless, it falls short in capturing the contextual relationships between words, a crucial aspect for precise determinations of information authenticity in fake news detection. Amid the advancements in deep learning technology, models based on neural networks now can acquire more profound and abstract features, enabling end-to-end learning. In the context of fake news detection, Ma et al. [9] employed Recurrent Neural Networks (RNNs) to input all news texts associated with a specific event. The final hidden state of the RNN was subsequently leveraged to discern fake news at the event level. Although RNNs demonstrate proficiency in capturing contextual information within text sequences, they confront challenges in capturing long-distance dependencies when handling extensive sequences. Addressing the early detection challenge in fake news, Yu et al. [10] presented a Convolutional Neural Network (CNN)-based approach. This method initially groups news about the same event, transforming the textual content of each newsgroup into a document vector. Subsequently, CNN extracts text features from multiple document vectors for fake news detection. In comparison to models like RNN, which necessitate the consideration of sequence information, CNN’s efficiency in text processing is notable for not relying on sequence information. However, CNN is typically employed to capture local features, with limited capacity for processing global information. Ma et al. [11] further improved the model’s performance by introducing generative adversarial networks to enhance the learning of textual representations in fake news detection. Generative Adversarial Networks enable enhanced and nuanced learning of text representations. The interplay between the generator and discriminator allows the model to acquire text representations that are both distinctive and abstract. It is noteworthy, however, that the application of GANs often necessitates considerable computing resources, such as high-performance GPUs, and entails prolonged training time. While the success of text feature-based fake news detection is evident in certain aspects, the consideration of textual features alone lacks comprehensiveness, as fake news may be accompanied by misleading images or charts.
With the continuous advancement of image processing technology, the decreasing difficulty in forging false images poses greater challenges for the general public in discerning the authenticity of news and, consequently, presents a more significant challenge for fake news detection [12]. Therefore, scholars have increasingly focused on the detection of fake images. Mahmood et al. [13] proposed a method that combines the smooth wavelet transform and Discrete Cosine Transform (DCT) to detect and locate copy-move operations in images. This method comprehensively captures image features in the frequency domain, aiding in the more precise detection of copy-move operations. However, it encounters challenges in effectively addressing intricate textures or semantic information. Farooq et al. [14], combining Local Binary Pattern (LBP) features and texture features, introduced a method using a universal algorithm based on LBP to detect passive image forgery. While this method adeptly captures local texture information within images, its comprehension of global structure and context is constrained. In situations involving intricate forgery techniques, it lacks the requisite discriminative capacity. Peng et al. [15] identified forged images by examining resampling traces in the images. In contrast to approaches that require a reference image for comparison, this method does not necessitate obtaining a reference image beforehand, making it practical for real-world applications. However, it is typically employed for the overall assessment of whether an image is manipulated and does not offer detailed localization of forged regions. Zeng et al. [16] employed a hybrid deep-learning model to detect steganographic operations in JPEG images. The model utilizes techniques such as quantization and truncation to enhance its robustness and generalization capabilities. Quan et al. [17] developed a convolutional neural network-based universal model capable of classifying images into natural and computer-generated categories, making it suitable for various fake image detection scenarios.
Despite the promising results that unimodal modal methods can offer to some extent, data in social networks often involve multimodal information such as text and images. Unimodal modal methods fall short of adequately capturing and processing this diversity and complexity. As a result, researchers are beginning to explore the combination of text and images to address the limitations of unimodal methods.
2.2 Multimodal-Based Fake News Detection
Currently, deep neural networks (DNNs) excel in nonlinear representation [18], making them a prominent choice for many multimodal representation learning methods aimed at enhancing the capability of fake news detection. Jin et al. [19] proposed a deep learning-based approach capable of learning multimodal content and social information from news posts. They introduced an attention mechanism to fuse this information and obtain multimodal features, enhancing the model’s focus on crucial information and improving the weight allocation for different modal data, allowing for more effective utilization of multimodal information. In EANN, Wang et al. [20] employed an adversarial network with a multimodal feature extractor to learn invariant features of events, acquiring multimodal features for each news article to facilitate fake news detection. Learning invariant features of events through adversarial networks enhances the model’s generalization, yielding favorable results across diverse events. In MVAE, Khattar et al. [21] utilized a multimodal variational autoencoder for fake news identification, inputting various modal features of posts into a bimodal variational autoencoder to obtain multimodal feature representations. The introduction of variational autoencoders aids in learning latent representations of data, thereby enhancing the model’s expressive and generalization capabilities. Cui et al. [22] introduced an end-to-end deep embedding framework (SAME) for fake news detection. In this model, the emotions of post publishers serve as the basis for discerning fake news. By embedding emotional features with other characteristics through deep learning, the model distinguishes between real and fake news. Leveraging post publishers’ emotions as a basis for judging fake news provides additional information, contributing to a more comprehensive understanding of the authenticity of news. SpotFake [23] utilizes a pre-trained BERT [24] model for text feature extraction from news posts and employs a pre-trained VGG-19 model on ImageNet [25] for image feature extraction. The use of pre-trained BERT models enables the learning of rich text representations. SpotFake+ [26], an enhanced version of SpotFake, utilizes an improved BERT variant, XLNet [27] for text feature extraction.
Despite the current technological advancements propelling the development of multimodal fake news detection, there remains limited exploration and utilization of relationships between different modalities. This paper aims to address this gap by introducing similarity representation learning and inference, filling the research void in understanding the relationships between news text and visual information. By exploring multimodal information and similarity relationships, this study seeks to comprehensively understand and learn the representations of news articles. Additionally, the introduction of adversarial networks for learning invariant features of events aims to advance the frontier of research in multimodal fake news detection.
This paper introduces a multimodal social media fake news detection model based on similarity reasoning and adversarial networks. The model comprises a multimodal feature extractor, similarity representation learning and inference, multimodal feature fusion, a fake news detector, and an event classifier. Initially, the model independently preprocesses text and images, subsequently extracting feature representations. Textual features are extracted using the BERT and Text-CNN models. Subsequently, an attention mechanism is introduced post-Text-CNN to capture the global features of the text. For image feature representation, the pre-trained VGG-19 model is employed to acquire local image features, followed by applying a Self-Attention mechanism to these local features to derive global image feature representations. The model learns local similarity representations for text and image through the Text-CNN model for text local features and VGG-19 for image local features. Simultaneously, it acquires global similarity representations for text and image through the global features extracted from text and images, respectively. All the local similarity representations and global similarity representations serve as nodes within a graph, and we calculate the edges connecting these nodes. The graph undergoes similarity reasoning, which involves updating the nodes and edges iteratively over the N steps of reasoning. The output of the global nodes from the final step is used as the inferred similarity representation. This representation is then passed through a fully connected layer to generate the ultimate similarity score. Concatenate the textual features extracted by the BERT model with the text local features extracted by the Text-CNN model to form the textual feature representation. Subsequently, perform feature fusion by concatenating this textual feature representation with the local image features extracted by VGG-19 and the similarity representation resulting from text-image inference. The event classifier is a neural network model with a structure consisting of two fully connected layers, each equipped with corresponding activation functions. It employs clustering algorithms to accurately categorize newly emerged news information into specific event classes. The computation loss of an event is indicative of the similarity of the event distributions, with a larger loss denoting a greater similarity. The fake news detector utilizes the Softmax function as the activation function for the output layer. This function transforms the output of the fully connected layer into activation values representing the probability of fake news. Through this mechanism, the model can classify input features, discerning whether the news is genuine or fake. The framework of the multimodal social media fake news detection model based on similarity reasoning and adversarial networks (EANBS) is illustrated in Fig. 2.
3.2 Multimodal Feature Extractor
3.2.1 Textual Feature Extractor
This paper utilizes two critical text feature extractors: The Text-CNN and the BERT model. The Text-CNN allows for a focused analysis of localized perspectives and fine-grained features within the text, while the BERT model excels at extracting deep-seated semantic characteristics from the text. The synergy between these two approaches enables a more efficient extraction of textual semantic features.
In Text-CNN, a convolutional layer is utilized to extract features at a local level. By applying convolutional operations, the model can capture nuanced characteristics within the text. This, in turn, facilitates a more comprehensive understanding of localized information present in the text, ultimately refining the representation of text features. The standard configuration of the Text-CNN model consists of an embedding layer, a convolutional layer, a pooling layer, and a fully connected layer. The arrangement of the Text-CNN model is visually depicted in Fig. 3.
Local textual feature: (a) The Embedding layer: Words or phrases can be depicted as continuous, low-dimensional vectors. In the embedding layer, a sentence
where
(b) The convolutional layer: After converting text into word vectors using Word2Vec, we employ CNN’s convolutional operation to capture local features within the text. This is achieved by configuring various convolution kernel sizes
where
(c) The pooling layer: Following the convolutional layer, Text-CNN typically employs a max-pooling operation to reduce the dimensionality of the output generated by the convolutional operation. This helps in extracting essential features from the text effectively, as illustrated in Eq. (3).
(d) The fully connected layer: Finally, the vector representation obtained from the preceding pooling layer is combined through the fully connected layer to yield the local text representation
Global textual feature: Incorporating the attention mechanism can help the neural network model focus more on essential words. This allows the model to prioritize information from these keywords while ignoring less significant segments. As a result, the influence of unnecessary data is reduced, enhancing the model’s ability to extract crucial information. This, in turn, boosts the model’s efficiency and accuracy.
The Self-Attention mechanism, a specific case of the general attention mechanism, stands out for its ability to learn intrinsic textual correlations. When combined with CNN or RNN, it significantly enhances the model’s learning capabilities and improves the interpretability of the neural network. The computational steps of the Self-Attention mechanism can be represented as Eqs. (4)–(7).
where
BERT, a pre-trained language representation model, is constructed with stacked bidirectional Transformer [28] encoder structures. In contrast to traditional unidirectional models, BERT comprehensively captures contextual nuances in text, proving particularly advantageous for intricate textual scenarios encountered in tasks such as fake news detection. Leveraging extensive unsupervised pre-training, BERT acquires rich language representations encompassing a broad spectrum of semantic knowledge. The incorporation of these pre-trained weights into our task endows the model with the ability to benefit from a wealth of prior knowledge, thereby enhancing its capacity to articulate text features. Through a self-attention mechanism, BERT adeptly learns global contextual information, facilitating a deeper understanding of semantic relationships across the entire text. This results in more comprehensive word vector representations. Consequently, we integrate Text-CNN and BERT as the two core modules for text feature extraction. First, tokenize the input text, converting the tokens into word embeddings and positional embeddings. Subsequently, input these embeddings into the BERT model. The feature extraction process by BERT is outlined in Eqs. (8), (9).
where
To synergistically leverage the advantages of both Text-CNN and BERT models, this study adopts a concatenation approach, as illustrated in Eq. (10).
Specifically, the text features extracted by BERT are concatenated with the local features extracted by Text-CNN to form the final representation of text features. This fusion strategy aims to integrate global and local information, enabling a more comprehensive and multidimensional expression of features in news text.
In the context of fake news detection tasks, the significance of this fusion strategy lies in its ability to enhance the model’s comprehension of complex, multilayered information. By integrating both global and local features, the model becomes adept at distinguishing between authentic and fake news, as it can more comprehensively capture the semantic and structural information embedded in news text.
3.2.2 Visual Feature Extractor
In general, the brain learns and comprehends visual information much faster than textual information. Based on this insight, we have considered the visual features of news. Integrating the extracted image features with textual features enhances the feature representation, leading to a more comprehensive understanding and assessment of fake news. VGG-19 stands as a classical convolutional neural network architecture, exhibiting exceptional performance in image classification tasks. The hierarchical structure of VGG-19, with its relative depth, facilitates the extraction of abstract features from images, demonstrating excellent adaptability to the intricate patterns and structures potentially present in images associated with fake news. Pre-training on large-scale image datasets endows VGG-19 with a rich feature representation. Leveraging these pre-trained weights allows the model to learn universal visual features from diverse images, a significant advantage in the context of fake news image detection. Therefore, this paper employs the pre-trained VGG-19 model for extracting image features. To better preserve image information, resize the input image to 256 × 256 pixels. Perform a center crop to reduce it to 224 × 224 pixels. Preprocess the cropped image and feed it into the VGG-19 model. Add a fully connected layer at the end, adjusting the final image feature dimension to
where
where
3.3 Similarity Representation Learning and Reasoning
In the realm of fake news detection, traditional methodologies often focus solely on extracting text and image features, disregarding the inherent parallels between the two. However, comprehending this intrinsic similarity is paramount for accurate fake news prediction. Thus, the introduction of similarity representation learning and inference emerges as a crucial avenue. This approach facilitates the absorption of shared semantic nuances between text and image, consequently elevating the accuracy and robustness of fake news detection.
3.3.1 Similarity Representation Learning
Traditional methods utilize the cosine or Euclidean distance to represent the similarity between two feature vectors, which can capture the relevance to a certain degree while lacking detailed correspondence. In this paper, we follow [29] to compute a similarity representation, which is a similarity vector instead of a similarity scalar, to capture more detailed associations between feature representations from different modalities. The similarity function representation can be represented by Eqs. (17), (18).
where
Global Similarity Representation. The global similarity,
Local Similarity Representation. To calculate the local similarity representations between local features found in visual and textual observations, apply textual-to-visual attention [30], calculate the cosine similarity between the region feature
The cosine similarity matrix is then normalized and can be represented in Eq. (20).
Next, we calculate the similarity between the word feature
where
The local similarity representation
To propagate similarity information among possible alignments at both local and global levels, a similarity graph is constructed with global similarity and local similarity as nodes. The relationship between two similarities serves as edges in the graph, i.e.,
where
In this paper, the similarity is inferred through similarity propagation, linear transformation, and a non-linear activation function. Specifically, the similarity is first propagated as Eq. (24).
where
where
We iterate reasoning for N steps, and the output of the global node at the last step of iterative reasoning is utilized as the similarity representation of the reasoning and inputted into the fully connected layer to obtain the final similarity score as Eq. (26).
The representations of textual features, visual features, and the similarity representations between text and image are combined to create a multimodal feature representation
where
This module is designed to implement a neural network for detecting fake news, which is built upon a multimodal feature extractor. Textual features, visual features, and their similar representations are combined to create a multimodal feature representation
where
where
The event classifier consists of two fully connected layers, and its role is to classify K events. It evaluates the performance and robustness of the feature extractor by comparing the differences between the feature representation and the original events. This approach eliminates the strict dependence on specific events in the collected dataset and enables better generalization to unseen events, which in turn guides the training of the feature extractor. The event classifier is denoted as
The minimization of the loss function is expressed as Eq. (32).
By using the loss function
During training, minimizing loss
where the coefficient
Weibo Dataset: The Weibo Dataset, as presented by Jin et al. [32], has been extensively employed in numerous studies focused on multimodal fake news. This dataset spans confirmed fake news from May 2012 to January 2016, officially verified by Weibo, and real news validated by China’s authoritative news source, Xinhua News Agency. During the data preprocessing phase, a meticulous multi-step approach was applied to ensure dataset quality. Initially, duplicate images were removed to alleviate redundancy in the data. Subsequently, low-quality images were filtered out to ensure that all images in the dataset maintained high clarity and usability standards. Only data samples featuring both textual and image modalities were utilized to prevent distributional biases in unimodal and multimodal experiments, thereby enhancing the persuasiveness and credibility of the results. The dataset was partitioned into training, validation, and testing sets in a 7:1:2 ratio.
Twitter Dataset: The Twitter dataset utilized in this study is sourced from the MediaEval2015 dataset [33], encompassing both a training set and a test set. Each news item in the dataset consists of supplementary images/videos, text, and labels. In the data preprocessing stage, punctuation, numbers, special characters, and short words were removed from the tweets. Given the emphasis of our work on textual and image information, tweets with videos were excluded. Examples of images and corresponding text in the dataset are illustrated in Figs. 4 and 5. The specific distribution of each dataset is presented in Table 1.
The specific model parameters for this experiment are detailed in Table 2. Throughout this paper, several experiments were conducted to optimize these parameter settings. By making repeated adjustments and conducting multiple experiments, we were able to identify the optimal parameter configurations. The Text-CNN filter window selects a fixed set of sizes [1,2,3,4], which serves to reduce the hyperparameter search space of the model. This simplification alleviates the complexity and tedium associated with the hyperparameter tuning process. By capturing features at various levels, the model becomes better suited to adapt to text inputs of differing lengths and complexities. This adaptation contributes to the enhancement of classification accuracy and robustness across various scenarios.
We used the traditional performance metrics namely accuracy, recall, precision, and F1 score to evaluate the proposed model. Here is a brief explanation of these metrics.
where True positive (TP): Fake news forecasted as fake; True negative (TN): Real news forecasted as real; False positive (FP): Real news forecasted as fake; False negative (FN): Fake news forecasted as real.
To validate the effectiveness of the proposed model, this study selects two categories of baseline models: Unimodal models and multimodal models.
The unimodal model employs either textual or visual information alone to detect fake news. Thus, this paper proposes the following two sample baselines:
•Textual. The model exclusively relies on the textual content within the post for post-classification, utilizing a pre-trained 300-dimensional Word2Vec model from Sogou Labs to represent word vectors. Firstly, Text-CNN is employed for text feature extraction, transforming the textual information into a feature representation. Subsequently, the extracted text features undergo processing through a 32-dimensional fully connected layer to accomplish the task of detecting fake news.
•Visual. The model relies exclusively on the image information within the post for post-classification, starting by extracting the image feature, denoted as
Multimodal approaches utilize information from multiple modalities to classify fake news.
•VQA [34]. Visual Question and Answer (VQA) is a system that provides answers to questions based on a given image. While the original VQA model was designed for a multiclassification task, the primary focus of this paper is on binary classification.
•NeuralTalk [35]. NeuralTalk is a model designed to generate subtitles for a given image. It obtains potential representations of the text sequence by averaging the output of the RNN (Recurrent Neural Network) at each time step. These latent representations are then passed to a fully connected layer for the prediction of fake news. Both the LSTM and the fully connected layer have a hidden size of 32.
•att-RNN [32]. att-RNN employs an attention mechanism that combines textual, visual, and social contextual features. It utilizes Long Short-Term Memory (LSTM) networks to extract textual features and integrates them with visual features through a cross-modal attention mechanism.
•EANN [20]. The Event Adversarial Neural Networks (EANN) consist of three components: The multimodal feature extractor, the fake news detector, and the event discriminator. The multimodal feature extractor extracts textual and visual features from posts and collaborates with the fake news detector to learn distinctive representations for fake news detection. The event discriminator is responsible for removing specific features related to events. All parameters used in training this model remain consistent with those of the original model.
•MVAE [21]. This methodology employs an encoding-decoding paradigm to capture shared representations encompassing both visual and textual modalities, to detect fake news. Through the training of a multimodal variational autoencoder, the approach involves the concatenation of text and visual features to derive multimodal representations. These representations undergo decoding, guided by reconstruction loss, to revert them to their original modalities. The resulting multimodal representations are strategically harnessed for the discrimination of fake news.
•BDANN [36]. Textual features in BDANN are extracted using a pre-trained BERT model, while visual features are obtained through a pre-trained VGG-19 model. Dependency on specific events is mitigated by incorporating a domain classifier.
•Roberta+CNN [37]. This framework incorporates a dedicated convolutional neural network model for image analysis and a sentence transformer for textual analysis. Features extracted from visual and textual modalities are embedded through dense layers, ultimately converging to predict deceptive imagery.
•MEAN [38]. This approach comprises two integral components: A multimodal generator and a dual discriminator. The multimodal generator is instrumental in extracting latent discriminative feature representations for both text and image modalities. For each modality, a decoder is employed to mitigate information loss during the generation process. The dual discriminator consists of a modality discriminator and an event discriminator. These discriminators are designed to classify features based on either modality or event, with network training guided by an adversarial scheme.
This paper employs conventional evaluation metrics for binary classification to assess the model’s performance. These metrics include accuracy, precision, recall, and F1 score. The experimental comparison results are presented in Table 3 and Fig. 6.
According to Table 3, on the Weibo dataset, the textual modality demonstrates a significant advantage in the task of detecting fake news compared to the visual modality. Text is more effective in conveying the core content of events, and its embedded semantic information directly facilitates the identification of fake news. In contrast, although the visual modality provides some visual information, its expressive capability is relatively limited, making it challenging to offer as rich semantic information as text. Therefore, in the fake news detection task, text modality proves to be more effective than the visual modality, better distinguishing between genuine and fake news information. While the unimodal model exhibits some effectiveness in fake news detection, their performance remains inadequate compared to multimodal models, further confirming the excellence of multimodal fake news detection methods. Multimodal fake news detection methods can fully leverage diverse information sources, such as text and images, to obtain more comprehensive and enriched feature representations. Text and images complement each other in expressing information, and through the effective fusion of these modalities’ features, the model can enhance its ability to identify fake news.
In the realm of multimodal fake news detection models, the MVAE exhibits superior performance by leveraging a multimodal variational autoencoder. Outperforming other models such as VAQ, NeuralTalk, and att-RNN, the EANN and BDANN models introduce Event Adversarial Neural Networks, achieving superior results in fake news detection. Through the incorporation of Event Adversarial Neural Networks, the EANN and BDANN models excel in learning and applying common features among events, thereby demonstrating robust performance in fake news detection tasks. Event Adversarial Networks contribute to the model’s ability to learn more general and transferable feature representations, mitigating reliance on specific events and enhancing both robustness and generalization. In comparison to the EANN and BDANN models, MEAN exhibits outstanding performance by learning both modality-invariant and event-invariant features through dual discriminators. On the Twitter dataset, the model’s performance is similar to that on the Weibo dataset.
The proposed EANBS fake news detection model demonstrates superior performance across various metrics on both Weibo and Twitter datasets compared to the contrastive models. On the Twitter dataset, the accuracy and precision of fake news detection surpass the best results in the comparative methods by 0.07% and 2.9%. On the Weibo dataset, the recall and precision of fake news detection surpass the best results in the comparative methods by 4% and 1%, respectively. Leveraging similarity representation learning to capture relationships between different modalities, the model comprehensively understands news content, thereby improving the accuracy and robustness of fake news detection. The introduction of adversarial networks further enhances the model’s performance, aiding in learning more robust feature representations and eliminating dependence on specific events. This improvement increases the model’s generalization capability towards unseen events, effectively identifying fake news and enhancing the overall quality of news on social media.
4.5 Analysis of Ablation Experiments
To verify the significance of the model components outlined in this paper, we created several model variants. These variants primarily fall into three types: EANBS-SIM, which eliminates similarity representation learning and reasoning; EANBS-BERT, which removes the BERT component; and EANBS-GAN, which excludes adversarial neural networks.
(1) Eliminate Similarity Representation Learning and Reasoning. The textual modality employs Text-CNN and BERT models to extract textual features, while the pre-trained VGG-19 model is used to extract visual features for the visual modality. Subsequently, the resulting feature vectors from both modalities are merged and used as input to both the event classifier and fake news detector.
(2) Eliminate BERT Component. The Text-CNN model is employed for the textual modality, while the pre-trained VGG-19 model is used for the visual modality. The similarity representation learning and reasoning module calculates the similarity between the extracted textual and visual features. Finally, textual features, visual features, and their similarity are concatenated as input for both the event classifier and the fake news detector.
(3) Eliminates the Adversarial Neural Networks. For the textual modality, both Text-CNN and BERT models are employed to extract textual features, while the pre-trained VGG-19 model is used for visual feature extraction. Subsequently, the similarity representation learning and reasoning module computes the similarity between the textual features obtained from Text-CNN and the visual features obtained from the pre-trained VGG-19. The resulting textual features, visual features, and their similarity are then combined and input into a fully connected layer with SoftMax for classification.
The results of ablation experiments are presented in Table 4, indicating that the removal of any component of the model results in a noticeable decrease in classification accuracy. This underscores the effectiveness of each model component in the experiments. The introduction of BERT provides our model with enhanced semantic understanding and more robust feature representation capabilities, aiding in capturing crucial features within fake news text. Utilizing the BERT model allows for more accurate comprehension of semantic relationships in text, enabling the identification of hidden information and effective discrimination between fake and genuine news. The similarity representation learning and inference model proves effective in capturing the relationship, i.e., similarity, between news text and image information in fake news detection. Through this model, common features between text and images are learned, facilitating inference regarding their degree of similarity. Such a model enables a more comprehensive understanding of news events, leading to more accurate assessments of news authenticity.
While the role of adversarial neural network models in fake news detection may be relatively weaker, comprehensive experiments combining BERT, similarity representation learning and inference models, and adversarial neural network models demonstrate superior performance compared to the ablation experiment results. This suggests the crucial complementary and synergistic effects among these models in enhancing fake news detection. By synergistically leveraging BERT’s semantic understanding, the association-capturing capabilities of similarity representation learning and inference models, and the feature fusion optimization of adversarial neural network models, we achieve outstanding results in fake news detection on two datasets. This multimodal combination approach offers new perspectives and technical means for fake news detection research, enhancing model performance and reliability.
To further assess the efficacy of the event classifier, we visualize the textual feature representations obtained by the model in this paper, both with and without adversarial neural networks, using the Weibo test dataset. Fig. 7 depicts the visualization of textual representations, where red dots represent labeled features of fake news, and blue dots represent labeled features of real news. Based on the feature distribution, it can be observed that the model without the adversarial neural networks can learn distinguishable features. However, the learned features are still intertwined when compared to the feature representations learned by the model in this paper. This also indicates that during the training phase, the event classifier endeavors to eliminate dependencies between feature representations and specific events. Through the minimax game, the multimodal feature extractor can acquire invariant feature representations of different events, enabling the learned generic features to be employed for transfer learning to discern the authenticity of breaking fake news in sudden events. This enhances the model’s transferability and its ability to generalize to new events, ultimately improving the performance of fake news detection.
To examine the influence of model parameters on performance, we conducted several experimental sets on two datasets, Weibo and Twitter, followed by a comparative analysis of their outcomes.
These experiments primarily focused on variations in the learning rate and different settings of the optimizer. Figs. 8 and 9 demonstrate the comparison of the experimental results.
(1) Effect of Learning Rate: Fig. 8 illustrates the impact of various learning rate values on the performance of the proposed model on two datasets, Weibo and Twitter. As shown in the figure, the model achieved its highest accuracy and F1 score on two datasets when the learning rate was set to 0.001. Therefore, in our experiments, we chose to set the learning rate to 0.001.
(2) Impact of Optimizers: Fig. 9 illustrates the impact of different optimizers on the performance of the proposed model on the microblogging dataset. In terms of model prediction accuracy, the use of the SGD optimizer results in a classification accuracy of around 0.85 after 50 rounds of training, while the Adam optimizer achieves a classification accuracy close to 1.0. Therefore, in our experiments, we opted to use the Adam optimizer for model training. The optimizer guides the various parameters of the loss function during the backpropagation process to update in the correct direction with an appropriate magnitude, continuously approaching the global minimum. In deep learning, selecting an appropriate optimizer can significantly improve both training efficiency and accuracy.
To verify the convergence of the proposed model, we selected parameters α = 10 and β = 0.75 for the experiments. Fig. 10 illustrates the changes in the loss function on two datasets, Weibo and Twitter. In the initial stages of training, the loss rapidly decreases, indicating that the choice of learning rate is appropriate, and the model has entered the gradient descent process. As training progresses, the loss function gradually stabilizes, indicating that the model has reached a certain state of equilibrium, demonstrating that the model has undergone effective learning.
To further illustrate the performance of our proposed method, we collected and analyzed some failure cases. Figs. 11a and 11b depict two instances where our proposed approach failed to detect fake news. In Fig. 11a, the excessive use of exclamation marks in the post’s text, along with a presentation format not typical of traditional news media, is likely to impact the model’s discriminative ability on the data. In Fig. 11b, the post’s text is exceptionally short, resulting in suboptimal performance of the proposed similarity reasoning and adversarial network models. Additionally, to explore enhanced performance, we intend to address these limitations in future work.
The spread of fake news not only damages the reputation of news organizations but also pollutes the online information landscape, posing a significant threat to the growth of social media. This paper aims to address the issue of detecting fake news in social media by constructing a multimodal social media detection model that utilizes similarity reasoning and adversarial networks. This is achieved by examining the similarities between the textual and visual components of fake news content. Through the analysis of the results from various feature experiments, we have discovered that the similarity features between the textual and visual aspects are highly effective in distinguishing between fake and real news. By implementing a game between the feature extractor and classifier, the model can acquire event-invariant representations by eliminating specific event-related features, thereby reducing the strong dependence on specific events in fake news. Moreover, the model can detect shared event characteristics in fake news, which improves its ability for feature transfer and the identification of fake news in emerging events. We have conducted numerous experiments on two datasets to demonstrate the model’s effectiveness. However, it is worth noting that we have primarily focused on the correlation between text and image in our model. As a result, we intend to explore the integration of video information features in our future research.
Acknowledgement: The authors would like to thank the anonymous referees for their valuable comments and helpful suggestions.
Funding Statement: This paper is supported by the National Natural Science Foundation of China (No. 62302540), with author F.F.S. For more information, please visit their website at https://www.nsfc.gov.cn/. Additionally, it is also funded by the Open Foundation of Henan Key Laboratory of Cyberspace Situation Awareness (No. HNTS2022020), where F.F.S is an author. Further details can be found at http://xt.hnkjt.gov.cn/data/pingtai/. The research is also supported by the Natural Science Foundation of Henan Province Youth Science Fund Project (No. 232300420422), and for more information, you can visit https://kjt.henan.gov.cn/2022/09-02/2599082.html. Lastly, it receives funding from the Natural Science Foundation of Zhongyuan University of Technology (No. K2023QN018), where F.F.S is an author. You can find more information at https://www.zut.edu.cn/.
Author Contributions: The authors confirm contribution to the paper as follows: Study conception and design: Fangfang Shan, Huifang Sun; data collection: Huifang Sun, Mengyi Wang; analysis and interpretation of results: Huifang Sun; draft manuscript preparation: Huifang Sun. All authors reviewed the results and approved the final version of the manuscript.
Availability of Data and Materials: The Weibo data used to support the findings of this study have been deposited on the website: https://drive.google.com/file/d/14VQ7EWPiFeGzxp3XC2DeEHi-BEisDINn/view?usp=sharing. The Twitter data used to support the findings of this study have been deposited on the website: https://github.com/MKLab-ITI/image-verification-corpus.
Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.
References
1. J. Wang, Y. Liu, and S. J. Ji, “Matrix factorized bilinear pooling for multi-modal fusion fake news detection (In Chinese),” Appl. Res. Comput., vol. 39, no. 10, pp. 2968–2973+2978, 2022. [Google Scholar]
2. P. Qi, J. Cao, and Q. Sheng, “Semantics-enhanced multi—modal fake news detection,” J. Comput. Res. Dev., vol. 58, no. 7, pp. 1456–1465, 2021. [Google Scholar]
3. J. Z. Zhang, J. Cao, and D. Radcliffe, “The impact of COVID-19 on journalism in emerging economies and the global south,” Youth Journalist, vol. 699, no. 7, pp. 99–100, 2021. [Google Scholar]
4. K. Shu, A. Sliva, S. Wang, J. L. Tang, and H. Liu, “Fake news detection on social media: A data mining perspective,” ACM SIGKDD Explor. Newsletter, vol. 19, no. 1, pp. 22–36, 2017. doi: 10.1145/3137597.3137600. [Google Scholar] [CrossRef]
5. O. Ajao, D. Bhowmik, and S. Zargari, “Sentiment aware fake news detection on online social networks,” in Int. Conf. Acoust., Speech Signal Process. (ICASSP), Brighton, UK, 2019, pp. 2507–2511. [Google Scholar]
6. M. A. Alonso, D. Vilares, C. Gómez-Rodríguez, and J. Vilares, “Sentiment analysis for fake news detection,” Electron., vol. 10, no. 11, pp. 1348, 2021. doi: 10.3390/electronics10111348. [Google Scholar] [CrossRef]
7. N. Islam et al., “Ternion: An autonomous model for fake news detection,” Appl. Sci., vol. 11, no. 19, pp. 9292, 2021. doi: 10.3390/app11199292. [Google Scholar] [CrossRef]
8. X. C. Liu, “Research on fake news detection based on machine learning algorithms,” Inf. Technol. Inf., no. 9, pp. 237–239, 2021. [Google Scholar]
9. J. Ma et al., “Detecting rumors from microblogs with recurrent neural networks,” in Twenty-Fifth Int. Joint Conf. Artif. Intell., New York, NY, USA, 2016, pp. 3818–3824. [Google Scholar]
10. F. Yu, Q. Liu, S. Wu, L. Wang, and T. Tan, “A convolutional approach for misinformation identification,” in Proc. Twenty-Sixth Int. Joint Conf. Artif. Intell., Melbourne, Australia, 2017, pp. 3901–3907. [Google Scholar]
11. J. Ma, W. Gao, and K. F. Wong, “Detect rumors on twitter by promoting information campaigns with generative adversarial learning,” in The World Wide Web Conf., San Francisco, USA, 2019, pp. 3049–3055. [Google Scholar]
12. Z. Jin, J. Cao, Y. Zhang, J. Zhou, and Q. Tian, “Novel visual and statistical image features for microblogs news verification,” IEEE Trans. Multim., vol. 19, no. 3, pp. 598–608, 2016. doi: 10.1109/TMM.2016.2617078. [Google Scholar] [CrossRef]
13. T. Mahmood, Z. Mehmood, M. Shah, and T. Saba, “A robust technique for copy-move forgery detection and localization in digital images via stationary wavelet and discrete cosine transform,” J. Vis. Commun. Image Rep., vol. 53, no. 5, pp. 202–214, 2018. doi: 10.1016/j.jvcir.2018.03.015. [Google Scholar] [CrossRef]
14. S. Farooq, M. H. Yousaf, and F. Hussain, “A generic passive image forgery detection scheme using local binary pattern with rich models,” Computers & Electrical Engineering, vol. 62, pp. 459–472, 2017. [Google Scholar]
15. A. Peng, Y. Wu, and X. Kang, “Revealing traces of image resampling and resampling antiforensics,” Adv. Multimed., vol. 2017, pp. 7130491, 2017. doi: 10.1155/2017/7130491. [Google Scholar] [CrossRef]
16. J. Zeng, S. Tan, B. Li, and J. Huang, “Large-scale JPEG image steganalysis using hybrid deep-learning framework,” IEEE Trans. Inf. Forens. Secur., vol. 13, no. 5, pp. 1200–1214, 2017. doi: 10.1109/TIFS.2017.2779446. [Google Scholar] [CrossRef]
17. W. Quan, K. Wang, D. M. Yan, and X. Zhang, “Distinguishing between natural and computer-generated images using convolutional neural networks,” IEEE Trans. Inf. Forens. Secur., vol. 13, no. 11, pp. 2772–2787, 2018. doi: 10.1109/TIFS.2018.2834147. [Google Scholar] [CrossRef]
18. Z. Zhao et al., “Social-aware movie recommendation via multimodal network learning,” IEEE Trans. Multimed., vol. 20, no. 2, pp. 430–440, 2017. doi: 10.1109/TMM.2017.2740022. [Google Scholar] [CrossRef]
19. Z. W. Jin, J. Cao, B. Wang, R. Wang, and Y. D. Zhang, “Rumor detection on social media with multimodal feature fusion,” J. Nanjing Univ. Inf. Sci. Technol., vol. 9, no. 6, pp. 583–592, 2017. [Google Scholar]
20. Y. Q. Wang et al., “EANN: Event adversarial neural networks for multi-modal fake news detection,” in Proc. 24th ACM SIGKDD Int. Conf. Knowl. Discov. Data Min., London, UK, 2018, pp. 849–857. [Google Scholar]
21. D. Khattar, J. S. Goud, M. Gupta, and V. Varma, “MVAE: Multimodal variational autoencoder for fake news detection,” in The World Wide Web Conf., San Francisco, CA, USA, 2019, pp. 2915–2921. [Google Scholar]
22. L. Cui, S. Wang, and D. Lee, “SAME: Sentiment-aware multi-modal embedding for detecting fake news,” in Proc. 2019 IEEE/ACM Int. Conf. Adv. Soc. Netw. Anal. Min., Vancouver, Canada, 2019, pp. 41–48. [Google Scholar]
23. S. Singhal, R. R. Shah, T. Chakraborty, P. Kumaraguru, and S. I. Satoh, “Spotfake: A multi-modal framework for fake news detection,” in 2019 IEEE Fifth Int. Conf. Multimed. Big Data, Changsha, China, 2018, pp. 39–47. [Google Scholar]
24. J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” arXiv preprint arXiv:1810.04805, 2018. [Google Scholar]
25. J. Deng et al., “ImageNet: A large-scale hierarchical image database,” in 2009 IEEE Conf. Comput. Vis. Pattern Recogn., Miami, Florida, USA, 2009, pp. 248–255. [Google Scholar]
26. S. Singhal, A. Kabra, M. Sharma, R. R. Shah, T. Chakraborty and P. Kumaraguru, “SpotFake+: A multimodal framework for fake news detection via transfer learning (student abstract),” in Proc. AAAI Conf. Artif. Intell., vol. 34, no. 10, pp. 13915–13916, 2020. doi: 10.1609/aaai.v34i10.7230. [Google Scholar] [CrossRef]
27. Z. L. Yang, Z. H. Dai, Y. M. Yang, J. Carbonell, R. Salakhutdinov and Q. V. Le, “XLNet: Generalized autoregressive pretraining for language understanding,” in Adv. Neural Inf. Process. Syst., Vancouver, BC, Canada, 2019, pp. 5753–5763. [Google Scholar]
28. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones and A. N. Gomez, “Attention is all you need,” in Adv. Neural Inf. Process. Syst., Long Beach, CA, USA, vol. 30, 2017, pp. 5998–6008. [Google Scholar]
29. H. Diao, Y. Zhang, L. Ma, and H. Lu, “Similarity reasoning and filtration for image-text matching,” in Proc. AAAI Conf. Artif. Intell., vol. 35, no. 2, pp. 1218–1226, 2021. doi: 10.1609/aaai.v35i2.16209. [Google Scholar] [CrossRef]
30. K. H. Lee, X. Chen, G. Hua, H. Hu, and X. He, “Stacked cross attention for image-text matching,” in Proc. Eur. Conf. Comput. Vis. (ECCV), Munich, Germany, 2018, pp. 201–216. [Google Scholar]
31. Z. Kuang et al., “Fashion retrieval via graph reasoning networks on a similarity pyramid,” in Proc. IEEE/CVF Int. Conf. Comput. Vis., Seoul, South Korea, 2019, pp. 3066–3075. [Google Scholar]
32. Z. Jin, J. Cao, H. Guo, Y. Zhang, and J. Luo, “Multimodal fusion with recurrent neural networks for rumor detection on microblogs,” in Proc. 25th ACM Int. Conf. Multimed., New York, USA, 2017, pp. 795–816. [Google Scholar]
33. C. Boididou et al., “Verifying multimedia use at mediaeval 2015,” in MediaEval 2015 Workshop, Wurzen, Germany, 2015. [Google Scholar]
34. S. Antol et al., “VQA: Visual question answering,” in Proc. IEEE Int. Conf. Comput. Vis., Santiago, Chile, 2015, pp. 2425–2433. [Google Scholar]
35. O. Vinyals, A. Toshev, S. Bengio, and D. Erhan, “Show and tell: A neural image caption generator,” in Proc. IEEE Conf. Comput. Vis. Pattern Recogn., Boston, USA, 2015, pp. 3156–3164. [Google Scholar]
36. T. Zhang et al., “BDANN: BERT-based domain adaptation neural network for multi-modal fake news detection,” in 2020 Int. Joint Conf. Neural Netw. (IJCNN), Glasgow, UK, 2020, pp. 1–8. [Google Scholar]
37. B. Singh and D. K. Sharma, “Predicting image credibility in fake news over social media using multi-modal approach,” Neural Comput. Appl., vol. 34, no. 24, pp. 21503–21517, 2022. doi: 10.1007/s00521-021-06086-4 [Google Scholar] [PubMed] [CrossRef]
38. P. Wei, F. Wu, Y. Sun, H. Zhou, and X. Y. Jing, “Modality and event adversarial networks for multi-modal fake news detection,” IEEE Signal Process. Letters, vol. 29, no. 2, pp. 1382–1386, 2022. doi: 10.1109/LSP.2022.3181893. [Google Scholar] [CrossRef]
Cite This Article
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.