This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

The Veracity Problem: Detecting False Information and its Propagation on Online Social Media Networks

Sarah Condran scondran@csu.edu.au 0000-0001-7813-2116 School of Computing, Mathematics and Engineering, Charles Sturt UniversityWagga WaggaNSWAustralia2650
Abstract.

Detecting false information on social media is critical in mitigating its negative societal impacts. To reduce the propagation of false information, automated detection provide scalable, unbiased, and cost-effective methods. However, there are three potential research areas identified which once solved improve detection. First, current AI-based solutions often provide a uni-dimensional analysis on a complex, multi-dimensional issue, with solutions differing based on the features used. Furthermore, these methods do not account for the temporal and dynamic changes observed within the document’s life cycle. Second, there has been little research on the detection of coordinated information campaigns and in understanding the intent of the actors and the campaign. Thirdly, there is a lack of consideration of cross-platform analysis, with existing datasets focusing on a single platform, such as X, and detection models designed for specific platform.

This work aims to develop methods for effective detection of false information and its propagation. To this end, firstly we aim to propose the creation of an ensemble multi-faceted framework that leverages multiple aspects of false information. Secondly, we propose a method to identify actors and their intent when working in coordination to manipulate a narrative. Thirdly, we aim to analyse the impact of cross-platform interactions on the propagation of false information via the creation of a new dataset.

false news, false information detection, social media networks
ccs: Information systems Social recommendation
Refer to caption
Figure 1. The Proposed Methodology

1. Introduction

The creation and spread of false information (aka fake news) is rapidly increasing, with online social media networks (OSMN) such as X111Formerly known as Twitter, Facebook, and Weibo contributing to its rise. False information can, often unbeknownst to them, manipulate how individuals responds to topics such as health, politics, and social life. One such example is a tweet in 2013 that claimed an explosion injured the US president, resulting in a loss of $130 billion in stock values (ElBoghdardy, 2013). Further, some research suggests false information is connected to election outcomes e.g. the 2016 US presidential elections (WTOE, 2016) and 2020 UK Brexit vote (Dennis et al., 2020).

OSMN enables anyone to access the latest information in a variety of formats (i.e. news articles, blogs) and sources (i.e. news outlets, public figures). The expansion of new formats and sources has decentralised the distribution of information, removing centralised control over its authenticity. This unregulated creation and spread of information places the onus of validating the truthfulness (or falsehood) of the information on the individual. However, an individual’s ability to identify falsehoods objectively is influenced by factors such as confirmation bias, which makes one trust and accept information that confirms their preexisting beliefs (Nickerson, 1998), and selective exposure, which is when one prefers to consume information that aligns with their beliefs (Freedman and Sears, 1965). External factors such as the bandwagon effect, which motivates one to perform an action because others are doing it (Leibenstein, 1950), and the validity effect, where one believes information after repeated exposure (Boehm, 1994) play a critical role.

A recent narrative which exemplified coordinated behaviors and the bandwagon effect is the Sydney Bondi Junction Stabbings (Nguyen and Workman, 2024). An image of the attacker circulated on social media, resulting in widespread misidentification of his heritage, religion, and identity. This false information initiated from small accounts, was amplified by verified accounts, and reported on by mainstream news. The coordinated behaviour exhibited both malicious and benign propagation patterns.

2. Literature Review

Detecting false information is crucial in alleviating the burden placed on individuals to verify the truthfulness of information, and the resulting consequences of this burden. There are three main streams of research key to this work which aims to address the problem of false information. These relate to (1) false information detection methods; (2) detection of coordinated behaviours; (3) cross-platform analysis of false information propagation.

False Information Detection Methods: To overcome the limitations of human-driven false information detection (e.g. scalability and bias (Tchakounté et al., 2020; Wallace et al., 2022)), AI-based decision support techniques have been developed. Most techniques fall into two types: content-based methods and context-based methods. Content-based methods are built using the information contained within the document, such as words and images (Zhang et al., 2024; Ghanem et al., 2021; Shu et al., 2019; Wang et al., 2024; Jin et al., 2022). While these methods can preemptively forestall the dissemination of false information, they are prone to adversarial manipulation of linguistic and stylistic features to evade detection (Jin et al., 2022). On the other hand, context-based methods use the information on the document’s propagation over OSMNs and the users who engage with it (Soga et al., 2024; Chowdhury et al., 2020; Bing et al., 2022; Min et al., 2022; Ruchansky et al., 2017). These methods are independent of linguistic and stylistic features and knowledge bases, but rely on information generated when a document is propagated on an OSMN, meaning false information is already spread, along with its related negative consequences. Although some hybrid models such as dEFEND (Shu et al., 2019) and CSI (Ruchansky et al., 2017) have been proposed, they use limited sources of information, making them prone to loss of reliability and effectiveness.

Detecting Coordination: There are two approaches to detect coordination, the first considers the similarity of actors (Dennis et al., 2020; Nizzoli et al., 2021). For example, (Nizzoli et al., 2021) proposed a network-based framework using iterative community detection to estimate the extent of coordination among actors. The second method considers the temporal synchronisation of actors. For example, (Zhang et al., 2021) incorporated a neural point process to identify synchronised behaviours. Moreover, most detection works treat the threshold of coordination as a distinct boundary within detection methods. However, it is a spectrum and varies for each community of actors (Giglietto et al., 2020). Methods that attempt to distinguish intent have found an inability to differentiate (Pacheco et al., 2021). While (Fazil and Abulaish, 2020) assumed all coordinated behaviours indicated inauthenticity, no other works were found that identify the intent of an actors when engaging in coordinated behaviours.

Existing Datasets: The available datasets for developing detection methods vary largely in terms of features for each document. For example, FakeNewsNet (Shu et al., 2020) provides both context and content data, while LIAR provides only content data. This means models developed on one dataset are often not transferable to other datasets. Additionally, most datasets include context data from one OSMN, making cross-platform analysis challenging. However (Ma et al., 2016) does include context data from both X and Weibo. The labelled datasets use manual labelling of documents and often require the manual identification of items corresponding to the document. Consequently, these datasets are static, and often out of date, leading to domain drift and diminished model performance in new domains.

3. Research Aim, Gaps and Questions

The aim of this research is to develop methods for the effective detection of false information on social media, and the analysis thereof, of its propagation.

The first research gap identified pertains to the detection of false information propagated on OSMNs. The existing methods consider the context and content features in isolation which provides a uni-dimensional view on a multi-dimensional concept. For example, in one dimension (e.g. content) an document may be labelled as true while in another (r.g. context), that same document may be labelled false. Further, the available data for a document and performance of a model can be impacted by the temporal and dynamic aspect of real world data. For example, consider a document observed at two times (d1.t1d_{1}.t_{1}, d1.t100d_{1}.t_{100}) where d1.t1d_{1}.t_{1} has 3 engagements and d1.t100d_{1}.t_{100} has 500 engagements. A model based on context data (e.g. number of user engagements) may assign a falsehood probability to d1.t1d_{1}.t_{1} based on significantly fewer data points compared to d1.t100d_{1}.t_{100}, thus affecting the quality of the prediction. This motivates the research question: RQ1 How can a model agnostic framework be developed to improve explainability and accuracy of existing false information detection techniques without incurring excessive computational overheads?

The second research gap identified is the detection of coordinated behaviours. The few works that detect coordination campaigns or actors are built on static datasets covering a limited range of topics and campaigns. However, no works look at identifying coordination on streaming data or without predefined campaigns to search for. Furthermore, no work has distinguished the intent of coordinated behaviours. This motivates the research question: RQ2 How can we identify actors working in coordination to maliciously amplify and manipulate the creation and propagation of information?

The third research gap is a lack of consideration of cross-platform interactions in detecting false information. There are no works to date which merge the context data from multiple OSMNs with the content data for each document to enhance the data available for analysis. Further, few works analyse whether the existing detection models are suitable for the different OSMNs. This motivates the research question: RQ3 How might the propagation of false information change over different online social media networks, and can existing detection methods account for the differences?

4. Methodology

The overarching aim of this work is addressed through the proposed methodology summarised in Figure 1, which outlines the approaches to solve the three research questions. The first research question RQ1 - Detection framework), will be solved through a three-part framework. This framework takes a document and its associated features as input, producing a probability of falsehood and an explanation as the output. (1) Base modeller: This component employees various false information detection base models (BMs) to generate probabilities of falsehood (p)(p). (2) Aggregator: This component combines the various probabilities (p)(p) into a final falsehood probability (Prob)(Prob). A novel dynamic aggregation method will be developed which accounts for the reliability of features the prediction is based on. For each instance of a document, a reliability weight (r)(r) will be assigned the the probability (p)(p). (3) Explainer: This component provides a tiered explanation of the contributions to the final probability (Prob)(Prob). Each tier offers greater detail of the BMs and the reliability factors.

The second research question RQ2 - Coordinated, solution is a two-part detection method which will first identify actors working in coordination and then determines the intent behind their behaviours. (1) Coordinated detector: This component identify the actors contributing to the a coordination campaign using isolated user characteristics (e.g. URLs, user mentions, hashtags) rather then explicit network structures (e.g. follower network, retweet information) which are often incomplete or unavailable. The pre-training and fine-tuning of models to reduce the required training dataset will be incorporated. (2) Intent detector: This component employs contrastive learning to identify an actors intent. That is, whether the actors harmful intent with malicious behaviours, or benign where an actor is engaging in coordinated activities without maliciousness for example as part of the bandwagon effect.

The third research question RQ3 - Dataset creation, solution is a three-part process to incorporate a new document into a multi-OSMN dataset. The dataset once created can enable the analysis of cross platform interactions. (1) Labeller: This segment uses an API to obtain a label for each document from a variety of manual factchecking sites. (2) Data Pull: This segment uses an API which collects all related context data from various OSMNs for each document. This novel method will enable the identification of items on a OSMN without prior manual identification. (3) Dataset: This segment will merge multiple context features and content features for one comprehensive data package for each document. Further, this segment employs entity linking to merge a user’s behaviour from multiple OMSNs.

5. Preliminary Results

Preliminary experimentation for the development of the Aggregator (RQ1) have returned favourable results for the use of a dynamic and adaptive weighting scheme. A brief outline of the experimental setup and results are presented below.

Datasets: Consistent with existing works, the benchmark datasets PolitiFact (Shu et al., 2020), GossipCop (Shu et al., 2020), and FakeHealth (Dai et al., 2020) are used to develop this aggregation method.

Base Models(BM): Three state-of-the-art base models are considered: (1) Fakeflow (FF) (Ghanem et al., 2021) utilises content features with Bidirectional Gated Recurrent Units (Bi-GRUs) to learn the flow of affective information throughout a document; (2) Publisher Credibility (PC) adapted from (Chowdhury et al., 2020) utilises document publisher features to train a probabilistic soft logic (PSL) model to calculate the credibility of a publisher; and (3) User Credibility (UC) adapted from (Chowdhury et al., 2020) utilises context features in the form of users who create an item to train a PSL model to calculate the credibility of that user.

Setup: The framework was built using python 3 running on Linux. And for all BMs the train-validation-test split was 70-10-20, with 10-fold cross validation.

Refer to caption
Figure 2. Reliability factor word_count

Aggregator: To develop a dynamic aggregation method based on the reliability of features, a set of reliability factors are produced for document did_{i}. The factors are defined on how they influence the features which contributes to a models prediction of falsehood. For instance, for context-based models such as FF, the features required are a set of words for each document did_{i}. Thus we make the assumption that a document did_{i} with a reasonable amount of words will provide FF with “sufficient” information to make a prediction of falsehood. This intuition is supported by the experiments shown in Figure 2 as measured by F1F1 scores for three datasets and different text lengths using FF.

Results: A detailed example (Figure 3) shows the working of the dynamic aggregation method where the weights are adaptively calculated for each instance of the document. Specifically, 3.a.iii illustrates the dynamic weightings assigned to the three base models for document d1d_{1} at time t2t_{2}. In Figure 3.b.iii the same document did_{i} at t168t_{168}, due to a higher level of user engagements due to the age of the document, has different weightings assigned to models which use user_history features. That is, UC had a higher contribution to the final prediction of falsehood.

Refer to caption
Figure 3. The Detection Framework

6. Conclusion

In this work, we identify three critical research gaps and propose solutions in the detection and mitigation of false information on social media. The proposed framework, once implemented, will improve false information detection by leveraging multiple base models in a dynamic manner. This will include developing novel methods for dynamic data aggregation based on reliability and providing hierarchical tiered explanation. The second proposed method involves a novel algorithm to identify both coordinated actors and their intent. Finally, the proposed dataset aims to provide a comprehensive view of a document and enable interrogation of cross-platform interactions.

Future research directions for the first research question include exploring additional reliability factors across a broader range of base models to determine whether there is a peak or a plateau in performance as reliability factors change. Another direction is to test the proposed framework on different online social media networks to assess its versatility and determine whether it is truly model-agnostic.

7. Acknowledgements

I would like to thank my supervisors Dr Michael Bewong, Dr Selasi Kwashie, Professor Md Zahidul Islam, and Associate Professor Irfan Altas for their support and guidance. This work was supported by Charles Sturt University.

References

  • (1)
  • Bing et al. (2022) Changsong Bing, Yirong Wu, Fangmin Dong, Shouzhi Xu, Xiaodi Liu, and Shuifa Sun. 2022. Dual Co-Attention-Based Multi-Feature Fusion Method for Rumor Detection. Information (Switzerland) 13 (1 2022), 25. Issue 1.
  • Boehm (1994) Lawrence E Boehm. 1994. The validity effect: A search for mediating variable. Personality and Social Psychology Bulletin 20 (1994), 285–293. Issue 3.
  • Chowdhury et al. (2020) Rajdipa Chowdhury, Sriram Srinivasan, and Lise Getoor. 2020. Joint Estimation of User and Publisher Credibility for Fake News Detection. International Conference on Information and Knowledge Management, Proceedings, 1993–1996.
  • Dai et al. (2020) Enyan Dai, Yiwei Sun, and Suhang Wang. 2020. Ginger Cannot Cure Cancer: Battling Fake Health News with a Comprehensive Data Repository. Proceedings of the International AAAI Conference on Web and Social Medi 14 (2020), 853–862.
  • Dennis et al. (2020) Assenmacher Dennis, Lena Clever, Janina Susanne Pohl, Heike Trautmann, and Christian Grimme. 2020. A Two-Phase Framework for Detecting Manipulation Campaigns in Social Media. Social Computing and Social Media. Design, Ethics, User Behavior, and Social Network Analysis 12194 (2020), 201–214.
  • ElBoghdardy (2013) Dina ElBoghdardy. 2013. Market quavers after fake AP tweet says Obama was hurt in White House explosions. The Washington Post (2013).
  • Fazil and Abulaish (2020) Mohd Fazil and Muhammad Abulaish. 2020. A socialbots analysis-driven graph-based approach for identifying coordinated campaigns in twitter. Journal of Intelligent and Fuzzy Systems 38 (2020), 3301–3305. Issue 3.
  • Freedman and Sears (1965) Jonathan L; Freedman and David O Sears. 1965. Selective Exposure. Advances in experimental social psychology 2 (1965), 57–97. https://www.sciencedirect.com/science/article/abs/pii/S0065260108601033
  • Ghanem et al. (2021) Bilal Ghanem, Simone Paolo Ponzetto, Paolo Rosso, and Francisco Rangel. 2021. FakeFlow: Fake News Detection by Modeling the Flow of Affective Information. arXiv preprint arXiv (1 2021). http://arxiv.org/abs/2101.09810
  • Giglietto et al. (2020) Fabio Giglietto, Nicola Righetti, Luca Rossi, and Giada Marino. 2020. It takes a village to manipulate the media: coordinated link sharing behavior during 2018 and 2019 Italian elections. Information, Communication & Society 23 (2020), 867–891. Issue 6.
  • Jin et al. (2022) Yiqiao Jin, Xiting Wang, Ruichao Yang, Yizhou Sun, Wei Wang, Hao Liao, and Xing Xie. 2022. Towards Fine-Grained Reasoning for Fake News Detection. Proceedings of the AAAI Conference on Artificial Intelligence 36 (2022). Issue 5. www.aaai.org
  • Leibenstein (1950) Harvey Leibenstein. 1950. Bandwagon, Snob, and Veblen Effects in the Theory of Consumers’ Demand. The quarterly journal of economics 64 (1950), 183–207. Issue 2.
  • Ma et al. (2016) Jing Ma, Wei Gao, Prasenjit Mitra, Sejeong Kwon, Bernard J Jansen, Kam-Fai Wong, and Meeyoung Cha. 2016. Detecting rumors from microblogs with recurrent neural networks. , 3818-3824 pages. https://ink.library.smu.edu.sg/sis_research
  • Min et al. (2022) Erxue Min, Yu Rong, Yatao Bian, Tingyang Xu, Peilin Zhao, Junzhou Huang, and Sophia Ananiadou. 2022. Divide-and-Conquer: Post-User Interaction Network for Fake News Detection on Social Media. WWW 2022 - Proceedings of the ACM Web Conference 2022, 1148–1158.
  • Nguyen and Workman (2024) Kevin Nguyen and Michael Workman. 2024. Benjamin Cohen was falsely accused of the Bondi Junction stabbings. Here’s how the lie spread around the world. https://www.abc.net.au/news/2024-04-15/how-misinformation-spread-after-bondi-junction-stabbing/103708210
  • Nickerson (1998) Raymond S Nickerson. 1998. Confirmation Bias: A Ubiquitous Phenomenon in Many Guises. Review of General Psychology 2 (1998), 175–220. Issue 2.
  • Nizzoli et al. (2021) Leonardo Nizzoli, Serena Tardelli, Marco Avvenuti, Stefano Cresci, and Maurizio Tesconi. 2021. Coordinated Behavior on Social Media in 2019 UK General Election. Proceedings of the International AAAI Conference on Web and Social Media. 15 (2021).
  • Pacheco et al. (2021) Diogo Pacheco, Pik-Mai Hui, Christopher Torres-Lugo, Bao Tran Truong, Alessandro Flammini, and Filippo Menczer. 2021. Uncovering Coordinated Networks on Social Media: Methods and Case Studies. www.aaai.org
  • Ruchansky et al. (2017) Natali Ruchansky, Sungyong Seo, and Yan Liu. 2017. CSI: A hybrid deep model for fake news detection. International Conference on Information and Knowledge Management, Proceedings Part F131841, 797–806.
  • Shu et al. (2019) Kai Shu, Limeng Cui, Suhang Wang, Dongwon Lee, and Huan Liu. 2019. Defend: Explainable fake news detection. Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 395–405.
  • Shu et al. (2020) Kai Shu, Deepak Mahudeswaran, Suhang Wang, Dongwon Lee, and Huan Liu. 2020. FakeNewsNet: A Data Repository with News Content, Social Context, and Spatiotemporal Information for Studying Fake News on Social Media. Big Data 8 (6 2020), 171–188. Issue 3.
  • Soga et al. (2024) Kayato Soga, Soh Yoshida, and Mitsuji Muneyasu. 2024. Exploiting stance similarity and graph neural networks for fake news detection. Pattern Recognition Letters 177 (1 2024), 26–32.
  • Tchakounté et al. (2020) Franklin Tchakounté, Ahmadou Faissal, Marcellin Atemkeng, and Achille Ntyam. 2020. A reliable weighting scheme for the aggregation of crowd intelligence to detect fake news. Information (Switzerland) 11 (6 2020). Issue 6.
  • Wallace et al. (2022) Shaun Wallace, Tianyuan Cai, Brendan Le, and Luis A. Leiva. 2022. Debiased Label Aggregation for Subjective Crowdsourcing Tasks. Conference on Human Factors in Computing Systems - Proceedings.
  • Wang et al. (2024) Bo Wang, Jing Ma, Hongzhan Lin, Zhiwei Yang, Ruichao Yang, Yuan Tian, and Yi Chang. 2024. Explainable Fake News Detection with Large Language Model via Defense Among Competing Wisdom. Proceedings of the ACM on Web Conference 2024, 2452–2463.
  • WTOE (2016) WTOE. 2016. Pope Francis Shocks World, Endorses Donald Trump for President, Releases Statement. https://web.archive.org/web/20161115024211/http://wtoe5news.com/us-election/pope-francis-shocks-world-endorses-donald-trump-for-president-releases-statement/
  • Zhang et al. (2024) Yuchen Zhang, Xiaoxiao Ma, Jia Wu, Jian Yang, and Hao Fan. 2024. Heterogeneous Subgraph Transformer for Fake News Detection. Proceedings of the ACM on Web Conference 2024, 1272–1282.
  • Zhang et al. (2021) Yizhou Zhang, Karishma Sharma, and Yan Liu. 2021. VigDet: Knowledge Informed Neural Temporal Point Process for Coordination Detection on Social Media. Conference on Neural Information Processing Systems. https://t.co/EV4SLEWXUv