Misinformation on social media spreads rapidly, often with serious consequences for public health and politics, yet existing detection tools struggle to explain why certain posts go viral. Researchers from the University of California, Berkeley, have developed a new AI framework called CausalMamba that addresses this gap by combining rumor classification with causal analysis to pinpoint the key messages driving misinformation. This approach moves beyond simply labeling content as true or false to reveal the underlying mechanisms of spread, offering a more actionable path for moderators and platforms to intervene effectively.
The core finding of the study is that CausalMamba can identify influential nodes within rumor propagation chains, such as specific tweets that act as central bridges in discussion threads. In experiments on the Twitter15 dataset, which includes 1,490 events with labels like True, False, Unverified, and Non-rumor, the model achieved competitive performance with an accuracy of 0.597 and a macro-F1 score of 0.598. More importantly, it learned latent causal graphs that show directional influence between tweets, allowing researchers to simulate interventions by removing top-ranked nodes. For example, in Event 17, removing the top three nodes identified via PageRank significantly disrupted graph connectivity, demonstrating how targeted actions could alter information flow.
Ology integrates three components: a Mamba-based sequence encoder to capture long-range dependencies in tweet content, a graph convolutional network (GCN) to model reply tree structures, and a differentiable causal module based on NOTEARS to learn directed acyclic graphs. The model processes each event as a graph where nodes represent tweets with features including BERT embeddings, temporal delays, and user information, concatenated into 833-dimensional inputs. It is trained with a joint loss function that balances classification accuracy and causal sparsity, using hyperparameters like a hidden dimension of 128 and a causal loss weight of 0.1.
From the paper show that CausalMamba outperforms baselines like BiLSTM-CNN (accuracy 0.505) and Transformer (accuracy 0.522), with the Mamba-GCN variant achieving the highest accuracy at 0.643. The inclusion of the causal module slightly reduced performance compared to Mamba-GCN, reflecting a trade-off between interpretability and predictive power. Qualitative analysis, as illustrated in Figure 2, reveals that the learned causal graphs highlight nodes with high PageRank scores, which are not always the source tweet but often intermediate replies that drive propagation. This enables counterfactual intervention simulations, where removing these nodes fragments the graph, validating the model's ability to identify actionable control points.
Of this research are significant for real-world applications, as it provides a unified framework for both detecting rumors and understanding their spread dynamics. By uncovering causal structures, platforms could prioritize moderation efforts on high-influence posts, potentially reducing the viral reach of misinformation without blanket censorship. The model's interpretability also addresses ethical concerns by offering transparency in AI decisions, aligning with responsible AI principles. However, the study notes limitations, such as evaluation being restricted to the Twitter15 dataset and a slight performance drop due to causal constraints, suggesting future work on dynamic loss weighting and extension to other datasets like Twitter16 or Weibo for broader validation.
Original Source
Read the complete research paper
About the Author
Guilherme A.
Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.
Connect on LinkedIn