Skip to ContentSkip to Navigation
Digital Competence Centre
your one-stop for research IT and data
Digital Competence Centre Contact

AI frameworks for fake news detection

An interview with Fadi Mohsen
10 March 2025

The rise of social media has transformed the way information spreads, making fake news detection a critical challenge. To address this issue, Dr. Fadi Mohsen, a UG researcher specializing in cybersecurity and machine learning, has developed a web framework that enables researchers to compare various machine learning algorithms for misinformation detection.

Dr. Mohsen’s initial research focus was on machine learning, but his fascination with the ever-evolving nature of security threats led him to cybersecurity. During his PhD, he identified vulnerabilities in Android systems and authentication protocols, working to develop countermeasures and improve security practices. Over the past nine years, his expertise has expanded. While cybersecurity remains his primary focus, he has consistently integrated machine learning, text mining, and large language models (LLMs) into his research.

In this interview, we explore his latest dataset and web framework, ‘Machine Learning Frameworks for Fake News Detection and Datasets’, designed to help researchers compare machine learning models for fake news detection.

By adopting an open science approach, researchers can collaborate more effectively, validate findings, and accelerate advancements in misinformation detection.

Could you provide an overview of your dataset and its significance in the context of fake news detection?

In this research, we utilized multiple datasets that have been previously referenced in the literature. To better align with real-world conditions, particularly for incremental learning scenarios, we derived new datasets from existing sources, including trimmed versions of WELFake, Scraped, and Kaggle 1.

One of the challenges we faced was writing a clear and detailed README file.

To promote reproducibility and further research, we have made the original datasets, newly generated datasets, and the scripts used for dataset generation publicly available. This ensures that other researchers can replicate our experiments, validate our findings, or extend our approach by applying it to new contexts. Our work provides a structured framework for evaluating fake news detection models across evolving datasets, which is crucial for tackling misinformation in dynamic environments.

You published your code along with your data. Did you face any challenges in preparing the code for publication?

Open research data and code are crucial in every field.

Yes, we built a framework that allows researchers to conduct comparative studies on mainstream machine learning algorithms and their ability to detect fake news articles. One of the challenges we faced was writing a clear and detailed README file to ensure that others could easily understand and use our framework. While our current method of sharing the code is sufficient, we realized that it lacks proper version control for tracking changes made by others.

You have chosen to publish your code and dataset in the repository DataverseNL. To what extent can open access to research data help to counter misinformation and fake news?

Open research data and code are crucial in every field, including fake news detection. The spread of misinformation can cause panic, uncertainty, and even real-world harm. By adopting an open science approach, researchers can collaborate more effectively, validate findings, and accelerate advancements in misinformation detection. Publishing datasets and code in open repositories like DataverseNL ensures Transparency & Reproducibility, Efficiency & Progress, and Stronger Countermeasures.

The utilization of AI and LLMs by cybercriminals to generate more believable and hard-to-detect fake news articles, images, and videos is alarming.

One of the big worries with AI is that we are unconsciously introducing biases in the algorithms. Did you identify any biases in your code? 

We utilize AI and machine learning in crafting cybersecurity solutions. Although we tend to tackle the possible biases introduced by the selected algorithms. Our approach does not dive deep into the core implementations of these algorithms; instead we focus on the experimentation settings to limit any possible biases. For example, in our fake news detection work, we purposely selected various and diverse datasets to mitigate potential biases and enhance the generalizability of our findings.

Given your expertise in the field of fake news detection, what is the most striking thing you have found in your work?

There are two things that strike me the most: first, the volume and the widespread use of fake content on the web. Second, the utilization of  AI and large language models (LLMs) by cyber criminals to generate more believable and hard to detect fake news articles, images, and videos. 

Users have the tendency to share information without verifying it — educating the public on the need to verify before sharing any information is crucial.

On the one hand, the knowledge and the technical capability to deal with fake news detection is evolving, while on the other hand, societies remain greatly vulnerable to the dangers of fake news and misinformation.  What macro-level strategies should be adopted to better protect societies from this global challenge?

In the world of security, humans are regarded as the weakest link. The root cause of vulnerabilities in systems are human malpractices e.g. a developer forgets to sanitize data before sending them to the backend or a system admin uses easy passwords. Furthermore, the exploitation of these vulnerabilities also occurs due to the human’s error, e.g., an employee clicking on an email attachment from a strange address or an employee forgetting to change her password. The same can be applied to the spread of fake news. Users have the tendency to share information without verifying it. Therefore, advancing the automated countermeasures is not enough to counter this volume of misinformation. Educating the public on the need to verify before sharing any information is crucial.

Last modified:11 March 2025 12.41 p.m.
Share this Facebook LinkedIn

More news

  • 11 March 2025

    Water: what if there is sometimes too much of it?

    A changing climate affects all sorts of things, from energy and food supplies to natural disasters such as floods. Researchers at the University of Groningen work on models to get a better grip on such changes, and to be able to make predictions.

  • 05 March 2025

    Women in Science

    The UG celebrates International Women’s Day with a special photo series: Women in Science.

  • 04 March 2025

    Why do we let our rainwater drain away through the sewage system?

    Regional water authorities are sometimes forced to discharge waste water from our sewage system untreated. ‘It then ends up in a canal or a lake,’ computer scientist Dilek Düştegör explains. She collaborates with municipalities and regional water...