z-logo
open-access-imgOpen Access
How concept drift can impair the classification of fake news
Author(s) -
Renato Silva,
Tiago A. Almeida
Publication year - 2021
Language(s) - English
Resource type - Conference proceedings
DOI - 10.5753/kdmile.2021.17469
Subject(s) - concept drift , fake news , harm , computer science , presidential election , online and offline , presidential system , politics , machine learning , artificial intelligence , psychology , internet privacy , social psychology , political science , data stream mining , law , operating system
Fake news is a serious problem that can influence political choices, harm people's physical and mental health, promote treatments without scientific evidence, and even incite violence. Machine learning methods are one of the leading solutions that have been studied for filtering fake news automatically. However, most studies do not consider the dynamic nature of news, creating static models and evaluating them offline through the traditional holdout or cross-validation. These studies naively assume that news characteristics do not change over time and, therefore, the performance of offline models is preserved as time goes on. In this study, we show how concept drift can impair the classification of fake news. We aim to verify whether the conclusions obtained in studies that disregarded the dynamic nature of the news are sustained. We analyzed how the performance of methods trained in an offline fashion is affected by the news update over time, including concept drift due to impacting events like the Covid-19 pandemic and the United States presidential election. The results showed that the performance of offline models is over-optimistic. Incremental learning methods should be preferred because they can adapt to changes in textual patterns over time.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here