The Parable of Google Flu: Traps in Big Data Analysis
Author(s) -
David Lazer,
Ryan Kennedy,
Gary King,
Alessandro Vespignani
Publication year - 2014
Publication title -
science
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 12.556
H-Index - 1186
eISSN - 1095-9203
pISSN - 0036-8075
DOI - 10.1126/science.1248506
Subject(s) - big data , covid-19 , fake news , internet privacy , world wide web , computer science , medicine , data mining , infectious disease (medical specialty) , disease , pathology
Large errors in flu prediction were largely avoidable, which offers lessons for the use of big data. In February 2013, Google Flu Trends (GFT) made headlines but not for a reason that Google executives or the creators of the flu tracking system would have hoped. Nature reported that GFT was predicting more than double the proportion of doctor visits for influenza-like illness (ILI) than the Centers for Disease Control and Prevention (CDC), which bases its estimates on surveillance reports from laboratories across the United States (1, 2). This happened despite the fact that GFT was built to predict CDC reports. Given that GFT is often held up as an exemplary use of big data (3, 4), what lessons can we draw from this error?
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom