Publication Details

Back An overview of methods for treating selectivity in big data sources


The official statistics community is now seriously considering big data as a significant data source for producing statistics. It holds the potential for providing faster, cheaper, more detailed and completely new types of statistics. However, the use of big data also brings several challenges. One of them is the non-probabilistic character of most sources of big data, as very often, they were not designed to produce statistics. The resulting selectivity bias is therefore a major concern when using big data. This paper presents a statistical approach to big data, searching for a definition meaningful from the statistical point of view and identifying their main statistical characteristics. It then argues that big data sources share many characteristics with Internet opt-in panel surveys and proposes this as a reference to address selectivity and coverage problems in big data. Coverage and the self-selection process are briefly discussed in mobile network data, Twitter, Google Trends and Wikipedia page views data. An overview of methods which can be used to address selectivity and eliminate, or mitigate, bias is then presented, covering both methods applied at individual level, i.e. at the level of the statistical unit, and at domain level, i.e. at the level of the produced statistics. Finally, the applicability of the methods to the several big data sources is briefly discussed and a framework for adjusting selectivity in big data is proposed.

Electronic format

Download publication (EN)
Release date: 5 July 2018

Additional information

Product code: KS-TC-18-004
ISBN 978-92-79-88769-7
ISSN 2315-0807
doi:10.2785/312232
Theme: General and regional statistics
Collection: Statistical working papers