TY - UNPD A1 - Bluhm, Benjamin A1 - Cutura, Jannic T1 - Econometrics at scale: Spark up big data in economics T2 - SAFE working paper series ; No. 266 N2 - This paper provides an overview of how to use "big data" for economic research. We investigate the performance and ease of use of different Spark applications running on a distributed file system to enable the handling and analysis of data sets which were previously not usable due to their size. More specifically, we explain how to use Spark to (i) explore big data sets which exceed retail grade computers memory size and (ii) run typical econometric tasks including microeconometric, panel data and time series regression models which are prohibitively expensive to evaluate on stand-alone machines. By bridging the gap between the abstract concept of Spark and ready-to-use examples which can easily be altered to suite the researchers need, we provide economists and social scientists more generally with the theory and practice to handle the ever growing datasets available. The ease of reproducing the examples in this paper makes this guide a useful reference for researchers with a limited background in data handling and distributed computing. T3 - SAFE working paper - 266 KW - Econometrics KW - Distributed Computing KW - Apache Spark Y1 - 2020 UR - http://publikationen.ub.uni-frankfurt.de/frontdoor/index/index/docId/51804 UR - https://nbn-resolving.org/urn:nbn:de:hebis:30:3-518045 N1 - An earlier version of this paper circulated as "Time Series Econometrics at Scale: A Practical Guide to Parallel Computing in (Py)Spark". IS - February 6, 2020 PB - SAFE CY - Frankfurt am Main ER -