Multidimensional Aggregations in Parallel Database Systems

Date

2017-12

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

Aggregations help computing summaries of a data set, which are ubiquitous in various big data analytics problems. In this dissertation, we provide two major technical contributions which work on parallel database systems that significantly extend the capabilities of aggregations, studying two complementary multidimensional mathematical structures: cubes and matrices. Cubes present a combinatorial problem in a set of discrete dimensions, widely studied in database systems. On the other hand, matrices are widely used in machine learning models, requiring iterative numerical methods taking input as multidimensional vectors. Both problems are difficult to solve on large data sets residing on secondary storage, and their algorithms are difficult to optimize in a parallel cluster. First, we extend cubes to intuitively show the relationship between measures aggregated at different grouping levels by introducing the percentage cube, a generalized database cube that takes percentages as its basic measure instead of simple sums. We show that the percentage cube is significantly harder to compute than the standard cube due to a higher exponential complexity. We propose SQL syntax and introduce novel query optimizations to materialize the percentage cube without any memory limitations. We compare our optimized queries with existing SQL functions, evaluating time, speed-up, and effectiveness of lattice pruning methods. Besides, we show columnar storage provides significant acceleration over row storage, the standard storage mechanism. Second, we study parallel aggregation on large matrices stored as tables and study how to compute a comprehensive data summarization matrix, called Gamma. Gamma generally fits in main memory, and it is shown to enable the efficient derivation of many machine learning models. Specifically, we show our Gamma summarization matrix benefits linear machine learning models, including PCA, linear regression, classification, and variable selection. We analytically show our summarization matrix captures essential statistical properties of the data set and we experimentally show Gamma allows iterative algorithms to iterate faster in main memory. Also, we show Gamma is further accelerated with array and columnar storage. We experimentally prove our parallel aggregations allow faster computation than existing machine learning libraries for model computations in R and Spark, two popular analytic platforms.

Description

Keywords

Databases, Multidimensional, Aggregation, Parallel, Cube, OLAP, Machine learning, Data summarization, PCA, Regression, Percentage

Citation

Portions of this document appear in: Zhang, Yiqun, Carlos Ordonez, Javier García-García, and Ladjel Bellatreche. "Optimization of Percentage Cube Queries." In EDBT/ICDT Workshops. 2017. And in: Zhang, Yiqun, Carlos Ordonez, and Wellington Cabrera. "Big data analytics integrating a parallel columnar DBMS and the R language." In 2016 16th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid), pp. 627-630. IEEE, 2016. And in: Ordonez, Carlos, Yiqun Zhang, and Wellington Cabrera. "The Gamma matrix to summarize dense and sparse data sets for big data analytics." IEEE Transactions on Knowledge and Data Engineering 28, no. 7 (2016): 1905-1918.