1. A distributed kernel summation framework for general-dimension machine learning
- Author
-
Dongryeol Lee, Richard Vuduc, Piyush Sao, and Alexander G. Gray
- Subjects
Theoretical computer science ,business.industry ,Computer science ,Kernel density estimation ,Parallel computing ,Machine learning ,computer.software_genre ,Computer Science Applications ,Kernel method ,Polynomial kernel ,Kernel embedding of distributions ,Kernel (statistics) ,Radial basis function kernel ,Distributed memory ,Artificial intelligence ,Tree kernel ,business ,computer ,Analysis ,Information Systems - Abstract
Kernel summations are a ubiquitous key computational bottleneck in many data analysis methods. In this paper, we attempt to marry, for the first time, the best relevant techniques in parallel computing, where kernel summations are in low dimensions, with the best general-dimension algorithms from the machine learning literature. We provide the first distributed implementation of kernel summation framework that can utilize: i various types of deterministic and probabilistic approximations that may be suitable for low and high-dimensional problems with a large number of data points; ii any multidimensional binary tree using both distributed memory and shared memory parallelism; and iii a dynamic load balancing scheme to adjust work imbalances during the computation. Our hybrid message passing interface MPI/OpenMP codebase has wide applicability in providing a general framework to accelerate the computation of many popular machine learning methods. Our experiments show scalability results for kernel density estimation on a synthetic ten-dimensional dataset containing over one billion points and a subset of the Sloan Digital Sky Survey Data up to 6144 cores. © 2013 Wiley Periodicals, Inc. Statistical Analysis and Data Mining, 2013
- Published
- 2013
- Full Text
- View/download PDF