Back to Search Start Over

Power/performance trade-offs of small batched LU based solvers on GPUs

Authors :
Antonino Tumeo
Massimiliano Fatica
Oreste Villa
Nitin A. Gawande
Source :
Euro-Par 2013 Parallel Processing ISBN: 9783642400469, Euro-Par
Publication Year :
2013

Abstract

In this paper we propose and analyze a set of batched linear solvers for small matrices on Graphic Processing Units (GPUs), evaluating the various alternatives depending on the size of the systems to solve. We discuss three different solutions that operate with different levels of parallelization and GPU features. The first, exploiting the CUBLAS library, manages matrices of size up to 32x32 and employs Warp level (one matrix, one Warp) parallelism and shared memory. The second works at Thread-block level parallelism (one matrix, one Thread-block), still exploiting shared memory but managing matrices up to 76x76. The third is Thread level parallel (one matrix, one thread) and can reach sizes up to 128x128, but it does not exploit shared memory and only relies on the high memory bandwidth of the GPU. The first and second solutions only support partial pivoting, the third one easily supports partial and full pivoting, making it attractive to problems that require greater numerical stability. We analyze the trade-offs in terms of performance and power consumption as function of the size of the linear systems that are simultaneously solved. We execute the three implementations on a Tesla M2090 (Fermi) and on a Tesla K20 (Kepler).

Details

ISBN :
978-3-642-40046-9
ISBNs :
9783642400469
Database :
OpenAIRE
Journal :
Euro-Par 2013 Parallel Processing ISBN: 9783642400469, Euro-Par
Accession number :
edsair.doi.dedup.....0d65eea9fdd4fde4b1a979ba61395985
Full Text :
https://doi.org/10.1007/978-3-642-40047-6_81