Alpaka - An Abstraction Library for Parallel Kernel Acceleration
Publikation: Beitrag in Buch/Konferenzbericht/Sammelband/Gutachten › Beitrag in Konferenzband › Beigetragen › Begutachtung
Beitragende
Abstract
Porting applications to new hardware or programming models is a tedious and error prone process. Every help that eases these burdens is saving developer time that can then be invested into the advancement of the application itself instead of preserving the status-quo on a new platform. The Alpaka library defines and implements an abstract hierarchical redundant parallelism model. The model exploits parallelism and memory hierarchies on a node at all levels available in current hardware. By doing so, it allows to achieve platform and performance portability across various types of accelerators by ignoring specific unsupported levels and utilizing only the ones supported on a specific accelerator. All hardware types (multi-and many-core CPUs, GPUs and other accelerators) are supported for and can be programmed in the same way. The Alpaka C++ template interface allows for straightforward extension of the library to support other accelerators and specialization of its internals for optimization.
Running Alpaka applications on a new (and supported) platform requires the change of only one source code line instead of a lot of #ifdefs.
Details
Originalsprache | Englisch |
---|---|
Titel | 2016 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) |
Seiten | 631-640 |
Seitenumfang | 10 |
ISBN (elektronisch) | 978-1-5090-3682-0 |
Publikationsstatus | Veröffentlicht - 2016 |
Peer-Review-Status | Ja |
Publikationsreihe
Reihe | 2016 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) |
---|---|
ISSN | 2164-7062 |
Konferenz
Titel | 30th IEEE International Parallel and Distributed Processing Symposium (IPDPS) |
---|---|
Dauer | 23 - 27 Mai 2016 |
Stadt | Chicago |
Land | Israel |
Externe IDs
Scopus | 84991727562 |
---|
Schlagworte
Schlagwörter
- Heterogeneous computing, HPC, C plus, CUDA, OpenMP, platform portability, performance portability