18 results on '"Extreme-Scale Computing"'
Search Results
2. The Landscape of Exascale Research: A Data-Driven Literature Analysis.
- Author
-
HELDENS, STIJN, HIJMA, PIETER, VAN WERKHOVEN, BEN, MAASSEN, JASON, BELLOUM, ADAM S. Z., and VAN NIEUWPOORT, ROB V.
- Abstract
The next generation of supercomputers will break the exascale barrier. Soon we will have systems capable of at least one quintillion (billion billion) floating-point operations per second (1018 FLOPS). Tremendous amounts of work have been invested into identifying and overcoming the challenges of the exascale era. In this work, we present an overview of these efforts and provide insight into the important trends, developments, and exciting research opportunities in exascale computing. We use a three-stage approach in whichwe (1) discuss various exascale landmark studies, (2) use data-driven techniques to analyze the large collection of related literature, and (3) discuss eight research areas in depth based on influential articles. Overall, we observe that great advancements have been made in tackling the two primary exascale challenges: energy efficiency and fault tolerance. However, as we look forward, we still foresee two major concerns: the lack of suitable programming tools and the growing gap between processor performance and data bandwidth (i.e., memory, storage, networks). Although we will certainly reach exascale soon, without additional research, these issues could potentially limit the applicability of exascale computing. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
3. PETSc DMNetwork: A Library for Scalable Network PDE-Based Multiphysics Simulations.
- Author
-
Abhyankar, Shrirang, Betrie, Getnet, Maldonado, Daniel Adrian, Mcinnes, Lois C., Smith, Barry, and Zhang, Hong
- Subjects
- *
LIBRARY information networks , *ELECTRIC circuits , *WATER distribution , *ELECTRIC power distribution grids , *USER interfaces - Abstract
We present DMNetwork, a high-level package included in the PETSc library for the simulation of multiphysics phenomena over large-scale networked systems. The library aims at applications that have networked structures such as those in electrical, gas, and water distribution systems. DMNetwork provides data and topology management, parallelization for multiphysics systems over a network, and hierarchical and composable solvers to exploit the problem structure. DMNetwork eases the simulation development cycle by providing the necessary infrastructure through simple abstractions to define and query the network components. This article presents the design of DMNetwork, illustrates its user interface, and demonstrates its ability to solve multiphysics systems, such as an electric circuit, a network of power grid and water subnetworks, and transient hydraulic systems over large networks with more than 2 billion variables on extreme-scale computers using up to 30,000 processors. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
4. Expansion Quality of Epidemic Protocols
- Author
-
Poonpakdee, Pasu, Di Fatta, Giuseppe, Kacprzyk, Janusz, Series editor, Camacho, David, editor, Braubach, Lars, editor, Venticinque, Salvatore, editor, and Badica, Costin, editor
- Published
- 2015
- Full Text
- View/download PDF
5. Connectivity Recovery in Epidemic Membership Protocols
- Author
-
Poonpakdee, Pasu, Di Fatta, Giuseppe, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Di Fatta, Giuseppe, editor, Fortino, Giancarlo, editor, Li, Wenfeng, editor, Pathan, Mukaddim, editor, Stahl, Frederic, editor, and Guerrieri, Antonio, editor
- Published
- 2015
- Full Text
- View/download PDF
6. Convergence Detection in Epidemic Aggregation
- Author
-
Poonpakdee, Pasu, Orhon, Neriman Gamze, Di Fatta, Giuseppe, Hutchison, David, editor, Kanade, Takeo, editor, Kittler, Josef, editor, Kleinberg, Jon M., editor, Kobsa, Alfred, editor, Mattern, Friedemann, editor, Mitchell, John C., editor, Naor, Moni, editor, Nierstrasz, Oscar, editor, Pandu Rangan, C., editor, Steffen, Bernhard, editor, Terzopoulos, Demetri, editor, Tygar, Doug, editor, Weikum, Gerhard, editor, an Mey, Dieter, editor, Alexander, Michael, editor, Bientinesi, Paolo, editor, Cannataro, Mario, editor, Clauss, Carsten, editor, Costan, Alexandru, editor, Kecskemeti, Gabor, editor, Morin, Christine, editor, Ricci, Laura, editor, Sahuquillo, Julio, editor, Schulz, Martin, editor, Scarano, Vittorio, editor, Scott, Stephen L., editor, and Weidendorfer, Josef, editor
- Published
- 2014
- Full Text
- View/download PDF
7. Big data and extreme-scale computing.
- Author
-
Asch, M., Moore, T., Badia, R., Beck, M., Beckman, P., Bidot, T., Bodin, F., Cappello, F., Choudhary, A., de Supinski, B., Deelman, E., Dongarra, J., Dubey, A., Fox, G., Fu, H., Girona, S., Gropp, W., Heroux, M., Ishikawa, Y., and Keahey, K.
- Subjects
- *
BIG data , *COMPUTING platforms , *DATA analysis , *HIGH performance computing , *SCIENTIFIC computing - Abstract
Over the past four years, the Big Data and Exascale Computing (BDEC) project organized a series of five international workshops that aimed to explore the ways in which the new forms of data-centric discovery introduced by the ongoing revolution in high-end data analysis (HDA) might be integrated with the established, simulation-centric paradigm of the high-performance computing (HPC) community. Based on those meetings, we argue that the rapid proliferation of digital data generators, the unprecedented growth in the volume and diversity of the data they generate, and the intense evolution of the methods for analyzing and using that data are radically reshaping the landscape of scientific computing. The most critical problems involve the logistics of wide-area, multistage workflows that will move back and forth across the computing continuum, between the multitude of distributed sensors, instruments and other devices at the networks edge, and the centralized resources of commercial clouds and HPC centers. We suggest that the prospects for the future integration of technological infrastructures and research ecosystems need to be considered at three different levels. First, we discuss the convergence of research applications and workflows that establish a research paradigm that combines both HPC and HDA, where ongoing progress is already motivating efforts at the other two levels. Second, we offer an account of some of the problems involved with creating a converged infrastructure for peripheral environments, that is, a shared infrastructure that can be deployed throughout the network in a scalable manner to meet the highly diverse requirements for processing, communication, and buffering/storage of massive data workflows of many different scientific domains. Third, we focus on some opportunities for software ecosystem convergence in big, logically centralized facilities that execute large-scale simulations and models and/or perform large-scale data analytics. We close by offering some conclusions and recommendations for future investment and policy review. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
8. The future of scientific workflows.
- Author
-
Deelman, Ewa, Peterka, Tom, Altintas, Ilkay, Carothers, Christopher D., van Dam, Kerstin Kleese, Moreland, Kenneth, Parashar, Manish, Ramakrishnan, Lavanya, Taufer, Michela, and Vetter, Jeffrey
- Subjects
- *
AUTOMATION , *WORKFLOW management , *COMPUTER scientists , *COMPUTER systems , *DISTRIBUTED computing - Abstract
Today’s computational, experimental, and observational sciences rely on computations that involve many related tasks. The success of a scientific mission often hinges on the computer automation of these workflows. In April 2015, the US Department of Energy (DOE) invited a diverse group of domain and computer scientists from national laboratories supported by the Office of Science, the National Nuclear Security Administration, from industry, and from academia to review the workflow requirements of DOE’s science and national security missions, to assess the current state of the art in science workflows, to understand the impact of emerging extreme-scale computing systems on those workflows, and to develop requirements for automated workflow management in future and existing environments. This article is a summary of the opinions of over 50 leading researchers attending this workshop. We highlight use cases, computing systems, workflow needs and conclude by summarizing the remaining challenges this community sees that inhibit large-scale scientific workflows from becoming a mainstream tool for extreme-scale science. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
9. Addressing failures in exascale computing.
- Author
-
Snir, Marc, Wisniewski, Robert W, Abraham, Jacob A, Adve, Sarita V, Bagchi, Saurabh, Balaji, Pavan, Belak, Jim, Bose, Pradip, Cappello, Franck, Carlson, Bill, Chien, Andrew A, Coteus, Paul, DeBardeleben, Nathan A, Diniz, Pedro C, Engelmann, Christian, Erez, Mattan, Fazzari, Saverio, Geist, Al, Gupta, Rinku, and Johnson, Fred
- Subjects
- *
COMPUTER system failures , *COMPUTERS conferences , *COMPUTER input-output equipment , *COMPUTER software , *APPLICATION software - Abstract
We present here a report produced by a workshop on ‘Addressing failures in exascale computing’ held in Park City, Utah, 4–11 August 2012. The charter of this workshop was to establish a common taxonomy about resilience across all the levels in a computing system, discuss existing knowledge on resilience across the various hardware and software layers of an exascale system, and build on those results, examining potential solutions from both a hardware and software perspective and focusing on a combined approach.The workshop brought together participants with expertise in applications, system software, and hardware; they came from industry, government, and academia, and their interests ranged from theory to implementation. The combination allowed broad and comprehensive discussions and led to this document, which summarizes and builds on those discussions. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
10. The Landscape of Exascale Research: A Data-Driven Literature Analysis
- Author
-
Stijn Heldens, Adam Belloum, Pieter Hijma, Ben van Werkhoven, Rob V. van Nieuwpoort, Jason Maassen, Multiscale Networked Systems (IvI, FNWI), System and Network Engineering (IVI, FNWI), Computer Systems, Network Institute, High Performance Distributed Computing, and Mathematics
- Subjects
data-driven analysis ,020203 distributed computing ,General Computer Science ,Research areas ,Computer science ,literature review ,high-performance computing ,Fault tolerance ,02 engineering and technology ,FLOPS ,Supercomputer ,Data science ,Exascale computing ,Additional research ,Theoretical Computer Science ,Data-driven ,020204 information systems ,extreme-scale computing ,0202 electrical engineering, electronic engineering, information engineering ,SDG 7 - Affordable and Clean Energy ,Efficient energy use - Abstract
The next generation of supercomputers will break the exascale barrier. Soon we will have systems capable of at least one quintillion (billion billion) floating-point operations per second (10 18 FLOPS). Tremendous amounts of work have been invested into identifying and overcoming the challenges of the exascale era. In this work, we present an overview of these efforts and provide insight into the important trends, developments, and exciting research opportunities in exascale computing. We use a three-stage approach in which we (1) discuss various exascale landmark studies, (2) use data-driven techniques to analyze the large collection of related literature, and (3) discuss eight research areas in depth based on influential articles. Overall, we observe that great advancements have been made in tackling the two primary exascale challenges: energy efficiency and fault tolerance. However, as we look forward, we still foresee two major concerns: the lack of suitable programming tools and the growing gap between processor performance and data bandwidth (i.e., memory, storage, networks). Although we will certainly reach exascale soon, without additional research, these issues could potentially limit the applicability of exascale computing.
- Published
- 2020
11. DCA++: A software framework to solve correlated electron problems with modern quantum cluster methods
- Author
-
Gonzalo Alvarez, Thomas C. Schulthess, Peter Staar, Urs R. Hähner, Michael S. Summers, Raffaele Solcà, and Thomas Maier
- Subjects
Workstation ,General Physics and Astronomy ,FOS: Physical sciences ,Parallel computing ,Dynamical cluster approximation ,computer.software_genre ,01 natural sciences ,010305 fluids & plasmas ,law.invention ,CUDA ,Condensed Matter - Strongly Correlated Electrons ,Software ,law ,Quantum cluster algorithms ,0103 physical sciences ,010306 general physics ,Sustainable software development ,Quantum ,Strongly Correlated Electrons (cond-mat.str-el) ,business.industry ,Strongly correlated electron systems ,Computational Physics (physics.comp-ph) ,Software framework ,Titan (supercomputer) ,Hardware and Architecture ,Scalability ,Strongly correlated material ,Continuous-time quantum Monte Carlo ,Extreme-scale computing ,business ,computer ,Physics - Computational Physics - Abstract
We present the first open release of the DCA++ project, a high-performance research software framework to solve quantum many-body problems with cutting edge quantum cluster algorithms. DCA++ implements the dynamical cluster approximation (DCA) and its DCA+ extension with a continuous self-energy. The algorithms capture nonlocal correlations in strongly correlated electron systems, thereby giving insight into high-Tc superconductivity. The code's scalability allows efficient usage of systems at all scales, from workstations to leadership computers. With regard to the increasing heterogeneity of modern computing machines, DCA++ provides portable performance on conventional and emerging new architectures, such as hybrid CPU–GPU, sustaining multiple petaflops on ORNL's Titan and CSCS’ Piz Daint supercomputers. Moreover, we show how sustainable and scalable development of the code base has been achieved by adopting standard techniques of the software industry. These include employing a distributed version control system, applying test-driven development and following continuous integration., Computer Physics Communications, 246, ISSN:0010-4655, ISSN:1879-2944
- Published
- 2020
12. Dynamic group communication for large-scale parallel data mining.
- Author
-
Katti, Amogh and Di Fatta, Giuseppe
- Subjects
PARALLEL computers ,DATA mining ,MULTICORE processors ,INTERNATIONAL communication ,DATA distribution ,COMPUTER algorithms ,ERROR analysis in mathematics - Abstract
Exascale systems are the next frontier in high-performance computing and are expected to deliver a performance of the order of 1018 operations per second using massive multicore processors. Very large- and extreme-scale parallel systems pose critical algorithmic challenges, especially related to concurrency, locality and the need to avoid global communication patterns. This work investigates a novel protocol for dynamic group communication that can be used to remove the global communication requirement and to reduce the communication cost in parallel formulations of iterative data mining algorithms. The protocol is used to provide a communication-efficient parallel formulation of the k-means algorithm for cluster analysis. The approach is based on a collective communication operation for dynamic groups of processes and exploits non-uniform data distributions. Non-uniform data distributions can be either found in real-world distributed applications or induced by means of multidimensional binary search trees. The analysis of the proposed dynamic group communication protocol has shown that it does not introduce significant communication overhead. The parallel clustering algorithm has also been extended to accommodate an approximation error, which allows a further reduction of the communication costs. The effectiveness of the exact and approximate methods has been tested in a parallel computing system with 64 processors and in simulations with 1024 processing elements. [ABSTRACT FROM PUBLISHER]
- Published
- 2013
- Full Text
- View/download PDF
13. Non-uniform data distribution for communication-efficient parallel clustering.
- Author
-
Goodall, Tabitha, Pettinger, David, and Di Fatta, Giuseppe
- Subjects
DATA distribution ,INTERNATIONAL communication ,PARALLEL computers ,CLUSTER analysis (Statistics) ,COMPUTER algorithms ,DATA mining - Abstract
Abstract: Global communication requirements and load imbalance of some parallel data mining algorithms are the major obstacles to exploit the computational power of large-scale systems. This work investigates how non-uniform data distributions can be exploited to remove the global communication requirement and to reduce the communication cost in parallel data mining algorithms and, in particular, in the k-means algorithm for cluster analysis. In the straightforward parallel formulation of the k-means algorithm, data and computation loads are uniformly distributed over the processing nodes. This approach has excellent load balancing characteristics that may suggest it could scale up to large and extreme-scale parallel computing systems. However, at each iteration step the algorithm requires a global reduction operation which hinders the scalability of the approach. This work studies a different parallel formulation of the algorithm where the requirement of global communication is removed, while maintaining the same deterministic nature of the centralised algorithm. The proposed approach exploits a non-uniform data distribution which can be either found in real-world distributed applications or can be induced by means of multi-dimensional binary search trees. The approach can also be extended to accommodate an approximation error which allows a further reduction of the communication costs. The effectiveness of the exact and approximate methods has been tested in a parallel computing system with 64 processors and in simulations with 1024 processing elements. [Copyright &y& Elsevier]
- Published
- 2013
- Full Text
- View/download PDF
14. The Block Jacobi-Davidson Eigensolver in PHIST
- Author
-
Thies, Jonas, Röhrig-Zöllner, Melven, Ernst, Dominik, Kreutzer, Moritz, Basermann, Achim, Hager, Georg, and Wellein, Gerhard
- Subjects
Jacobi-Davidson ,sparse matrices ,extreme-scale computing ,High Performance Computing ,eigenvalues - Abstract
not available
- Published
- 2018
15. Holistic Performance Engineering for Sparse Iterative Solvers
- Author
-
Thies, Jonas, Ernst, Dominik, and Röhrig-Zöllner, Melven
- Subjects
High-Performance Computing ,Parallel Iterative Methods ,Extreme-Scale Computing ,High Performance Computing ,Sparse Linear Algebra ,Eigenvalue Solvers - Abstract
In many applications, sparse (linear and/or eigenvalue) solvers take up a large fraction of the overall runtime. We believe that the increasingly complex hardware of today's and future HPC systems has lead to a gap in the understanding of the performance achieved by actual applications, many of which are still using a monolithic `MPI only' approach despite the heterogeneous nature of the hardware. We have developed a new sparse solver library PHIST (https://bitbucket.org/essex/phist/) that defines a simple "kernel interface" layer inspired by MPI. Algorithms implemented in PHIST are portable in terms of software and performance as they only call building blocks of linear algebra via this interface. We have introduced simple performance models for these basic building blocks at the interface level, so that regardless of the backend providing the implementation, an overview of the optimization potential on the kernel level can be obtained, and performance pitfalls in the application (e.g. strided memory accesses) may be revealed. Available backends for PHIST include established libraries such as Trilinos/Epetra or PETSc, as well as more recent "MPI+X" approaches as implemented in Trilinos/Tpetra or our own kernel library GHOST (https://bitbucket.org/essex/ghost).
- Published
- 2018
16. Block Krylov and Jacobi-Davidson methods on heterogenous systems
- Author
-
Thies, Jonas, Ernst, Dominik, and Röhrig-Zöllner, Melven
- Subjects
High-Performance Computing ,eigenvalue problems ,Hybrid Parallel ,extreme-scale computing ,High Performance Computing ,sparse linear algebra - Abstract
Over the past five years we have developed two open source software packages called GHOST and PHIST (https://bitbucket.org/essex/[ghost|phist]. We discuss the software and performance engineering techniques used when designing these libraries and show some examples of use. GHOST provides optimized implementations of memory-bounded linear algebra operations on heterogenous CPU/GPU systems. PHIST provides the software infrastructure for implementing iterative sparse matrix algorithms in a portable and efficient way by introducing a kernel interface layer inspired by the message passing interface (MPI). Implementations of the interface are verified using an extensive test suite and performance models. Going beyond the isolated optimzation of linear algebra kernels, phist allows algorithm-level performance optimizations like kernel fusion and overlapping of communication and computation. To make phist algorithms easy to integrate into existing applications, we provide implementations of the kernel interface for various commonly used libraries such as Trilinos, PETSc and Eigen, and a Fortran+MPI reference implementation. Besides the standard C interface, Pyton, C++ and Fortran bindings are automatically generated for all functions. We show how the new libraries can be used to boost the performance of existing implementations of Block Krylov solvers in the Trilinos package Anasazi, and present results for our own implementation of the block Jacobi-Davidson QR method applied to model problems from quantum physics.
- Published
- 2018
17. PHIST: a Pipelined, Hybrid-parallel Iterative Solver Toolkit
- Author
-
Röhrig-Zöllner, Melven and Thies, Jonas
- Subjects
High-Performance Computing ,Parallel Iterative Methods ,High Performance Computing ,Extreme-Scale Computing ,Sparse Linear Algebra ,Eigenvalue Solvers - Published
- 2018
18. DCA++: A software framework to solve correlated electron problems with modern quantum cluster methods.
- Author
-
Hähner, Urs R., Alvarez, Gonzalo, Maier, Thomas A., Solcà, Raffaele, Staar, Peter, Summers, Michael S., and Schulthess, Thomas C.
- Subjects
- *
SOFTWARE frameworks , *MICROCOMPUTER workstations (Computers) , *COMPUTER systems , *MANY-body problem , *ELECTRONS , *SUPERCOMPUTERS , *PORTABLE computers - Abstract
We present the first open release of the DCA++ project, a high-performance research software framework to solve quantum many-body problems with cutting edge quantum cluster algorithms. DCA++ implements the dynamical cluster approximation (DCA) and its DCA + extension with a continuous self-energy. The algorithms capture nonlocal correlations in strongly correlated electron systems, thereby giving insight into high- T c superconductivity. The code's scalability allows efficient usage of systems at all scales, from workstations to leadership computers. With regard to the increasing heterogeneity of modern computing machines, DCA++ provides portable performance on conventional and emerging new architectures, such as hybrid CPU–GPU, sustaining multiple petaflops on ORNL's Titan and CSCS' Piz Daint supercomputers. Moreover, we show how sustainable and scalable development of the code base has been achieved by adopting standard techniques of the software industry. These include employing a distributed version control system, applying test-driven development and following continuous integration. Program Title: DCA++ Program Files doi: http://dx.doi.org/10.17632/482jm5cv77.1 Licensing provisions: BSD-3-Clause Programming language: C++14 and CUDA Nature of problem: Understanding the fascinating physics of strongly correlated electron systems requires the development of sophisticated algorithms and their implementation on leadership computing systems. Solution method: The DCA++ code provides a highly scalable and efficient implementation of the dynamical cluster approximation (DCA) and its DCA + extension. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.