Future exascale systems must be optimized for both power and performance at scale in order to achieve DOE’s goal of a sustained petaflop within 20 Megawatts by 2022 [1]. Massive parallelism of the future systems combined with complex memory hierarchies will form a barrier to efficient application and architecture design. These challenges are exacerbated with emerging complex architectures such as GPGPUs and Intel Xeon Phi as parallelism increases orders of magnitude and system power consumption can easily triple or quadruple. Therefore, we need techniques that can reduce the search space for optimization, isolate power-performance bottlenecks, identify root causes for software/hardware inefficiency, and effectively direct runtime scheduling.
Revised: April 27, 2015 |
Published: September 23, 2013
Citation
Song S., N.R. Tallent, and A. Vishnu. 2013.Exploring Machine Learning Techniques For Dynamic Modeling on Future Exascale Systems. In Modeling & Simulation of Exascale Systems & Applications: Workshop on Modeling & Simulation of Exascale Systems & Applications, September 18-19, 2013, Seattle, Washington. Washington Dc:US Department of Energy, Office of Advanced Scientific Computing Research.PNNL-SA-105672.