Abstract

HPC systems expose configuration options to assist optimization. Configurations such as parallelism, thread and data mapping, or prefetching have been explored but with a limited optimization objective (e.g., performance) and a fixed problem size. Unfortunately, efficient strategies in one scenario may poorly generalize when applied in new contexts.We investigate the impact of configuration options and different problem sizes over performance and energy. Well-adapted NUMA-related options and cache-prefetchers provide significantly more gains for energy (5.9×) than performance (1.85×). Moreover, reusing optimization strategies from performance to energy only provides 40% of the gains found when natively optimizing for energy, while transferring strategies across problem sizes limits to 70% of the original gains.We fill this gap with ML: simple decision trees predict the best configuration for a target size using only information collected on another size. Our models achieve 88% of the native gains when cross-predicting across performance and energy, and 85% across problem sizes.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.