Exascale Computing
Exascale computing focuses on developing and utilizing supercomputers capable of performing at least a quintillion (10<sup>18</sup>) calculations per second, enabling unprecedented computational power for scientific simulations and AI applications. Current research emphasizes optimizing performance and energy efficiency through techniques like advanced cooling systems, efficient distributed training algorithms (including graph neural networks and various parallel strategies for large language models), and novel data compression methods (leveraging deep learning for lossy compression). These advancements are crucial for tackling complex scientific problems across diverse fields, from materials science and climate modeling to high-energy physics and geospatial analysis, by enabling larger-scale simulations and more sophisticated AI models.
Papers
Employing Artificial Intelligence to Steer Exascale Workflows with Colmena
Logan Ward, J. Gregory Pauloski, Valerie Hayot-Sasson, Yadu Babuji, Alexander Brace, Ryan Chard, Kyle Chard, Rajeev Thakur, Ian Foster
Exploring GPU-to-GPU Communication: Insights into Supercomputer Interconnects
Daniele De Sensi, Lorenzo Pichetti, Flavio Vella, Tiziano De Matteis, Zebin Ren, Luigi Fusco, Matteo Turisini, Daniele Cesarini, Kurt Lust, Animesh Trivedi, Duncan Roweth, Filippo Spiga, Salvatore Di Girolamo, Torsten Hoefler