2024 ICL Retreat Held at Ijams Nature Center

ICL 2024 Retreat Photo

The 2024 ICL Retreat took place on August 14, 2024, at Ijams Nature Center. This year’s smaller group size allowed for a streamlined, one-day program that began with breakfast and concluded with a formal dinner following a full day of presentations and discussions. The natural beauty of Ijams provided an ideal backdrop for the retreat’s activities. Unfortunately, Jack was unable to attend in person due to a positive COVID-19 diagnosis; however, he presented his talk “An Overview of High Performance Computing” at the ICL Friday Lunch on September 13.

ICL Launches NSF-funded MINCER

ICL is symbolically “mincing” or breaking down the complexity of large volumes of performance monitoring data with MINCER, an NSF-funded project titled “Monitoring Infrastructure for Network and Computing Environment Research (MINCER)“. This project is led by Heike Jagode as the PI and includes a subcontract with the University of Texas at El Paso, partnering with Shirley Moore and Deepak Tosh. Spanning three years, the project started in August 2024, with a total budget of $2 million, of which ICL’s share is $1.4 million.

MINCER aims to provide comprehensive insights into distributed heterogeneous computing environments to foster research and education in platform reliability, efficiency, and security through three main thrusts: (1) Enhanced performance and power monitoring capabilities for AI architectures to support research aimed at improving AI system performance and utilization, thus enabling informed decisions on optimization and system configurations. (2) Unique network-related measurements to support research in network anomaly detection, resilience, and energy efficiency. (3) Integration of the MINCER monitoring infrastructure with research platforms like Chameleon and the Open Science Grid to enhance the understanding of cloud and edge computing dynamics, as well as the relationship between data transfer and computation in large-scale distributed computing.

HPCwire Recognizes Jack Dongarra in Inaugural HPCwire 35 Legends

HPCwire 25 Legends Jack DongarraIn celebration of its 35th anniversary, HPCwire announced the inaugural HPCwire 35 Legends — an annual list recognizing 35 luminaries who have made HPC what it is today. Included among the initial roster is ICL founder Jack Dongarra. HPCwire writes:

ACM Turing Award winner Dr. Jack Dongarra’s career in HPC began with his education in mathematics and computer science, culminating in a Ph.D. from the University of New Mexico, where h was introduced to numerical linear algebra, a foundational field in HPC.

Dongarra’s pivotal work at Argonne National Laboratory involved developing algorithms for the LINPACK library, which became a standard for measuring supercomputer performance. Dongarra told HPCwire this experience exposed him to the challenges and opportunities in scientific computing, motivating him to pursue a career dedicated to improving performance and efficiency of computational methods.

His contributions at the University of Tennessee and ORNL bridged academic research and practical HPC applications, leading to the development of the LAPACK and ScaLAPACK libraries, as well as the creation of the famed Top500 list of supercomputers.

Read the full profile and see all 35 Legends

Wall Street Journal Article Highlights China’s Growing Secrecy in Supercomputing Amid U.S. Tech Restrictions

The July 23 Wall Street Journal article, “China Is Getting Secretive About Its Supercomputers,” explores China’s increasing secrecy around its supercomputing advancements due to U.S. export restrictions on advanced chips. The article includes insights from ICL founder Jack Dongarra, who, through his work with the Top500 list, provides perspective on the global supercomputing race, emphasizing China’s potential to have faster machines despite their lack of public submissions.

Read the full article

Sparse BLAS Working Group BoF at SC24

While sparse matrix computations are at the heart of many scientific and engineering applications, there exists no widely adopted interface standard. ICL’s Hartwig Anzt is part of a cross-institutional effort involving academia and industry to define an API for sparse linear algebra operations. Hartwig presented an overview of this effort at a recent ICL Lunch Talk.

A community-driven approach is crucial for ensuring that the resulting standard caters to the diverse needs of users and gains widespread adoption. To this end, Hartwig will be leading a Sparse BLAS BoF at SC24 on Thursday, November 21 to present a blueprint of the Sparse BLAS standard, discuss the considerations behind key design choices, and solicit feedback from the community. There will be an option to join the BoF virtually. See the website for full details.

Congratulations

PEARC24 Best Student Paper for PyMAGMA: A Python Interface for MAGMA

Congratulations to current UTK EECS and former ICL student Julian Halloy, whose work on PyMAGMA: A Python Interface for MAGMA was recently recognized with a Best Student Paper award at the PEARC24 conference. Julian worked with co-author Stan Tomov on the work. From the abstract:

“The Matrix Algebra for GPU and Multicore Architectures (MAGMA) library is used for dense linear algebra computations on both CPUs and GPUs. MAGMA offers an extensive selection of linear algebra routines similar to those found in BLAS and LAPACK, but optimized and specifically designed for GPU acceleration. However, as MAGMA is a C library, C/C++ and MAGMA’s C function interfaces can be challenging for non-computer scientists to learn and use. To address this problem, PyMAGMA provides a user-friendly Python interface. PyMAGMA employs C++ object-oriented programming methodologies to simplify the interfaces, particularly by reducing the number of arguments required for routines and eliminating the need for specifying the floating-point data format and arithmetic precision used. In particular, function overloading is employed to abstract routines and avoid precision specification. We show that the PyMAGMA library provides access to high-performance linear algebra using GPUs, while offering the typical benefits of Python, including ease of learning, versatility of use, extensive access to external libraries, large community and support, cross-platform compatibility, and capability for interactive work.”

Conference Reports

CCDSC 2024

CCDSC 2024 Group Photo

The 2024 Workshop on Clusters, Clouds, and Data for Scientific Computing (CCDSC) was held at La Maison des Contes near Lyon, France, September 3rd – 6th, and brought together over 60 leading experts in HPC, clusters, and cloud computing from around the world.
This workshop is part of a biennial series that began in 1992, alternating between the U.S. and France. The purpose of this workshop, which is by invitation only, is to evaluate the state-of-the-art and future trends for cluster computing and the use of computational clouds for scientific computing. This year’s edition explored the evolving intersection of HPC, AI, and cloud technologies, with a focus on addressing scientific computing challenges. The venue, a historic château dating back to the 19th century, provided a picturesque backdrop for the event.

This slideshow requires JavaScript.

IEEE/ACM ICONS 2024

ICL’s Tokey Tahmid recently had the opportunity to attend the IEEE/ACM International Conference on Neuromorphic Systems, held July 30-August 2 at George Mason University in Arlington, Virginia. The goal of ICONS is to bring together leading researchers in neuromorphic computing to present new research, develop new collaborations, and provide a forum to publish work in this area.

Tahmid presented “Towards Scalable and Efficient Spiking Reinforcement Learning for Continuous Control Tasks”, work co-authored by his advisors: ICL’s Mark Gates and Piotr Luszczek and UTK’s Catherine Schuman. This work addresses the question of scalability and efficiency of Spiking Neural Networks in real-world continuous control problems such as mobile robots. The authors address these challenges by adopting a Deep Reinforcement Learning based Spiking Neural Network algorithm paired with MPI for distributed training, and mixed-precision for optimization. Demonstrated results showed significant improvement in performance scalability and efficiency.

Recent Releases

PAPI 7.20b

The PAPI team issued a new release of PAPI, version 7.20b.  This release introduces a new component, “rocp_sdk”, which supports AMD GPUs/APUs through the ROCprofiler-SDK interface, currently still under development and testing. The release also includes general improvements to the PAPI code, enhancing both design and functionality, as well as various bug fixes. See the PAPI Github page for a more complete list of new features and to download the software.

Interview

Dong Jun Woun

Graduate Research Assistant
Dong Jun Woun Then

Where are you from originally?

I’m from Daejeon, South Korea. Even though it’s been many years since I left, I still root for the Hanwha Eagles, my favorite baseball team in the KBO league. One of my fondest childhood memories is visiting the Sungsingdam bakery, which has the best fried peanut streusel filled with red bean paste.

Can you summarize your educational and professional background?

I recently graduated from the University of Tennessee with a bachelor’s degree in Computer Science, with minors in Cyber Security and Machine Learning. I’m continuing my studies as a graduate student at UT and joining ICL as a graduate research assistant under the guidance of Heike Jagode. During my undergraduate studies, I had the opportunity to intern at Oak Ridge National Laboratory, where I was mentored by Dr. Prasanna Date.

How did you first hear about ICL, and what made you want to work here?

Throughout my undergraduate years at UT, I read and heard about ICL. I’ve always aimed to apply my skills to make a positive difference in the world, and ICL’s mission resonated with that goal. I’m excited to be part of ICL and look forward to contributing to the innovative projects here.

What are your main research interests and what do you work on at ICL?

My primary research interests lie in machine learning, performance counting, and cybersecurity. As a Research Intern at Oak Ridge National Laboratory, I had the chance to compare a machine learning model’s performance on a quantum computer versus a traditional computer.

Additionally, I’ve always been interested in hardware benchmarking, inspired by watching videos from channels like Hardware Unboxed, Gamer’s Nexus, and BullsLab. I’ve used benchmarking tools to optimize performance on my personal devices. I’m thrilled to join the performance group at ICL and work on benchmarking for High-Performance Computers.

What would you like to do next in your career?

I aim to continue applying my skills to create and improve technology, driving meaningful change. Although I initially entered college as a Civil Engineering major, my passion for using engineering to make a positive impact has remained consistent.

What are your interests/hobbies outside of work?

In my free time, I enjoy learning about history, particularly the industrial era. I also like working out, lifting weights and running. Additionally, I’m a fan of video games, mainly strategy games like Civilization 6 or Stellaris. I also enjoy multiplayer games like Chained Together or League of Legends.

Tell us something about yourself that might surprise people.

I’ve experienced moving every two to three years, having relocated four times. I’ve lived in Daejeon, Korea for 9 years, Budapest, Hungary for four years, and Tennessee for 9 years, spending five years in Clarksville and four years in Knoxville.

Recent Papers

  1. Cojean, T., P. Nayak, T. Ribizel, N. Beams, Y-H. Mike Tsai, M. Koch, F. Göbel, T. Grützmacher, and H. Anzt, Ginkgo - A math library designed to accelerate Exascale Computing Project science applications,” The International Journal of High Performance Computing Applications, August 2024. DOI: 10.1177/10943420241268323
  2. Dongarra, J., and D. Keyes, The co-evolution of computational physics and high-performance computing,” Nature Reviews Physics, August 2024. DOI: 10.1038/s42254-024-00750-z
  3. Gates, M., A. Abdelfattah, K. Akbudak, M. Al Farhan, R. Alomairy, D. Bielich, T. Burgess, ébastien. Cayrols, N. Lindquist, D. Sukkari, et al., Evolution of the SLATE linear algebra library,” The International Journal of High Performance Computing Applications, September 2024. DOI: 10.1177/10943420241286531
  4. Luszczek, P., A. Castaldo, Y. M. Tsai, D. Mishler, and J. Dongarra, Numerical eigen-spectrum slicing, accurate orthogonal eigen-basis, and mixed-precision eigenvalue refinement using OpenMP data-dependent tasks and accelerator offload,” The International Journal of High Performance Computing Applications, vol. 303, issue 136, September 2024. DOI: 10.1177/10943420241281050

Recent Conferences

  1. AUG
    -
    Open Science Grid School 2024 Madison, Wisconsin
    Gerald Ragghianti
    Gerald
    Gerald Ragghianti
  2. AUG
    -
    2024 CSSI PI meeting Charlotte, North Carolina
    Aurelien Bouteiller
    Aurelien
    Mark Gates
    Mark
    Aurelien Bouteiller, Mark Gates
  3. AUG
    -
    Deborah Penchoff
    Deborah
    Deborah Penchoff
  4. AUG
    -
    Ahmad Abdelfattah
    Ahmad
    Ahmad Abdelfattah
  5. SEP
    -
    CCDSC 2024 Lyon, France
    Jack Dongarra
    Jack
    Piotr Luszczek
    Piotr
    Jack Dongarra, Piotr Luszczek
  6. SEP
    -
    Heidelberg Laureate Forum Heidelberg, Germany
    Jack Dongarra
    Jack
    Maksim Melnichenko
    Maksim
    Jack Dongarra, Maksim Melnichenko

Upcoming Conferences

  1. OCT
    -
    SIAM MDS24 Atlanta, Georgia
    Maksim Melnichenko
    Maksim
    Maksim Melnichenko

Recent Lunch Talks

  1. AUG
    23
    Michael Jantz
    Michael Jantz
    UTK EECS
    The ECP SICM High-Level Interface: Automated Management of Complex Memories with Application Guidance PDF
  2. SEP
    6
    Tokey Tahmid
    Tokey Tahmid
    Low Precision and Efficient Programming Languages for Sustainable AI PDF
  3. SEP
    13
    Jack Dongarra
    Jack Dongarra
    An Overview of High Performance Computing PDF
  4. SEP
    19
    Hartwig Anzt
    Hartwig Anzt
    ICL / TU Munich
    Sparse BLAS Working Group: On the path to defining a standard for sparse BLAS operations PDF

Upcoming Lunch Talks

  1. OCT
    4
    Jada Huskey & Kiley Compton
    Jada Huskey & Kiley Compton
    UTK College of Engineering Research Office
  2. OCT
    11
    Piotr Luszczek
    Piotr Luszczek
    MIT / ICL
    Great Expectations for SC24 or All Quiet on HPC Front? PDF
  3. OCT
    18
    Shirley Moore
    Shirley Moore
    UTEP
    Performance Modeling of Distributed Deep Learning Applications… without Buying Another Nuclear Power Plant PDF

People

  1. Deborah Penchoff
    Deborah Penchoff recently started a new position as an Assistant Professor at the University of Central Florida Department of Chemistry. Congratulations and best wishes to Deborah on her new position!
  2. Dong Jun Woun
    Dong Jun Woun joined ICL as a Graduate Research Assistant working with the Performance group. Welcome Jun!
  3. Dayle Flanigan
    ICL welcomes Dayle Flanigan to its main office as a new Financial Assistant.

Dates to Remember

Save the Date

Save the date! The SC24 ICL dinner will be on Wednesday,November 20th. Time and location TBA.

Open Positions at ICL

Post Doctoral Research Associate

ICL has an opening for a full-time postdoctoral research associate in the linear algebra group. We invite applicants with fresh PhD degrees, or those who are graduating within 3 months, to apply for this opportunity.

The position is available immediately. The candidate’s primary responsibility is to work on the distributed implementation of the Fast Fourier Transform (FFT) on large scale HPC systems with GPUs. This includes performance benchmarking on several systems with different MPI implementations, performance optimizations, profiling, analysis, and developing new features for distributed FFT computation as necessary.

More Information

Around the Web

Ewa Deelman

@thoefler