News and Announcements

Top500 – U.S. Reclaims the Top Spot

June 2012 TOP500 HandoutIn the 39th edition of the Top500, IBM’s Sequoia has beat Fujitsu’s K Computer to become the fastest supercomputer on the planet. The Sequoia, a BlueGene/Q powered behemoth, is housed at the Lawrence Livermore National Laboratory in Livermore, California, and achieved 16.32 petaflop/s on the LINPACK benchmark using 1,572,864 cores.  Sequoia’s raw performance is staggering, yes, but the machine is also one of the most energy efficient systems on the Top500.

The new #1 was not the only activity at the top of the list, as several machines were bumped in position, and 2 new systems made the top 10. Argonne National Laboratory’s Mira, another IBM BlueGene/Q machine in the U.S., debuted in 3rd place with 8.15 petaflop/s on the LINPACK benchmark using 786,432 cores. Italy has an IBM BlueGene/Q of their own, Fermi, which marks their first entry into the top 10 with 1.72 petaflop/s using 163,840 cores. In all, 4 of the top 10 supercomputers are IBM BlueGene/Q systems.

Rank Site Computer
1 DOE/NNSA/LLNL
United States
Sequoia – BlueGene/Q, Power BQC 16C 1.60 GHz, Custom
IBM
2 RIKEN Advanced Institute for Computational Science (AICS)
Japan
K computer, SPARC64 VIIIfx 2.0GHz, Tofu interconnect
Fujitsu
3 DOE/SC/Argonne National Laboratory
United States
Mira – BlueGene/Q, Power BQC 16C 1.60GHz, Custom
IBM
4 Leibniz Rechenzentrum
Germany
SuperMUC – iDataPlex DX360M4, Xeon E5-2680 8C 2.70GHz, Infiniband FDR
IBM
5 National Supercomputing Center in Tianjin
China
Tianhe-1A – NUDT YH MPP, Xeon X5670 6C 2.93 GHz, NVIDIA 2050
NUDT
See the full list at Top500.org.

Hear Rich Brueckner and Jack Dongarra discuss the new TOP500 list by listening to the following podcast from InsideHPC:

In this podcast, Jack Dongarra from the University of Tennessee discusses the June 2012 TOP500 list of the world’s most powerful supercomputers. The TOP500 is celebrating its 20th anniversary this year and Jack tells us more about how it got started and how the evolution of supercomputer performance measurement continues to evolve.

[audio:http://icl.cs.utk.edu/newsletter/files/2012-07/audio/DongarraPodcast.mp3]

ICL Retreat 2012

As a reminder, mark your calendars for August 16-17 for the 2012 ICL Retreat! This year, we return to the Buckberry Lodge in Gatlinburg.

Shirley Moore

Shirley Moore, who has worked for ICL since January 1993, accepted a faculty position at the University of Texas at El Paso (UTEP) for the fall of 2012. She will be an Associate Professor of Computational and Computer Science (split position between the graduate Computational Science Program and the Computer Science Department). Her research at UTEP will focus on using code restructuring and performance tools to enhance the performance, maintainability, and evolvability of parallel scientific applications.

On Friday, June 29th, Shirley gave her final lunch talk as an ICL employee where she reviewed many of the things she worked on in her 19.5 years at ICL, and discussed what the future holds for her at UTEP. Shirley was given a warm ICL send off from director Jack Dongarra and the rest of the ICL crew. A collection of photos from Shirley’s time at ICL can be viewed here. Good luck, Shirley!

LINPACK is Still Going Strong

In a recent HPCwire interview, Tom Tabor had an in-depth discussion with Hans Meuer about the impact of the Top500 list over the span of its nearly 20 year existence. Hans explained how the list is compiled based on the best LINPACK performance, and how Jack Dongarra—dubbed the “father of LINPACK” by Meuer—joined the effort in 1993.

The use of LINPACK to rank every machine on the Top500 is significant, especially considering the list is going on 20 years old and has used basically the same benchmark in an application where the machines have become significantly larger and more powerful. A bit of what Meuer had to say about LINPACK in the Top500:

We have been criticized for choosing LINPACK from the very beginning, but now in the 20th year, I believe that it was this particular choice that has contributed to the success of TOP500. Back then and also now, there simply isn’t an appropriate alternative to LINPACK. Any other benchmark would appear similarly specific, but would not be so readily available for all systems in question. One of LINPACK’s advantages is its scalability, in the sense that it has allowed us for the past 19 years to benchmark systems that cover a performance range of more than 11 orders of magnitude.

To read more about the LINPACK benchmark and the role of the Top500 in HPC, check out the HPCwire interview in its entirety, here.

Conference Reports

International Conference on Computational Science

This year’s International Conference on Computational Science (ICCS) was held in Omaha, Nebraska on June 4-6, and was the twelfth meeting of the ICCS. The theme for this year’s meeting was “Empowering Science through Computing,” in order to emphasize the growing importance of computational science theory and practice.

Several ICLers attended the meeting, including Jack Dongarra, Peng Du, Ichitaro Yamazaki, and Mark Gates. Jack moderated a session of GPU related papers in which Mark presented work on block-asynchronous multigrid smoothers—a team effort with Hartwig and Stan. Ichi also presented Yulu’s mutli-core, multi-GPU paper and some work on multi-core LU and Cholesky factorizations. In a separate session, Peng presented a paper on fault tolerant LU. ICL alumnus Marc Baboulin also presented a paper on the random butterfly applied to LU, a paper that was co-authored by Jack, Simplice, and Stan.

International Supercomputing Conference

  • 010
  • 009
  • 008
  • 007
  • 006
  • 005
  • 004
  • 003
  • 002
  • 001
  • 000

The 2012 International Supercomputing Conference (ISC’12), held on June 17-21 at the Congress Center in Hamburg, Germany, drew a record 2,403 attendees and 175 of the world’s leading high performance computing vendors and research organizations.

During the proceedings, the University of Tennessee’s Jack Dongarra was named an ISC Fellow along with Thomas Ludwig of the German Climate Computing Center. Jack also gave a talk highlighting the latest accomplishments of the MAGMA project during the Application Performance session, and presented Top500 awards during the conference’s opening session.

Jakub Kurzak, Hatem Ltaief, and Jack Dongarra also gave a tutorial on Dense Linear Algebra Libraries for High Performance Computing, including a retrospective of LINPACK, LAPACK, and ScaLAPACK, a look at developments in PLASMA and MAGMA, and the ongoing efforts in linear algebra software for distributed memory machines (DAGuE/DPLASMA).

Recent Releases

MAGMA 1.2.1 Released

MAGMA 1.2.1 is now available for download. This release includes the following updates:

  • Removed dependency on CUDA device driver.
  • Added magma_malloc_cpu to allocate CPU memory aligned to 32-byte boundary for performance and reproducibility.
  • Added QR with pivoting in CPU interface (functions [zcsd]geqp3).
  • Added hegst/sygst Fortran interface.
  • Improved performance of gesv CPU interface by 30%.
  • Improved performance of ungqr/orgqr CPU and GPU interfaces by 30%; more for small matrices.
  • Numerous bug fixes.

Visit the MAGMA website to download the tarball.

clMAGMA 0.3 Released

clMAGMA 0.3 is now available. This release provides OpenCL support for the main two-sided matrix factorizations, the basis for eigenproblem and SVD solvers. clMAGMA 0.3 includes the following new functionalities:

  • Reduction to upper Hessenberg form by similarity orthogonal transformations (routines magma_{z|c|d|s}gehrd ).
  • Reduction to upper/lower bidiagonal form by similarity orthogonal transformations (routines magma_{z|c|d|s}gebrd ).
  • Reduction to tridiagonal form by similarity orthogonal transformations ( routines magma_{zhe|che|dsy|ssy}trd ).

Visit the MAGMA website to download the tarball.

Interview

Wesley Bland Then

Wesley Bland

Where are you from, originally?
I was born in Knoxville and have spent most of my life here.

Can you summarize your educational background?
I did my undergrad at Tennessee Tech in Cookeville, TN. While I was there, I spent my summers at ORNL working internships with two different groups. I came to UTK in 2007 and received my M.S. in 2009. I’m still here and still working on my Ph.D. which I’m targeting for next year.

Tell us how you first learned about ICL.
I learned about the ICL when I came to grad school at UTK. I spent my first year here as a graduate teaching assistant for CS 100. Wes Alvaro was in one of my classes and talked about the ICL, so I came to one of the recruiting lunches. That summer I talked to George and his team, and joined the ICL in the fall.

What made you want to work for ICL?
The ICL seemed like the group at UTK that was doing work most closely related to my previous research. I was interested in joining a group where I would be able to keep working on HPC problems while pursuing my Ph.D. I ended up doing something completely different than my undergraduate work, but it’s been great to learn new areas of computer science.

What are you working on while at ICL?
I work in the distributed computing group on MPI-level fault tolerance. Most recently, I’ve been working with George, Thomas, and Aurelien on an effort to standardize and implement fault tolerance in the MPI forum. I’m also working with teams from the University of Georgia and Vanderbilt University to bring FT techniques to their applications.

What do you plan on doing after graduation?
My goal is to eventually get back into a university as a professor. I’ve always enjoyed teaching and want to continue doing that. In the meantime, I’d like to try something a little different than academic research. I’d like to try out a job in industry for a while.

What are your interests/hobbies outside of work?
I enjoy all the usual nerd things: books, video games, side programming projects, science fiction. I also coach a basketball team at my church as well as participating in various musical groups. I follow Manchester United whenever possible, and I play with my dog Ada (named after who you think she’s named after) with my wife Julie.

Tell us something about yourself that might surprise people.
I started programming when I was 8. The library at school had a book called “Learn to make your own computer games” which had a bunch of simple BASIC programs like Guess a Number and Tic-Tac-Toe. I checked it out and compiled them all on our old computer at home. Between that and the fact that both of my parents were computer scientists, I guess there wasn’t much choice about what I was going to end up doing.

Recent Papers

  1. Baboulin, M., S. Donfack, J. Dongarra, L. Grigori, A. Remi, and S. Tomov, A Class of Communication-Avoiding Algorithms for Solving General Dense Linear Systems on CPU/GPU Parallel Machines,” Proc. of the International Conference on Computational Science (ICCS), vol. 9, pp. 17-26, June 2012.
  2. Song, F., and J. Dongarra, A Scalable Framework for Heterogeneous GPU-Based Clusters,” The 24th ACM Symposium on Parallelism in Algorithms and Architectures (SPAA 2012), Pittsburgh, PA, USA, ACM, June 2012.  (3.39 MB)
  3. Anzt, H., S. Tomov, M. Gates, J. Dongarra, and V. Heuveline, Block-asynchronous Multigrid Smoothers for GPU-accelerated Systems,” ICCS 2012, Omaha, NE, June 2012.  (608.95 KB)
  4. Song, F., S. Tomov, and J. Dongarra, Enabling and Scaling Matrix Computations on Heterogeneous Multi-Core and Multi-GPU Systems,” 26th ACM International Conference on Supercomputing (ICS 2012), San Servolo Island, Venice, Italy, ACM, June 2012.  (5.88 MB)
  5. Du, P., P. Luszczek, and J. Dongarra, High Performance Dense Linear System Solver with Resilience to Multiple Soft Errors,” ICCS 2012, Omaha, NE, June 2012.  (1.27 MB)
  6. Yamazaki, I., S. Tomov, and J. Dongarra, One-Sided Dense Matrix Factorizations on a Multicore with Multiple GPU Accelerators,” The International Conference on Computational Science (ICCS), June 2012.
  7. Bosilca, G., A. Bouteiller, E. Brunet, F. Cappello, J. Dongarra, A. Guermouche, T. Herault, Y. Robert, F. Vivien, and D. Zaidouni, Unified Model for Assessing Checkpointing Protocols at Extreme-Scale,” University of Tennessee Computer Science Technical Report (also LAWN 269), no. UT-CS-12-697, June 2012.  (2.76 MB)
  8. Baboulin, M., D. Becker, G. Bosilca, A. Danalis, and J. Dongarra, An efficient distributed randomized solver with application to large dense linear systems,” ICL Technical Report, no. ICL-UT-12-02, July 2012.  (626.26 KB)
  9. Langou, J., B. Hoffman, and B. King, How LAPACK library enables Microsoft Visual Studio support with CMake and LAPACKE,” University of Tennessee Computer Science Technical Report (also LAWN 270), no. UT-CS-12-698, July 2012.  (501.53 KB)
  10. Abdelfattah, A., J. Dongarra, D. Keyes, and H. Ltaeif, Optimizing Memory-Bound Numerical Kernels on GPU Hardware Accelerators,” VECPAR 2012, Kobe, Japan, July 2012.  (737.28 KB)
  11. Kasichayanula, K., D. Terpstra, P. Luszczek, S. Tomov, S. Moore, and G. D. Peterson, Power Aware Computing on GPUs,” SAAHPC '12 (Best Paper Award), Argonne, IL, July 2012.  (658.06 KB)
  12. Haidar, A., H. Ltaeif, and J. Dongarra, Toward High Performance Divide and Conquer Eigensolver for Dense Symmetric Matrices,” SIAM Journal on Scientific Computing (Accepted), July 2012.

Recent Lunch Talks

  1. JUN
    1
    Kiran Kasichayanula
    Kiran Kasichayanula
    Power Aware Computing on GPUs PDF
  2. JUN
    8
    James Plank
    James Plank
    EECS
    Smoke, Mirrors and Erasure Codes PDF
  3. JUN
    15
    Mark Gates
    Mark Gates
    Multi-GPU Hessenberg reduction in MAGMA PDF
  4. JUN
    22
    Branislav Jansik
    Branislav Jansik
    Aarhus University
    DEC code implementation and parallelization PDF
  5. JUN
    29
    Shirley Moore
    Shirley Moore
    Shirley Moore Photo Gallery URL
  6. JUL
    6
    Mathieu Faverge
    Mathieu Faverge
    Algorithms mixing LU and QR kernels for multi-core cluster systems PDF
  7. JUL
    13
    Aurelien Bouteiller
    Aurelien Bouteiller
    Fault Tolerance Methods: Challenges and Opportunities PDF
  8. JUL
    20
    Erlin Yao
    Erlin Yao
    A Case Study of Designing Efficient Algorithm-based Fault Tolerant Application for Exascale Parallelism PDF
  9. JUL
    27
    Anthony Danalis
    Anthony Danalis
    From Serial Loops to Parallel Execution on Distributed Systems PDF

Upcoming Lunch Talks

  1. AUG
    3
    Stanimire Tomov
    Stanimire Tomov
    clMAGMA: Heterogeneous High-Performance Linear Algebra with OpenCL PDF
  2. AUG
    10
    Ichitaro Yamazaki
    Ichitaro Yamazaki
    Pivoting strategies for solving symmetric indefinite linear systems PDF
  3. AUG
    24
    Azzam Haidar
    Azzam Haidar
    Eigenproblem Algorithms PDF
  4. AUG
    31
    Yutaka IshikawaMitsuhisa Sato
    Yutaka Ishikawa and Mitsuhisa Sato
    Japan
    Updates on XcalableMP PGAS Language PDF

People

  1. Fengguang Song
    Fengguang Song has accepted a job as a researcher in the Computer Science Lab at the Samsung Research Center in San Jose, California.  He starts October 1st and will remain with ICL through September.  Congratulations and good luck!
  2. Shirley Moore
    Shirley Moore, who has worked for ICL since January 1993, has accepted a faculty position at the University of Texas at El Paso (UTEP) for the fall of 2012.  Good luck, Shirley!