News and Announcements

HPCG: A New Performance Metric

jack_dongarra

Since 1993, ICL’s Jack Dongarra has maintained a ranking of the world’s top performing supercomputers. The semiannual TOP500 list is compiled using ICL’s High Performance Linpack benmark (HPL). HPL and the TOP500 are the most widely recognized and discussed metrics for ranking HPC systems. However, the Linpack benchmark, which relies on solving dense linear equations to calculate performance, is becoming less relevant as it does not depict the real-world performance of modern HPC applications, which are moving towards differential equations.

Very aware of this performance ‘gap,’ Jack teamed up with Michael Heroux from Sandia National Laboratories to develop the High Performance Conjugate Gradient (HPCG) benchmark, which—because it is composed of computations and data access patterns more commonly found in applications—will hopefully allow for a better correlation to real scientific application performance and drive computer system design and implementation in directions that will better impact performance improvement.

Linpack will not be retired, however. Instead, HPCG will serve as an alternative ranking of the TOP500 list, allowing a re-ranking of the systems on the list to “real” applications—not unlike the re-shuffling of the list done for the Green500. Dongarra hopes to debut the HPCG benchmark at SC13.

News of PaRSEC Award Spreads

As reported last month, DOE awarded funding to ICL for the Parallel Runtime Scheduling and Execution Controller (PaRSEC) project. The award, which allocates ICL/UTK $1 million dollars over three years, was enough to grab the attention of HPC Wire (by way of TN Today). Congratulations once again!

The Parallel Runtime Scheduling and Execution Controller (PaRSEC) is a generic framework for architecture aware scheduling and management of micro-tasks on distributed many-core heterogeneous architectures. Applications we consider can be expressed as a Direct Acyclic Graph of tasks with labeled edges designating data dependencies. DAGs are represented in a compact problem-size independent format that can be queried on-demand to discover data dependencies in a totally distributed fashion.

PaRSEC assigns computation threads to the cores, overlaps communications and computations and uses a dynamic, fully-distributed scheduler based on architectural features such as NUMA nodes and algorithmic features such as data reuse. The framework includes libraries, a runtime system, and development tools to help application developers tackle the difficult task of porting their applications to highly heterogeneous and diverse environments.

SILAS Project Funded

The Sustained Innovation for Linear Algebra Software (SILAS) project, a collaborative effort between UTK, U of C Denver, and UC Berkeley, was recently funded by an NSF SSI call. SILAS updates two of the most widely used numerical libraries in the history of computational science and engineering—LAPACK and ScaLAPACK, (abbreviated Sca/LAPACK)—enhancing and hardening them for this ongoing revolution in processor architecture and system design.

The primary impact of SILAS is a direct function of the importance of the Sca/LAPACK libraries to many branches of computational science. The Sca/LAPACK libraries are the community standard for dense linear algebra and have been adopted and/or supported by a large community of users, computing centers, and HPC vendors.

Application domains where the Sca/LAPACK libraries are used include (among a host of other examples) airplane wing design, radar cross-section studies, flow around ships and other off-shore constructions, diffusion of solid bodies in a liquid, noise reduction, and diffusion of light through small particles. Enhancing these libraries through the SILAS project with state-of-the-art methods and algorithms, and adapting them for new and emerging platforms, will have a large impact on the research and education community, government laboratories, and private industry.

ADAPT Project Funded

The ADAPT project, recently funded under an NSF SSE call, enhances, hardens, and modernizes the Open MPI library and creates a viable foundation for a new generation of Open MPI components by implementing fundamental software techniques that can be used in many-core systems to efficiently execute MPI-based applications, and to tolerate fail-stop process failures, at scales ranging from current large systems to the extreme scale systems that are coming soon.

To improve the efficiency of Open MPI, ADAPT integrates, as a core component, knowledge about the hardware architecture, and allows all layers of the software stack full access to this information. Process placement, distributed topologies, file accesses, point-to-point and collective communications can then adapt to such topological information, providing more portability.

The ADAPT team, which includes UTK and the University of Houston, is also updating the current collective communication layer to allow for a task-based collective description contained at a group-level, which in turn adjusts to the intra and inter-node topology. Planned expansion of the current code with resilient capabilities allows Open MPI to efficiently survive hard and soft error types of failures. These capabilities can be used as building blocks for all currently active fault tolerance proposals in the MPI standard body.

Interview

Blake Haugen Then

Blake Haugen

Where are you from, originally?

I grew up in a small town (~10,000 people) in Northeast Iowa called Waverly, known for having the largest horse sale in the nation. One of the greatest perks of living in Waverly was the smell that came from the local Nestle plant.

Can you summarize your educational background?

I attended a small Lutheran, liberal arts school called Wartburg College. My majors were Engineering Science and Computer Science with a double minor in Math and Music. Upon graduation in May of 2010, I chose to attend the University of Tennessee to pursue my master’s degree in Computer Science, which I completed in May of 2012. I am currently working toward my PhD at UT.

Tell us how you first learned about ICL.

My first exposure to ICL was through a scientific computing course I took during my junior year of college. A large portion of the material covered was parallel computing and we discussed the Top 500 list. This class sparked my interest in HPC as well as ICL.

What made you want to work for ICL?

I knew I wanted to study HPC and scientific computing which made the ICL a real “no-brainer.” The world-class research and open source software speak for themselves.

What are you working on while at ICL?

Most of my time is spent working on the PLASMA project where I focus on “autotuning.” This includes improving algorithm performance prediction in an attempt to automatically select algorithm parameters to maximize performance.

If you weren’t working at ICL, where would you like to be working and why?

If I wasn’t working at ICL, I would likely be employed at a software company developing tools used by the scientific and business communities. However, if I had pursued a different career path, I could have been a musician or music teacher.

What are your interests/hobbies outside work?

I have played the trombone since elementary school and am currently a member of the East Tennessee Concert Band. I also enjoy cooking but I know I can always order a pizza when things don’t go according to plan.

Tell us something about yourself that might surprise people.

Last year I began participating in a curling league. Yes, the one with brooms and stones on the ice! You would be shocked to learn the sport exists outside the winter Olympics and has a vibrant community even here in Knoxville, TN.

Recent Papers

  1. Donfack, S., S. Tomov, and J. Dongarra, Dynamically balanced synchronization-avoiding LU factorization with multicore and GPUs,” University of Tennessee Computer Science Technical Report, no. ut-cs-13-713, July 2013.  (659.77 KB)
  2. Bland, W., P. Du, A. Bouteiller, T. Herault, G. Bosilca, and J. Dongarra, Extending the scope of the Checkpoint-on-Failure protocol for forward recovery in standard MPI,” Concurrency and Computation: Practice and Experience, July 2013. DOI: 10.1002/cpe.3100  (3.89 MB)
  3. Dong, T., V. Dobrev, T. Kolev, R. Rieben, S. Tomov, and J. Dongarra, Hydrodynamic Computation with Hybrid Programming on CPU-GPU Clusters,” University of Tennessee Computer Science Technical Report, no. ut-cs-13-714, July 2013.  (866.68 KB)
  4. Ma, T., G. Bosilca, A. Bouteiller, and J. Dongarra, Kernel-assisted and topology-aware MPI collective communications on multi-core/many-core platforms,” Journal of Parallel and Distributed Computing, vol. 73, issue 7, pp. 1000-1010, July 2013. DOI: 10.1016/j.jpdc.2013.01.015  (1.4 MB)
  5. Donfack, S., J. Dongarra, M. Faverge, M. Gates, J. Kurzak, P. Luszczek, and I. Yamazaki, On Algorithmic Variants of Parallel Gaussian Elimination: Comparison of Implementations in Terms of Performance and Numerical Properties,” University of Tennessee Computer Science Technical Report, no. UT-CS-13-715, July 2013, 2012.  (358.98 KB)
  6. Nelson, J., Analyzing PAPI Performance on Virtual Machines,” ICL Technical Report, no. ICL-UT-13-02, August 2013.  (437.37 KB)
  7. Kurzak, J., P. Luszczek, and J. Dongarra, LU Factorization with Partial Pivoting for a Multicore System with Accelerators,” IEEE Transactions on Parallel and Distributed Computing, vol. 24, issue 8, pp. 1613-1621, August 2013. DOI: http://doi.ieeecomputersociety.org/10.1109/TPDS.2012.242  (1.08 MB)
  8. Bouteiller, A., F. Cappello, J. Dongarra, A. Guermouche, T. Herault, and Y. Robert, Multi-criteria Checkpointing Strategies: Response-Time versus Resource Utilization,” Euro-Par 2013, Aachen, Germany, Springer, August 2013.  (431.84 KB)

Recent Lunch Talks

  1. JUL
    19
    Anthony Danalis
    Anthony Danalis
    Creating a new operation with DPLASMA: a step by step guide PDF
  2. AUG
    23
    Michela Taufer
    Michela Taufer
    University of Delaware
    On the effectiveness of application-aware self-management for scientific discovery in volunteer computing systems PDF
  3. AUG
    30
    Jeff Larkin
    Jeff Larkin
    NVIDIA
    OpenACC 2.0 Highlights PDF

Upcoming Lunch Talks

  1. SEP
    6
    Yves Robert
    Yves Robert
    On the Combination of Silent Error Detection and Checkpointing PDF
  2. SEP
    13
    Jakub Kurzak
    Jakub Kurzak
    Parallel Ultra Light Systolic Array Runtime PDF
  3. SEP
    20
    Piotr Luszczek
    Piotr Luszczek
    Energy and Power Consumption Trends PDF
  4. SEP
    27
    Dan Terpstra
    Dan Terpstra
    Small Scale Water Treatment in Developing Countries PDF

Visitors

  1. Volodymyr Turchenko
    Volodymyr Turchenko from Research Institute of Intelligent Computer Systems, Ternopil National Economic University, Ukraine will be visiting from August 27 through August 27. Volodymyr will be interacting with the Linear Algebra and Distributed Computing groups.
  2. Shirley Moore
    Shirley Moore from the University of Texas at El Paso will be visiting from July 15 through August 16. Shirley will be working with the performance analysis group druing her visit.

People

  1. Thananon Patinyasakdikul
    Thananon Patinyasakdikul will be joining ICL as a new graduate student and will be working with the Distributed Computing group. Welcome aboard!
  2. Reazul Hoque
    Reazul Hoque will also be joining ICL as a new graduate student and working with the Distributed Computing group. Welcome aboard!
  3. Shirley Moore
    ICL alum Shirley Moore,  now at the University of Texas at El Paso, will be visiting ICL through August 16th. Shirley will be working with the performance analysis group and attending the ICL annual retreat.

Visitors

  1. Volodymyr Turchenko
    Volodymyr Turchenko from Research Institute of Intelligent Computer Systems, Ternopil National Economic University, Ukraine will be visiting from August 27 through August 27. Volodymyr will be interacting with the Linear Algebra and Distributed Computing groups.
  2. Shirley Moore
    Shirley Moore from the University of Texas at El Paso will be visiting from July 15 through August 16. Shirley will be working with the performance analysis group druing her visit.

Dates to Remember

ICL Retreat 2013

Mark your calendars for August 1516 for the 2013 ICL Retreat! This year, the retreat will be held at the RT Lodge in Maryville.

New Academic Year Begins

The new academic year begins on Wednesday, August 21st. Good luck to our faculty and students!