TOP
Search the Dagstuhl Website
Looking for information on the websites of the individual seminars? - Then please:
Not found what you are looking for? - Some of our services have separate websites, each with its own search option. Please check the following list:
Schloss Dagstuhl - LZI - Logo
Schloss Dagstuhl Services
Seminars
Within this website:
External resources:
  • DOOR (for registering your stay at Dagstuhl)
  • DOSA (for proposing future Dagstuhl Seminars or Dagstuhl Perspectives Workshops)
Publishing
Within this website:
External resources:
dblp
Within this website:
External resources:
  • the dblp Computer Science Bibliography


Dagstuhl Seminar 10181

Program Development for Extreme-Scale Computing

( May 02 – May 07, 2010 )


Permalink
Please use the following short url to reference this page: https://www.dagstuhl.de/10181

Organizers




Schedule

Press Room

Press Reviews


Summary

The number of processor cores available in high-performance computing systems is steadily increasing. A major factor is the current trend to use multi-core and many-core processor chip architectures. In the November 2009 list of the TOP500 Supercomputer Sites, 98.4% of the systems listed have more than 2048 processor cores and the average is about 9300. While these machines promise ever more compute power and memory capacity to tackle today's complex simulation problems, they force application developers to greatly enhance the scalability of their codes to be able to exploit it. This often requires new algorithms, methods or parallelization schemes as many well-known and accepted techniques stop working at such large scales. It starts with simple things like opening a file per process to save checkpoint information, or collecting simulation results of the whole program via a gather operation on a single process, or previously unimportant order O(n2)-type operations that now quickly dominate the execution. Unfortunately many of these performance problems only show up when executing with very high numbers of processes and cannot be easily diagnosed or predicted from measurements at lower scales. Detecting and diagnosing these performance and scalability bottlenecks requires sophisticated performance instrumentation, measurement and analysis tools. Simple tools typically scale very well but the information they provide proves to be less and less useful at these high scales. Clearly, understanding performance and correctness problems of applications requires running, analyzing, and drawing insight into these issues at the largest scale.

Consequently, a strategy for software development tools for extreme-scale systems must address a number of dimensions. First, the strategy must include elements that directly address extremely large task and thread counts. Such a strategy is likely to use mechanisms that reduce the number of tasks or threads that must be monitored. Second, less clear but equally daunting, is the fact that several planned systems will be composed of heterogeneous computing devices. Performance and correctness tools for these systems are very immature. Third, the strategy requires a scalable and modular infrastructure that allows rapid creation of new tools that respond to the unique needs that may arise as extreme-scale systems evolve. Further, a successful tools strategy must enable productive use of systems that are by definition unique. Thus, it must provide the full range of traditional software development tools, from debuggers and other code correctness tools such as memory analyzers, performance analysis tools as well as build environments for complex codes that rely on a diverse and rapidly changing set of support libraries.

Many parallel tools research groups have already started to work on scaling their methods, techniques, and tools to extreme processor counts. In this Dagstuhl seminar, we wanted participants from Universities, government laboratories and industry to report on their successes or failures in scaling their tools, review existing working and promising new methods and techniques, and discuss strategies for solving unsolved issues and problems.

This meeting was the forth in a series of seminars related to the topic "Performance Analysis of Parallel and Distributed Programs", with previous meetings being the Dagstuhl Seminar 07341 on "Code Instrumentation and Modeling for Parallel Performance Analysis" in August 2007, Seminar 02341 on "Performance Analysis and Distributed Computing" held in August 2002, and Seminar 05501 on "Automatic Performance Analysis" in December 2005.

The seminar brought together a total of 46 researchers and developers working in the area of performance from universities, national research laboratories and, especially important, from three major computer vendors. The goals were to increase the exchange of ideas, knowledge transfer, foster a multidisciplinary approach to attacking this very important research problem with direct impact on the way in which we design and utilize parallel systems to achieve high application performance.


Participants
  • Dorian C. Arnold (University of New Mexico - Albuquerque, US) [dblp]
  • David Boehme (Forschungszentrum Jülich, DE) [dblp]
  • Michael J. Brim (University of Wisconsin - Madison, US)
  • Holger Brunst (TU Dresden, DE) [dblp]
  • Marc Casas Guix (Barcelona Supercomputing Center, ES)
  • I-hsin Chung (IBM TJ Watson Research Center - Yorktown Heights, US) [dblp]
  • Bronis R. de Supinski (LLNL - Livermore, US)
  • Luiz DeRose (Cray Inc. - Saint Paul, US)
  • Evelyn Duesterwald (IBM TJ Watson Research Center - Yorktown Heights, US)
  • Karl Fürlinger (LMU München, DE) [dblp]
  • Jim Galarowicz (The Krell Institute - Ames, US)
  • Todd Gamblin (LLNL - Livermore, US) [dblp]
  • Markus Geimer (Jülich Supercomputing Centre, DE) [dblp]
  • Michael Gerndt (TU München, DE) [dblp]
  • Judit Gimenez (Barcelona Supercomputing Center, ES) [dblp]
  • Tobias Hilbrich (TU Dresden, DE)
  • Jeffrey K. Hollingsworth (University of Maryland - College Park, US) [dblp]
  • Kevin A Huck (Barcelona Supercomputing Center, ES)
  • Karen L. Karavanic (Portland State University, US) [dblp]
  • Bettina Krammer (University of Versailles, FR)
  • Dieter Kranzlmüller (LMU München, DE) [dblp]
  • Madhavi Krishnan (University of Wisconsin - Madison, US)
  • Jesus Labarta (Barcelona Supercomputing Center, ES) [dblp]
  • David Lecomber (Allinea Software Ltd - Warwick, GB)
  • Chee Wai Lee (University of Oregon, US)
  • Matthew LeGendre (University of Wisconsin - Madison, US)
  • German Llort (Barcelona Supercomputing Center, ES)
  • David K. Lowenthal (University of Arizona - Tucson, US) [dblp]
  • Allen D. Malony (University of Oregon - Eugene, US) [dblp]
  • Barton P. Miller (University of Wisconsin - Madison, US) [dblp]
  • Bernd Mohr (Jülich Supercomputing Centre, DE) [dblp]
  • Kathryn Mohror (LLNL - Livermore, US) [dblp]
  • David Montoya (Los Alamos National Lab., US) [dblp]
  • Frank Mueller (North Carolina State University - Raleigh, US) [dblp]
  • Matthias S. Müller (TU Dresden, DE) [dblp]
  • Ramesh V. Peri (Intel - Austin, US)
  • Heidi Poxon (Cray Inc. - Saint Paul, US) [dblp]
  • Philip Roth (Oak Ridge National Laboratory, US) [dblp]
  • Nick Rutar (University of Maryland - College Park, US)
  • Martin Schulz (LLNL - Livermore, US) [dblp]
  • Harald Servat (Barcelona Supercomputing Center, ES)
  • Zoltan Szebenyi (Forschungszentrum Jülich, DE)
  • Nathan Tallent (Rice University - Houston, US) [dblp]
  • Roland Wismüller (Universität Siegen, DE)
  • Brian J. N. Wylie (Jülich Supercomputing Centre, DE)
  • Mary Zosel (LLNL - Livermore, US)

Related Seminars
  • Dagstuhl Seminar 02341: Performance Analysis and Distributed Computing (2002-08-18 - 2002-08-23) (Details)
  • Dagstuhl Seminar 05501: Automatic Performance Analysis (2005-12-12 - 2005-12-16) (Details)
  • Dagstuhl Seminar 07341: Code Instrumentation and Modeling for Parallel Performance Analysis (2007-08-19 - 2007-08-24) (Details)

Classification
  • Modeling/simulation
  • optimization/scheduling
  • porgramming languages/compiler
  • sw-engineering

Keywords
  • program instrumentation
  • performance analysis
  • parallel computing