Workshop abstracts



MPI in a Multicore environment

Håkon Bugge, Scali Inc.

This talk will discuss how the longer term trends such as multicores, memory bandwidth, and transistor budget will influence computing the following years. Which role will/should the MPI layer play in this? Further, the responsibility of the MPI layer vs. other parts of the operating environment will be discussed. Special focus will be on assignment of processes to cores for a pure MPI model and for a hybrid OpenMP/MPI environment as well. Also, techniques for optimizing the execution of a single jobs vs. throughput of several jobs will be discussed


High Performance Computing in Norway

Roy Dragseth, University of Tromsø

The talk will first give a quick overview of the HPC ecosystem in Norway and an outline of the current development in our country: two new HPC systems are installed in Bergen and Tromsø, a nationwide storage system with a dedicated lambda network is under implementation and national grid infrastructure is being established.

The second part of the talk will give an outline of the design goals and targeted applications of the new 60 TFlops cluster currently being installed at the University of Tromsø.


Modelling of materials, where we were, where we are and where are we going?

Olle Eriksson, Uppsala University

In this talk I will make a quick review of where the field of materials modelling has evolved from, by giving examples of methodology and scientific questions which were of interest some 20-30 years ago. The dramatic change of the capabilities of high performance computation, both in terms of hardware as well has software has changed this field, in a way few scientific disciplines can compare with. A few examples of research activities at Uppsala university will be given, reflecting at least to some degree which direction this field is moving in. The examples involve dynamical properties of materials, biological applications and a full automatic software system with which to undertake modelling of materials properties.


Biomedical modelling and simulation: patient specific models for diagnosis

Matts Karlsson, Linköping University

One important application of computational mechanics in biomedicine deals with creating subject specific models of the human cardiovascular systems for enhanced diagnostics and intervention planning. Magnetic resonance imaging enables non-invasive time-resolved three-dimensional data including individual anatomy as well as velocities of the blood, the myocardium and vessel walls. With the use of supercomputers we are able to perform predictive simulations of the blood flow to map the wall shear stress distribution in arteries in order to understand the mechanisms behind atherosclerosis and other diseases.


Silent Corruptions

Peter Kelemen, CERN

The capacity of the CERN Computer Center is scaling up as we prepare for LHC. As we currently have more than 4 petabytes of online magnetic storage (and expecting another 1.5 petabytes before the end of the year), data integrity issues become more and more visible. Since January 2007, CERN has has been systematically collecting and analysing observations of silent data corruptions. We present the motivations, tools and current findings of the ongoing investigation.


NDGF, a Nordic Tier-1 for WLCG

Josva Kleist, Nordic DataGrid Facility

The Tier-1 facility operated by the Nordic DataGrid Facility (NDGF) differs significantly from other Tier-1s in several aspects: It is not located one or a few locations but instead distributed throughout the Nordic, it is not under the governance of a single organization but instead a "virtual" Tier-1 build out of resources under the control of a number of different national organizations.

We present the technical implications of these aspects as well as the high-level design of this distributed Tier-1. The focus will be on the challenges involved in the creation of a storage system based on dCache. dCache is well known and respected as a powerfull distributed storage resource manager, and was chosen for implementing the storage aspects of the nordic Tier 1. In contrast to classic dCache deployments, we deploy dCache over a WAN with limitted bandwith, high latency, frequent network failures, and spanning many administrative domains. These properties provide unique challenges, covering topics such as security, administration, maintenance, upgradability, reliability, and performance. Our initial focus has been on implementing the GridFTP 2 OGF recommendation in dCache and the Globus Toolkit. Compared to GridFTP 1, GridFTP 2 allows for more intelligent data flow between clients and storage pools, thus enabling more efficient use of our limitted bandwith.


Building the European High-Performance Ecosystem

Kimmo Koski, Finnish IT center (CSC)

During 2006-2007 there has been intense work in Europe in order to increase European competitiveness in high-end computing. Various new activities have been started in addition to the existing European grid projects, such as Deisa or EGEE. European Strategy Forum for Research Infrastructures (ESFRI) has published a roadmap including plans 35 new major European infrastructures, most of them requiring high end computing, data management and software development. In addition, High-Performance Computing in Europe Taskforce (HET) was established in June 2006 with a target to draft a strategy for European HPC Ecosystem focusing petaflop computing. As a result for the successful strategy work, a collaboration among 14 countries was established and a project proposal for establishing European petaflop/s centers submitted to EU. The EU project, Partnership for Advanced Computing in Europe (PACE), will start in the beginning of 2008.

Basic tool for modeling the European HPC ecosystem is the performance pyramid. One of the key arguments from HET work—later implemented in PACE proposal—is to develop the different levels in the pyramid in a balanced way: enabling sufficient top-class resources, but at the same time invest considerably in boosting the collaboration, scaling the software, building the competencies and developing the national/regional infrastructures to support in creating a competitive and sustainable European HPC service.

The talk will review the current European collaboration in HPC and future plans in that context. Nordic status and possible opportunities are discussed. Relations between infrastructure development and scientific communities requiring HPC capacity are covered with a few practical examples. Viewpoints for efficient Nordic impact in the European HPC Ecosystem are given to boost the discussion.


Scaling and Other Bad Ideas in High Performance Computing

Erik Lindahl, Stockholm University

Superficially, molecular dynamics is a very straightforward algorithm, but with complex memory access patterns and moving particles performance has historically been limited to a few percent of the theoretical hardware peak capacity. I will present our long-term efforts in resolving this problem with our GROMACS molecular simulation toolkit, in particular our strong focus on absolute simulation performance rather than steps or relative scaling - and why parallelization in some cases even can be an extremely bad idea. Much of our optimization work is generally applicable to x86 code, and I will also present a brand new domain decomposition implementation based on "neutral territory" interaction partitioning. I will also show how reviving MPMD (multiple program, multiple data) parallelization ideas has enabled us to get strong scaling of particle simulations all the way down to 200 atoms per CPU, and how to get the most out of programs like these on large dual quad-core clusters like Neolith.


News from SNAC and SweGrid

Mats Nylén, HPC2N

The new policy for the Swedish National Allocation Committee (SNAC) is a major development of the allocation process for users of HPC resurces in Sweden. I will discuss the reasons for the change of policy and the new policy itself. The consequnces for users will be specially highlighted.

SweGrid is currently going through a major upgrade, mostly due to Swedens participation in the LHC experiments in CERN. SweGrid also provides Swedens contribution to NDGF. I will discuss the current status of the upgrade and outline the future plans for SweGrid.


Hierarchic Data Structures for Sparse Matrix Representation in Large-scale DFT/HF Calculations

Pawel Salek, KTH

Hierarchic data structure for efficient representation of sparse matrices is presented. The data structure is to allow an optimal hardware use in typical matrix operations, with sparsity patterns specific to large-scale Hartree-Fock or density functional calculations. Matrices appearing in these problems are often only semi-sparse and sparsity needs to be enforced. Algorithm for systematic truncation of small elements is presented. Efficient and accurate handling of sparsity allows to achieve linear or NlogN scaling in both storage and computational time and effectively replace the diagonalization algorithms used for determination of density from a fixed Kohn-Sham/Fock matrix. The data structure is hierarchic: At lowest level, small dense submatrices are stored, enabling use of BLAS libraries for their multiplication. Higher levels form a tree for flexibility and logarithmic cost of random-matrix element access. The data structure flexibility simplifies implementation of various matrix manipulation algorithms. Example implementations of matrix-matrix multiplications and inverse Cholesky factorization are presented. OpenMP parallelization of the developed matrix library is discussed as well.


Open Architectures for High Performance Computing

Robert Starmer, Cisco Systems

High Performance Computing (HPC) has a long history of exploting the newest technologies in order to achieve the best possible performance. Using these leading-edge technologies can have great benefits, such as rapidly accelerating scientific and industry applications. But these advances come with a cost: the bugs and instabilities typical to early adopters of immature complex systems. Utilizing open source software methodologies for new advanced technologies, particularly within the context of the HPC community, has proven to be an effective accelerator to achieve stability and performance. Additionally, HPC researchers have been invaluable in providing new insights and techniques with advanced hardware because of their access to the source code. In several notable cases, open source has therefore shortened the HPC "new product hardening" cycle and been successful in promoting widespread adoption.

This talk will discuss two HPC open source software projects that have advanced through the "new and immature" phases to become stable, production-quality systems through the help of the open source community. Cisco engineers actively participate in both of these projects:

  1. The OpenFabrics Alliance is a set of industry partners who jointly develop and maintain an open source drivers, tools, and middleware for low latency / high bandwidth InfiniBand networks.

  2. The Open MPI Project is an open source MPI-2 implementation that is developed and maintained by a consortium of academic, research, and industry partners. Open MPI embodies a fascinating blend of leading-edge research and industry-hardened product testing methodologies.


Virtualization: wLCG use cases and review of current applicable research

Owen Synge, German Electron Synchrotron, DESY, Hamburg

This talk reports upon the Virtualization users workshop held earlier this year at DESY. wLCG use cases and applications of Virtualization on the worker node became the focus of the talk. This talk will summaries the current status of the many HPC projects related to this area both from many research projects.


IBM Blue Gene/P - An Overview of a Petaflop Capable System

Carl Tengwall, IBM

In June 26, 2007, IBM announced Blue Gene/P as the leading edge offering in its massively parallel Blue Gene supercomputer line, succeeding Blue Gene/L. When fully configured, Blue Gene/P is designed to scale to at least 262,144 (256 K) quad-processor nodes, with a peak performance of 3.56 PetaFLOP/s. This presentation describes our vision of this petascale system [i.e. a system capable of delivering over a quadrillion (1015) floating point operations per second] from the hardware point of view, i.e., it provides an overview of the system architecture, chip design, system packaging and the software infrastructure. The system software design will be covered for Blue Gene/P focusing on the areas that have changed from the Blue Gene/L software design, and also our on-going directions of research for Blue Gene software.


High Performance Computing at HP

Martin Walker, Hewlett-Packard

This talk will make some observations on the size, shape, and dynamics of the high-performance computing market, then describe the view from HP on meeting the challenges that arise from this analysis. The talk will conclude with some remarks on facing the future of HPC.


info@nsc.liu.se