HPC 2010

 

High Performance Computing, GRIDS and clouds

 

An International Advanced Workshop

 

 

 

June 21 – 25, 2010, Cetraro, Italy

 

 

 

 

 

 

 

 

Final Programme

 

 

 

 

International Programme Committee

 

 

FRANK BAETKE

Global HPC Technology, Hewlett Packard

U.S.A.

CHARLIE CATLETT

Argonne National Laboratory and University of Chicago

U.S.A.

JACK DONGARRA

Innovative Computing Laboratory Computer Science Dept., University of Tennessee

U.S.A.

IAIN DUFF

Rutherford Appleton Laboratory and CERFACS

U.K. and FRANCE

IAN FOSTER

Math & Computer Science Div., Argonne National Laboratory

and Dept of Computer Science, The University of Chicago

U.S.A.

GEOFFREY FOX

Community Grid Computing Laboratory, Indiana University

U.S.A.

WOLFGANG GENTZSCH

The DEISA Project and Open Grid Forum

GERMANY

LUCIO GRANDINETTI

Dept. of Electronics, Informatics and Systems, University of Calabria

ITALY

CHRIS JESSHOPE

Faculty of Science, Informatics Institute, University of Amsterdam

NETHERLANDS

GERHARD JOUBERT

Technical University Clausthal

GERMANY

CARL KESSELMAN

University of Southern California

Information Sciences Institute

U.S.A.

JANUSZ KOWALIK

University of Gdansk

POLAND

THOMAS LIPPERT

Institute for Advanced Simulation, Juelich Supercomputing Centre

GERMANY

MIRON LIVNY

Computer Sciences Dept., University of Wisconsin

U.S.A.

IGNACIO LLORENTE

Distributed Systems Architecture Group, Dpt. de Arquitectura de Computadores y Automática, Facultad de Informática, Universidad Complutense de Madrid

SPAIN

ALBERTO MASONI

INFN – National Institute of Nuclear Physics – Italy, EU-IndiaGrid2

ITALY

SATOSHI MATSUOKA

Department of Mathematical and Computing Sciences, Tokyo Institute of Technology

JAPAN

PAUL MESSINA

Argonne Leadership Computing Facility

U.S.A.

SILVIO MIGLIORI

ENEA - Italian National Agency for New Technologies, Energy and the Environment

ITALY

DANIEL REED

Microsoft Research

U.S.A.

GILAD SHAINER

Mellanox Technologies and HPC Advisory Council

U.S.A.

PETER SLOOT

University of Amsterdam

NETHERLANDS

DOMENICO TALIA

Dept. of Electronics, Informatics and Systems, University of Calabria

ITALY

 

 

 

Organizing Committee

 

JACK

DONGARRA

LUCIO Grandinetti

MEHIDDIN Al baali

m. CARMEN incutti

MEHDI SHEIKHALISHAHI

MANOJ DEVARE

VOLODYMYR TURCHENKO

 

 

 

 

 Sponsors

 

MICROSOFT

 

 

 

 

AMD

BULL

 

 

HEWLETT PACKARD

 

 

IBM

 

 

MELLANOX TECHNOLOGIES

 

 

T-PLATFORMS

 

 

 

 

 

 

 

 

 

 

Amazon Web Services

 

 

CLUSTERVISION

 

 

CRS4  Center for Advanced Studies, Research and Development in Sardinia

 

 

ENEA - Italian National Agency for New Technologies, Energy and the Environment

 

 

EUINDIAGRID

 

 

Harvard Biomedical HPC

 

 

HPC Advisory Council

 

 

IEEE Computer Society

 

 

Inside HPC

 

 

INSTITUTE FOR THEORETICAL COMPUTER SCIENCE - Tsinghua University, China

 

 

INTEL

 

 

JUELICH SUPERCOMPUTING CENTER, Germany

 

 

KISTI -Korea Institute of Science and Technology Information

 

 

NEC

 

 

NICE

 

 

Platform Computing

 

 

SCHOOL of COMPUTER SCIENCE and TECHNOLOGY- Huazhong University, China

 

 

TABOR COMMUNICATIONS – HPC Wire

 

 

T-Systems

 

 

UNIVERSITY OF CALABRIA, Italy

 

 

 

 

Free Subscriptions! Stay up-to-date with the latest news for supercomputing professionals. Subscribe to HPCwire today and gain access to peer discussions, premium content, and HPC industry resource tools including market trends, career placement, and live coverage of major high performance computing events.

Request your FREE subscription today!

 

 

 

insideHPC.com is the web's leading source for up-to-the-minute news, commentary, exclusive features and analysis about everything HPC with an average readership of 1.5 million page views per month. If your organization wants to reach the global community of supercomputing center managers, HPC thought leaders, senior corporate executives, program managers, and HPC practitioners, ranging from computational scientists to system administrators, then you’ll want to turn to insideHPC.

Too busy to keep up? Subscribe to the daily email updates, RSS news feed, or even follow us on Twitter. As a service to the HPC community, there is no cost to subscribe to insideHPC.

 

 

 

Free Amazon Web Service credits for all HPC 2010 delegates

 

Amazon is very pleased to be able to donate $100 in service credits to all HPC 2010 delegates, which will be delivered via email. Since early 2006, Amazon Web Services (AWS) has provided companies of all sizes with an infrastructure web services platform in the cloud. With AWS you can requisition compute power, storage, and other services–gaining access to a suite of elastic IT infrastructure services as you demand them. With AWS you have the flexibility to choose whichever development platform or programming model makes the most sense for the problems you’re trying to solve.

 

 

 

 

Speakers

 

Paolo Anedda

CRS4 Center for Advanced Studies, Research and Development in Sardinia

Cagliari

ITALY

 

Piotr Arlukowicz

University of Gdansk

POLAND

 

Marcos Athanasoulis

Harvard Medical School

Harvard University

USA

 

Frank Baetke

Global HPC Technology

Hewlett Packard

Richardson, Texas

USA

 

Bruce Becker

South African National Grid

Pretoria

SOUTH AFRICA

 

Gianfranco Bilardi

Dept. of Electronics and Informatics

Faculty of Engineering

University of Padova

Padova

ITALY

 

George Bosilca

Innovative Computing Lab

University of Tennessee

Knoxville

USA

 

Marian Bubak

University of Science and Technology

Krakow

POLAND

and

Informatics Institute, University of Amsterdam

Asmterdam

THE NETHERLANDS

 

Charlie Catlett

Argonne National Laboratory

Argonne, IL

USA

 

Mathias Dalheimer

Fraunhofer Institute for Industrial Mathematics

GERMANY

 

Tim David

Centre for Bioengineering

University of Canterbury

Christchurch

NEW ZEALAND

 

Manoj Devare

Dept. of Electronics, Informatics and Systems

University of Calabria

Rende, CS

ITALY

 

Sudip S. Dosanjh

SANDIA National Labs

Albuquerque, NM

USA

 

Skevos Evripidou

Department of Computer Science

University of Cyprus

Nicosia

CYPRUS

 

Jose Fortes

Advanced Computing and Information Systems (ACIS) Lab

and

NSF Center for Autonomic Computing (CAC)

University of Florida

Gainesville, FL

USA

 

Ian Foster

Argonne National Laboratory

and

Dept. of Computer Science

The University of Chicago

Argonne & Chicago, IL

USA

 

Guang Gao

University of Delaware

Department of Electrical and Computer Engineering

Newark, Delaware

USA

 

Alfred Geiger

T-Systems Solutions for Research GmbH

Stuttgart

GERMANY

 

Wolfgang Gentzsch

DEISA Distributed European Infrastructure for Supercomputing Applications

and

OGF

GERMANY

 

Vladimir Getov

School of Electronics and Computer Science

University of Westminster, London

UNITED KINGDOM

 

Dror Goldenberg

Mellanox Technologies

Sunnyvale, CA

USA

 

Jean Gonnord

CEA - The French Nuclear Agency

Choisel

FRANCE

 

Sergei Gorlatch

Universität Münster

Institut für Informatik

Münster

GERMANY

 

Lucio Grandinetti

Dept. of Electronics, Informatics and Systems

University of Calabria

Rende, CS

ITALY

 

Weiwu Hu

Institute of Computing Technology

Chinese Academy of Sciences

Beijing

CHINA

 

Christopher Huggins

ClusterVision

Amsterdam

THE NETHERLANDS

 

Chris Jesshope

Informatic Institute, Faculty of Science

University of Amsterdam

Amsterdam

THE NETHERLANDS

 

Peter Kacsuk

MTA SZTAKI

Budapest

HUNGARY

 

Carl Kesselman

Information Sciences Institute

University of Southern California

Marina del Rey, Los Angeles, CA

USA

 

Janusz Kowalik

University of Gdansk

POLAND

 

Valeria Krzhizhanovskaya

St. Petersburg State Polytechnic University

RUSSIA

and

University of Amsterdam

THE NETHERLANDS

 

Marcel Kunze

Karlsruhe Institute of Technology

Steinbuch Centre for Computing

Karlsruhe

GERMANY

 

Tim Lanfear

NVIDIA Ltd

Reading

UNITED KINGDOM

 

Simon Lin

Academia Sinica Grid Computing (ASGC)

Institute of Physics

Taipei

TAIWAN

 

Thomas Lippert

Juelich Supercomputing Centre

Juelich

GERMANY

 

Miron Livny

Computer Sciences Dept.

University of Wisconsin

Madison, WI

USA

 

Ignacio Llorente

Dpt. de Arquitectura de Computadores y Automática

Facultad de Informática

Universidad Complutense de Madrid

Madrid

SPAIN

 

Satoshi Matsuoka

Dept. of Mathematical and Computing Sciences

Tokyo Institute of Technology

Tokyo

JAPAN

 

Timothy G. Mattson

Intel Computational Software Laboratory

Hillsboro, OR

USA

 

Paul Messina

Argonne National Laboratory

Argonne, IL

U.S.A.

 

Ken Miura

Center for Grid Research and Development

National Institute of Informatics, Tokyo

JAPAN

 

Leif Nordlund

AMD

SWEDEN

 

Jean-Pierre Panziera

Extreme Computing Division

Bull

FRANCE

 

Christian Perez

INRIA

FRANCE

 

Raoul Ramos Pollan

CCETA-CIEMAT Computing Center

SPAIN

 

B.B. Prahlada Rao

Programme SSDG

C-DAC Knowledge Park

Bangalore

INDIA

 

Ulrich Rüde

Lehrstuhl fuer Simulation

Universitaet Erlangen-Nuernberg

Erlangen

GERMANY

 

Bernhard Schott

Platform Computing

Frankfurt

GERMANY

 

Satoshi Sekiguchi

Information Technology Research Institute

National Institute of Advanced Industrial Science and Technology

JAPAN

 

Alex Shafarenko

Dept. of Computer Science

University of Hertfordshire

Hatfield

UNITED KINGDOM

 

Mark Silberstein

Technion-Israel Institute of Technology

Haifa

ISRAEL

 

Leonel Sousa

INESC

and

TU Lisbon, Lisbon

PORTUGAL

 

Domenico Talia

Dept. of Electronics, Informatics and Systems

University of Calabria

Rende, CS

ITALY

 

Dmitry Tkachev

T-Platforms

Moscow

RUSSIA

 

Amy Wang

Institute for Theoretical Computer Science

Tsinghua University

Beijing

CHINA

 

Robert Wisniewski

IBM  Watson Research Center

Yorktown Heights, NY

USA

 

Matt Wood

Amazon Web Services

Amazon

UNITED KINGDOM

 

Hongsuk Yi

Supercomputing Center

KISTI Korea Institute of Science and Technology Information

Daejeon

KOREA

 

 

Workshop Agenda

 

Monday, June 21st

Session

Time

Speaker/Activity

 

9.00 – 9.15

Welcome Address

Session I

 

State of the art and future scenarios

 

9.15 – 9.50

I. FOSTER

“Thinking outside the box: How cloud, grid, and services can make us smarter”

 

9.50 – 10.25

C. JESSHOPE

“General-purpose parallel computing - a matter of scale”

 

10:25 – 11:00

g. gao

Dataflow Models for Computation. State of the Art and Future Scenarios”

 

11:00 – 11:30

COFFEE BREAK

               

11:30 – 12:05

R. WISNIEWSKI

Software Challenges and Approaches for Extreme-Scale Computing”

 

12:05 – 12:40

S. MATSUOKA

Hetero – Acceleration the Yellow Brick Road onto Exascale?”

 

12:40 – 12:50

CONCLUDING REMARKS

Session II

 

Emerging computer systems and solutions

 

17:00 – 17:25

F. BAETKE

“Standards-based Peta-scale Systems – Trends, Implementations and Solutions”

 

17:25 – 17:50

D. GOLDENBERG

“Driving InfiniBand Technology to Petascale Computing and Beyond”

 

17:50 – 18:15

A. GEIGER

“Status and Challenges of a Dynamic Provisioning Concept for HPC-Services”

 

18:15 – 18:45

COFFEE BREAK

 

18:45 – 19:10

D. TKACHEV

Clustrx: A New Generation Operating System Designed for HPC”

 

19:10 – 19:35

C. HUGGINS

“Managing complex cluster architectures with Bright Cluster Manager”

 

19:35 – 20:00

B. SCHOTT

“DGSI: Federation of Distributed Compute Infrastructures”

 

20:00 – 20:10

CONCLUDING REMARKS

 

 

Tuesday, June 22nd

Session

Time

Speaker/Activity

Session III

 

Advances in HPC technology and systems I

 

9:00 – 9:25

S. DOSANJH

Exascale Computing and the Role of Co-design”

 

9:25 – 9:50

J.P. PANZIERA

“Beyond the Petaflop

 

9:50 – 10:15

V. GETOV

Component-oriented Approaches for Software Development and Execution in the Extreme-scale Computing Era”

 

10:15 – 10:40

S. SEKIGUCHI

“Development of High Performance Computing and the Japanese planning”

 

10:40 – 11:05

T. LIPPERT

“PRACE: Europe's Supercomputing Research Infrastructure”

               

11:05 – 11:35

COFFEE BREAK

 

11:35 – 12:00

T. MATTSON

“The Future of Many Core Processors: a Tale of Two Processors”

 

12:00 – 12:25

L. NORDLUND

“AMD current and future solutions for HPC Workloads”

 

12:25 – 12:50

S. EVRIPIDOU

“The Data-Flow model of Computation in the Multi-core era”

 

12:50 – 13:00

CONCLUDING REMARKS

Session IV

 

Advances in HPC technology and systems II

 

17:00 – 17:25

G. BILARDI

“Network Oblivious Algorithms”

 

17:25 – 17:50

G. BOSILCA

“Distributed Dense Numerical Linear Algebra Algorithms on massively parallel heterogeneous architectures”

 

17:50 – 18:15

P. ANEDDA

“Mixing and matching virtual and physical HPC clusters”

 

18:15 – 18:45

COFFEE BREAK

 

18:45 – 20:00

PANEL DISCUSSION 1: “Challenges and opportunities in exascale computing

Chair: P. Messina

Panelists: S. Dosanjh, J. Gonnord, D. Goldenberg, T. Lippert, J.P. Panziera, R. Wisniewski, S. Sekiguchi, S. Matsuoka

 

 

Wednesday, June 23rd

Session

Time

Speaker/Activity

Session V

 

Grid and cloud technology and systems

 

9:00 – 9:25

M. LIVNY

“Distributed Resource Management: The Problem That Doesn’t Go Away”

 

9:25 – 9:50

D. TALIA

“Service-Oriented Distributed Data Analysis in Grids and Clouds”

 

9:50 – 10:15

P. KACSUK

Integrating Service and Desktop Grids at Middleware and Application Level

 

10:15 – 10:40

J. FORTES

“Cross-cloud Computing”

 

10:40 – 11:05

V. KRZHIZHANOVSKAYA

“Dynamic workload balancing with user-level scheduling for parallel applications on heterogeneous Grid resources”

               

11:05 – 11:35

COFFEE BREAK

Session VI

 

Cloud technology and systems I

 

11:35 – 12:00

C. CATLETT

“Rethinking Privacy and Security: How Clouds and Social Networks Change the Rules”

 

12:00 – 12:25

I. LLORENTE

“Innovations in Cloud Computing Architectures”

 

12:25 – 12:50

M. KUNZE

“The OpenCirrus Project. Towards an Open-source Cloud Stack”

 

12:50 – 13:00

CONCLUDING REMARKS

Session VII

 

Cloud technology and systems II

 

16:30 – 17:00

M. WOOD

“Orchestrating the Cloud: High Performance Elastic Computing”

 

17:00 – 17:25

M. DEVARE

“A Prototype implementation of Desktop Clouds”

 

17:25 – 17:50

M. SILBERSTEIN

“Mechanisms for cost-efficient execution of Bags of Tasks in hybrid cloud-grid environments”

 

17:50 – 18:15

M. DALHEIMER

“Cloud Computing and Enterprise HPC”

 

18:15 – 18:45

COFFEE BREAK

 

18:45 – 20:00

PANEL DISCUSSION 2: “State of the Cloud: Early Lessons Learned With Commercial and Research Cloud Computing”

Chair: C. Catlett

Panelists: I. Foster, I. Llorente, M. Dalheimer, M. Kunze

 

 

Thursday, June 24th

Session

Time

Speaker/Activity

Session VIII

 

Infrastructures, tools, products, solutions for HPC, grids and clouds

 

9:00 – 9:25

A. WANG

“PAIMS: Precision Agriculture Information Monitoring System”

 

9:25 – 9:50

T. MATTSON

“Design patterns and the quest for General Purpose Parallel Programming”

 

9:40 – 10:15

W. HU

“A Multicore Processor Designed for Petaflops Computation”

 

10:15 – 10:40

L. SOUSA

Efficient Execution on Heterogeneous Systems

 

10:40 – 11:05

T. LANFEAR

High-Performance Computing with NVIDIA Tesla GPUs

               

11:05 – 11:35

COFFEE BREAK

 

11:35 – 12:00

J. KOWALIK

“Hybrid Computing for Solving High Performance Computing Problems”

 

12:00 – 12:50

P. ARLUKOWICZ

“An Introduction to CUDA Programming: A Tutorial”

 

12:50 – 13:00

CONCLUDING REMARKS

Session IX

 

National and international HPC, grid and cloud infrastructures and projects

 

16:30 – 16:55

K. MIURA

“Cyber Science Infrastructure in Japan - NAREGI Grid Middleware Version 1 and Beyond -”

 

16:55 – 17:20

R. RAMOS POLLAN

“The road to sustainable eInfrastructures in Latin America”

 

17:20 – 17:40

B. BECKER

“The South African National Grid: Blueprint for Sub-Saharan e-Infrastructure”

 

17:40 – 18:00

B.B. PRAHLADA RAO

“GARUDA: Indian National Grid Computing Initiative”

 

18:00 – 18:30

COFFEE BREAK

 

18:30 – 18:50

S. LIN

“Building e-Science and HPC Collaboration in Asia”

 

18:50 – 19:10

M. BUBAK

“PL-Grid: the first functioning National Grid Initiative in Europe”

 

19:10 – 19:35

W. GENTZSCH

“DEISA and the European HPC Ecosystem”

 

19:35 – 20:00

H. YI

“HPC Infrastructure and Activity in Korea”

 

20:00 – 20:10

CONCLUDING REMARKS

 

 

Friday, June 25th

Session

Time

Speaker/Activity

Session X

 

Challenging applications of HPC, grids and clouds

 

9:00 – 9:25

C. KESSELMAN

“The Grid as Infrastructure for Sharing BioMedical Information: The Biomedical Informatics Research Network”

 

9:25 – 9:50

T. DAVID

“System Level Acceleration for Multi-Scale Modelling in Physiological Systems”

 

9:50 – 10:15

M. ATHANASOULIS

Building shared HPC facilities: the Harvard Orchestra experience

 

10:15 – 10:40

S. GORLATCH

Towards Scalable Online Interactive Applications on Grids and Clouds

 

10:40 – 11:05

U. RUEDE

Simulation and Animation of Complex Flows Using 294912 Processor Cores

 

11:05 – 11:35

COFFEE BREAK

 

11:35 – 12:00

A. SHAFARENKO

“Asynchronous computing of irregular applications using the SVPN model and S-Net coordination”

 

12:00 – 12:25

M. BUBAK

“Towards Collaborative Workbench for Science 2.0 Applications”

 

12:25 – 12:50

C. PEREZ

“On High Performance Software Component Models”

 

12:50 – 13:00

CONCLUDING REMARKS

 

CHAIRMEN

 

Sudip S. Dosanjh

SANDIA National Labs

Albuquerque, NM

USA

 

 

Wolfgang Gentzsch

DEISA Distributed European Infrastructure

for Supercomputing Applications

and

OGF

GERMANY

 

 

Satoshi Matsuoka

Dept. of Mathematical and Computing Sciences

Tokyo Institute of Technology

Tokyo

JAPAN

 

 

 

Chris Jesshope

Informatic Institute, Faculty of Science

University of Amsterdam

Amsterdam

THE NETHERLANDS

 

 

Ian Foster

Argonne National Laboratory

and

Dept. of Computer Science

The University of Chicago

Argonne & Chicago, IL

U.S.A.

 

 

Ian Foster

Argonne National Laboratory

and

Dept. of Computer Science

The University of Chicago

Argonne & Chicago, IL

U.S.A.

 

 

 

Carl Kesselman

Information Sciences Institute

University of Southern California

Marina del Rey, Los Angeles, CA

USA

 

 

Guang Gao

University of Delaware

Department of Electrical and Computer Engineering

Newark, Delaware

USA

 

 

Miron Livny

Computer Sciences Dept.

University of Wisconsin

Madison, WI

USA

 

 

Gerhard Joubert

Technical University Clausthal

GERMANY

 

PANELS

PANEL 1

Challenges and opportunities in exascale computing

 

Numerous workshops have identified scientific and engineering computational grand challenges that could be addressed with exascale computing resources.  However, the technology expected to be available to build affordable exascale systems in the next decade leads to architectures that will be very difficult to program and to manage. Will completely new programming models be needed?  And new numerical algorithms and mathematical models? Will a co-design approach that involves application teams from the beginning of the exascale initiative make the programming challenges tractable?  The panel participants will debate these issues and other related topics.

 

Chairman: P. Messina

Panelists: S. Dosanjh, J. Gonnord, D. Goldenberg, T. Lippert, J.P. Panziera, R. Wisniewski, S. Sekiguchi, S. Matsuoka

Back to Session IV

PANEL 2

State of the Cloud: Early Lessons Learned With Commercial and Research Cloud Computing

 

This panel will discuss insights gained using cloud technologies and services for scientific computing.  These range from security to performance, from costs to flexibility.  Each panelist will briefly discuss one or more of these challenges, offering examples of solutions as well as difficulties related to scientific use of clouds.

 

Chairman: C. Catlett

Panelists: I. Foster, I. Llorente, M. Dalheimer, M. Kunze

 

Back to Session VII

 

 

ABSTRACTS

Thinking outside the box: How cloud, grid, and services can make us smarter

 

Ian Foster

Math & Computer Science Div., Argonne National Laboratory

Argonne, IL and

Dept of Computer Science, The University of Chicago

Chicago, IL, U.S.A.

 

Whitehead observed that "civilization advances by extending the number of important operations which we can perform without thinking about them." Thanks to Moore's Law, these operations can nowadays involve increasingly complex information manipulation and computation. The outsourcing of computing via approaches such as utility computing, on-demand computing, grid computing, software as a service, and cloud computing can further enhance human capabilities, by freeing computer applications from the limiting confines of a single computer. Software that thus runs "outside the box" can be more powerful (Google, TeraGrid), dynamic (Animoto, caBIG), and collaborative (FaceBook, myExperiment). It can also be cheaper, due to economies of scale in hardware and software. Simultaneously, service-oriented architectures make it easier to integrate data and software from many sources. The combination of new functionality and new economics inspires new applications, reduces barriers to entry for application providers, and in general disrupts the computing ecosystem. I discuss new applications that outside-the-box computing enables; the hardware and software architectures that make these new applications possible; and the social dimensions of outside-the-box computing.

 

Back to Session I

General-purpose parallel computing - a matter of scale

 

Chris Jesshope

 

Faculty of Science, Informatics Institute

University of Amsterdam

Amsterdam, NETHERLANDS

 

The question this talk will pose is whether it possible to achieve the holy grail of general purpose parallel computing. One of the major pitfalls to this goal is the many and varied approaches to parallel programming, yet we believe it is possible to provide a generic virtualisation layer that provides the necessary API to support this variety of concerns. Another question is whether this interface can be implemented efficiently across a range of architectures, where by efficiency we mean not only meeting non-functional constraints on throughput and latency but also managing constraints on energy dissipated and heat distribution in the target devices, which is becoming increasingly important. To meet these constraints it is likely that the target processors will be highly heterogeneous and that the run-time system will need to support both data-driven scheduling to manage the asynchrony that comes with this territory as well as to provide dynamic resource management to allow the overall system to adapt and meet the these potentially conflicting requirements. We will present the Self-adaptive virtual processor and describe work on various implementations including in the ISA of a multi-core, as an interface to FPGA programming and across a variety of existing conventional and not-so conventional multi-core platforms.

Back to Session I

Dataflow Models for Computation. State of the Art and Future Scenarios

 

Guang Gao

University of Delaware, Department of Electrical and Computer Engineering

Newark, Delaware, USA

 

The emerging trend on multi-core chips is changing the technology landscape of computing system in the scale that has not been witnessed since the Intel microprocessor chip commissioned in early 1970s. However, the implication of this technology revolution is profound: its success can only be ensured if we can successfully (productively) implement parallel computer architecture on such chips as well as its associated software technology.

We start with a brief note on the fundamental work on dataflow models of computation in the last century that goes back to 1960s/ 1970s.  We then comment on the state of the art development of dataflow models to address the new challenges in parallel architecture and software models presented by the multi-core chip technology.  Finally, we present some hypotheses on the future scenarios of advances of dataflow models.

Back to Session I

Software Challenges and Approaches for Extreme-Scale Computing

 

Robert Wisniewski

IBM  Watson Research Center

Yorktown Heights, NY, USA

 

The drive to exascale contains a series of challenges for technology.  The solutions that will be developed from a technology perspective are going to lead to a related series of challenges from a system software perspective.

Some of the prime determiners of the software challenges will include the technology solutions to meet the power budget and to achieve the requisite reliability.  Also, trends in memory and I/O costs, and their relative ratios to compute are changing unfavorably.  When investigations began a couple years ago into software for exascale, there was a feeling revolutionary approaches would be needed in many spaces.  As the challenges were examined in greater detail, there is a growing sense that both because of time constraints, and achievable evolutionary technology, that while there are some areas that will require significantly new approaches to achieve exascale, other areas can support exascale in an evolutionary manner.  In the this talk I will lay out the major technology challenges with likely their solutions, and how that will impact the system software for exascale.  I will then describe some of the key approaches IBM is taking to address those impacts on system software.

Back to Session I

Hetero – Acceleration the Yellow Brick Road onto Exascale?

 

Satoshi Matsuoka

 

Dept. of Mathematical and Computing Sciences

Tokyo Institute of Technology, Tokyo, JAPAN

 

Since the first commodity x86 cluster Wigraf achieving paltry 10s~100s Megaflops in 1994,  we  have experienced several orders of magnitude boost in performance. However, the first Petaflop was achieved with the LANL RoadRunner, a Cell-based "accelerated" cluster, and in 2010 we may see the first (GP)GPU-based cluster reaching Petaflops. Do such non-CPU "accelerator” merely push the flops superficially, or are they fundamental to scaling? Based on experiences from TSUBAME, the first  GPU-accelerated cluster on the Top500, we show that GPUs not only achieve higher performance but also better scaling, and in fact their true nature as multithreaded massively-parallel vector processor would be fundamental for Exascale. Such results are being reflected onto the design of TSUBAME2.0 and its successors.

Back to Session I

Standards-based Peta-scale Systems – Trends, Implementations and Solutions

 

Frank Baetke

Global HPC Technology, Hewlett Packard, Richardson, Texas, USA

 

HP’s HPC product portfolio which has always been based on standards at the processor, node and interconnect level lead to a successful penetration of the High Performance Computing market across all application segments. Specifically the c-Class Blade architecture is now fully established as a reference for HPC-Systems as the TOP500 list clearly shows. The rich portfolio of compute, storage and workstation blades comprises a family a components call the Proliant BL-series complementing the well-established rack-based Proliant DL family of nodes. To address additional challenges at the node and systems level HP recently introduced the Proliant SL-series.

Beyond acquisition cost, the other major factor is power and cooling efficiency.  This is primarily an issue of cost for power, but also for the power and thermal density of what can be managed in a data center. To leverage the economics of scale established HPC centers as well as providers of innovative services are evaluating new concepts which have the potential to make classical data center designs obsolete. Those new concepts provide significant advantages in terms of energy efficiency, deployment flexibility and manageability. Examples of this new approach, often dubbed POD for Performance Optimized Datacenter, including a concept to scale to multiple PFLOPS at highest energy efficiency will be shown.

Finally details of a new Peta-scale system to be delivered later this year will be shown and discussed.

Back to Session II

Driving InfiniBand Technology to Petascale Computing and Beyond

 

Dror Goldenberg

Mellanox Technologies, Sunnyvale, CA, USA

 

PetaScale and Exascale systems will span tens-of-thousands of nodes, all connected together via high-speed connectivity solutions. With the growing size of clusters and CPU cores per cluster node, the interconnect needs to provide all of the following features: highest bandwidth, lowest latency, multi-core linear scaling, flexible communications capabilities, autonomic handling of data traffics, high reliability and advanced offload capabilities. InfiniBand has emerged to be the native choice for PetaScale clusters, and was chosen to be the connectivity solution for the first Petaflop system, and is being used for 60% of the world Top100 supercomputers (according to the TOP500 list). With the capabilities of QDR (40Gb/s) InfiniBand, including adaptive routing, congestion control, RDMA and quality of service, InfiniBand shows a strong roadmap towards ExaScale computing and beyond. The presentation will cover the latest InfiniBand technology, advanced offloading capabilities, and the plans for InfiniBand EDR solutions.

Back to Session II

Status and Challenges of a Dynamic Provisioning Concept for HPC-Services

 

Alfred Geiger

T-Systems Solutions for Research GmbH

Stuttgart, GERMANY

 

The presentation will first of all describe the state of the art in commercial HPC-provisioning. Existing concepts are primarily based on shared services that can be accessed via grid-middleware. However currently we observe a mismatch between these provisioning concepts and the expectations of HPC-customers. The request is for a cloud-like model for the provisioning of temporarily dedicated resources. Treating cloud as a business model rather than a technology, the service-providers are doing first steps in this direction. On the other side there are still significant obstacles on the way to a service that fully meets the customer-requirements. In this contribution a possible roadmap on the way to dynamic HPC-services will be discussed together with short-term workarounds for missing pieces of technology. Furthermore the technology-gaps will be identified.

Back to Session II

Clustrx: A New Generation Operating System Designed for HPC

 

Dmitry Tkachev

Research and Development Director T-Platforms

Moscow, RUSSIA

 

Clustrx is a new generation HPC OS architected specifically for high performance computing.  Designed by a team of HPC experts, Clustrx is the first OS in which all the operating system functionality - the HPC stack and workload management subsystem - are fully integrated into a single software package.  Designed with an innovative, real-time management and monitoring system, Clustrx eliminates any limits of scalability and manageability for multi-petaflops clusters and simplifies the shared use of supercomputing resources for grid environments.  Clustrx is the HPC operating system designed to enable the eventual migration from petascale to exascale.

Back to Session II

Managing complex cluster architectures with Bright Cluster Manager

 

Christopher Huggins

ClusterVision, Amsterdam, THE NETHERLANDS

 

Bright Cluster Manager makes clusters of any size easy to install, use and manage, and is the cluster management solution of choice for many universities, research institutes and companies across the world. In this presentation, ClusterVision will give some examples on how Bright Cluster Manager makes it easy to install, use, monitor, manage and scale large and complex cluster infrastructures.

 

Back to Session II

DGSI: Federation of Distributed Compute Infrastructures

 

Bernhard Schott

Platform Computing GERMANY

 

DGSI (D-Grid Scheduler Interoperability project) develops the DCI-Federation Protocol enabling dynamic combination and transparent use of Cloud and Grid resources. The protocol is open source and technology agnostic and will be implemented in 5 different Grid technologies in the project. Presented implementation examples are based on Platform Computing Cloud and Grid technology: Platform ISF and Platform LSF.

Back to Session II

Exascale Computing and the Role of Co-design

 

Sudip Dosanjh

SANDIA National Labs

Albuquerque, NM, USA

 

Achieving a thousand-fold increase in supercomputing technology to reach exascale  computing (1018 operations per second) in this decade will revolutionize the way supercomputers are used. Predictive computer simulations will play a critical role in achieving energy security, developing climate change mitigation strategies, lowering CO2 emissions and ensuring a safe and reliable 21st century nuclear stockpile. Scientific discovery, national competitiveness, homeland security and quality of life issues will also greatly benefit from the next leap in supercomputing technology. This dramatic increase in computing power will be driven by a rapid escalation in the parallelism incorporated in microprocessors. The transition from massively parallel architectures to hierarchical systems (hundreds of processor cores per CPU chip) will be as profound and challenging as the change from vector architectures to massively parallel computers that occurred in the early 1990’s. Through a collaborative effort between laboratories and key university and industrial partners, the architectural bottlenecks that limit supercomputer scalability and performance can be overcome. In addition, such an effort will help make petascale computing pervasive by lowering the costs for these systems and dramatically improving their power efficiency.

 

The U.S. Department of Energy’s strategy for reaching exascale includes:

   Collaborations with the computer industry to identify gaps

   Prioritizing research based on return on investment and risk assessment

   Leveraging existing industry and government investments and extending technology in strategic technology focus areas

   Building sustainable infrastructure with broad market support

      Extending beyond natural evolution of commodity hardware to create new markets

      Creating system building blocks that offer superior price/performance/programmability at all scales (exascale, departmental, and embedded)

   Co-designing hardware, system software and applications

 

The last element, co-design, is a particularly important area of emphasis. Applications and system software will need to change as architectures evolve during the next decade. At the same time, there is an unprecedented opportunity for the applications and algorithms community to influence future computer architectures. A new co-design methodology is needed to make sure that exascale applications will work effectively on exascale supercomputers.

 

Back to Session III

Beyond the Petaflop

 

Jeanne-Pierre Panziera

Extreme Computing Division, Bull, France

 

As Petaflop-size systems are currently being deployed, a formidable

challenge has been set for the HPC community: one exaflop within 8-10 years.

Relying on technology evolution alone is not enough to reach this goal. A

disruptive approach that encompasses all aspects of hardware, software and

application design is required.

Back to Session III

Component-oriented Approaches for Software Development and Execution in the Extreme-scale Computing Era

 

Vladimir Getov

School of Electronics and Computer Science

University of Westminster, London, U.K.

 

The complexity of computing at extreme scales is increasing rapidly, now matching the complexity of the simulations running on them. This complexity arises from the interplay of variety of factors such as level of parallelism (systems in this range currently use hundreds of thousands of processing elements and are envisioned to reach millions of threads of parallelism), availability of parallelism in algorithms, productivity, design of novel runtime system software, deep memory hierarchies, heterogeneity, reliability and resilience, and power consumption, just to name a few. The quest for higher processing speed has become only one of many challenges when designing novel high-end computers. While this complexity is qualitatively harder and multidimensional, addressing successfully the unprecedented conundrum of challenges in both hardware and software is a key to rapidly unlocking the potential of extreme-scale computing within the next 10-15 years.

 

In recent years, component-based technologies have emerged as a modern and promising approach to software development of complex parallel and distributed applications. The adoption of software components could increase significantly the development productivity, but the lack of longer-term experience and the increasing complexity of the target systems demand more research results in the field. In particular, the search for the most appropriate component model and corresponding programming environments is of high interest and importance. The higher level of complexity as described above involves a wider range of requirements and resources which demand dedicated support for dynamic intelligent (non-functional) properties and flexibility that could be provided in an elegant way by adopting a component-based approach.

 

Figure 1: Component-based Development and Execution Platform Architecture

This paper will present the design methodology of a generic component-based platform for both applications and system frameworks to have a single, seamless, “invisible” system image. The block diagram in Figure 1 depicts the generic architecture of our component-based development and execution platform.

 

We will argue that the software development could be simplified by adopting the component-oriented paradigm, where much better productivity can be achieved because of the higher level of abstraction. At the same time this approach enables in a natural way the introduction of autonomic support at runtime including automatic reconfiguration and tuning. This is illustrated by the model-to-solution pipeline block diagram presented in Figure 2. The main functions – compose, deploy, monitor and steer – are being implemented in our component-based platform.

 

Figure 2: Component-centric Problem-to-Solution Pipeline

 

The full paper will include more details about our initial experience that also cover some other important aspects of the development and execution cycle such as validation and dynamic verification. The conclusions include also some ideas and plans for future research in this area.

Back to Session III

Development of High Performance Computing and the Japanese planning

 

Satoshi Sekiguchi

Information Technology Research Institute

National Institute of Advanced Industrial Science and Technology, JAPAN

 

At the well-known SHIWAKE in November 2009, the Government Revitalisation Unit gave sentence to freeze on The Next-Generation Supercomputer Project, however it has been survived under the conditions of engaging more people to enjoy the benefit of the extreme performance and building national scale infrastructure to support science, engineering and other businesses. One of the effort just started to make this happen is to form so-called "HPCI" which intends to provide a venue to gather computing resources and people around. As a member of planning HPCI working group, I will introduce an outline of the discussion and its plan for the future.

Back to Session III

PRACE: Europe's Supercomputing Research Infrastructure

 

Thomas Lippert

Juelich Supercomputing Centre, Juelich, GERMANY

 

Within the last two years a consortium of 20 European countries has prepared the legal and technical prerequisites for the establishment of a leadership-class supercomputing infrastructure in Europe. The consortium named "Partnership for Advanced Computing in Europe" has carried out a preparatory phase project supported by the European Commission. The statutes of the new association, a Belgian "association sans but lucrative", were signed in April 2010 and its inauguration took place in June 2010. So far, four members have committed to provide compute cycles worth € 100 Million each in the 5 years period until 2015. The hosting countries in succession foresee the installation of machines of the highest performance class (Tier-0), providing a diversity of architectures beyond Petaflop/s.

Access to the infrastructure is provided on the basis of scientific quality through a pan-European peer review system under the guidance of the scientific steering committee (SSC) of PRACE. The SSC is a group of leading European peers from a variety of fields in computational science and engineering. Proposals can be submitted in form of projects or as programs by communities. In May 2010 a first early-access call was issued, and the provision of computer time through PRACE is foreseen commencing in August 2010 by the supercomputer JUGENE of Research Centre Jülich. Regular provision will start in November 2010.

PRACE's Tier-0 supercomputing infrastructure will be complemented by national centres (Tier-1) of the PRACE partners. In the tradition of DEISA, the Tier-1 centres will provide limited access to national systems for European groups - granted through national peer review - under the synchronization and governance of PRACE.

Back to Session III

The Future of Many Core Processors: a Tale of Two Processors

 

Tim Mattson

Principal Engineer Intel Labs

Hillsboro, OR, USA

 

We all know by now that many core chips are the future of the microprocessor industry.  They are the only way to deliver a steady progression of performance improvements in response to Moore’s law while staying within a fixed power envelope. But how will those cores be connected?  How simple will the cores be?  What programming models will be most effective?  Can we develop a memory model that will “feed the beast” and let us sustain high throughput?

 

These and other questions are the aim of Intel Lab’s Terascale research program. As part of this program we envision a sequence of research chips.  2007’s 80-core terascale processor is the first of these chips. Recently, we announced our next research chip, the 48 core SCC processor.

In this talk, I will discuss these two research processors and what they tell us about the future of many core processors. I will also discuss the critical problem of how to program many core chips. I am a software person so I will emphasize the applications programmers’ point of view; though I will attempt to address lower level hardware and software issues as well.

 

Back to Session III

AMD current and future solutions for HPC Workloads

 

Leif Nordlund

AMD, SWEDEN

 

Reflections on the Multicore CPU and the GPGPU architecture developments, Possible ways to get closer to Heterogeneous computing and Open standards vs Vendor specific implementations.

Back to Session III

The Data-Flow model of Computation in the Multi-core era

 

Skevos (Paraskevas) Evripidou

Department of Computer Science, University of Cyprus

Nicosia, CYPRUS

 

The sequential model of execution has dominated computing since the inception of digital computers. Research and development on parallel processing was also dominated by the use of sequential processors with some hardware and/or software extensions. Proponents of alternative models that are naturally parallel, such as Data-flow, have been citing the shortcomings and limitations of the sequential model versus the inherit advantages of the alternative models. However, the microprocessor designers were able to use the extra transistor given to them by Moore’s law to continue improving the performance by building more complex chips with very large caches. At the beginning of this decade the most severe limitation of the sequential model, namely its inability to tolerate long memory latencies (Memory Wall) has slowed down the performance gains and finally stopped them. Furthermore, the very the complex designs in combination with the very large caches have given a rise to the power consumption (Power wall). These two walls force the industry to change direction and switch to multiple cores per chip which in turn has elevated concurrency as the major challenge in achieving high performance. The switch to multi-core was basically an engineering remedy: lowering the frequency at a level that the memory wall could be overcome and replicate the basic sequential core.  New concurrent models/paradigms are needed in order to fully utilize the potential of Multi-core chips. 

 

The Dynamic Data-flow model is a formal model that can handle concurrency efficiently in a distributed manner and it can tolerate memory and synchronization latencies.  In this work we make the case that the Dynamic Data-flow model of execution can be efficiently combined with the control flow model in order to provide efficient concurrency control and at the same time fix the limitations of the control flow that cause the switch to multi-core. The combined model will benefit from the tolerance to the memory and synchronization latencies of the data-flow model, thus bypassing the Memory wall.  The combined Micro-architecture will be simpler and more power efficient because they will be no need for the complex modules that were added to overcome the memory latencies in the control-flow model. Furthermore, the data-flow scheduling will significantly, reduce the size of the caches, thus overall reducing the effect of the Power wall.

Back to Session III

Network Oblivious Algorithms

 

Gianfranco Bilardi

University of Padova, Italy

 

The design of algorithms that can run unchanged yet efficiently on a variety of machines characterized by different degrees of parallelism and communication capabilities is a highly desirable goal. We propose a framework for network-obliviousness based on a model of computation where the only parameter is the problem's input size. Algorithms are then evaluated on a model with two parameters, capturing parallelism and synchronization requirements of communication.

 

We show that optimality in the evaluation model implies near optimality in the Decomposable BSP model (D-BSP), which deploys logarithmically many parameters in the number of processors to effectively describe a wide and significant class of parallel platforms.  For a special class of "wise" network-oblivious algorithms one can actually establish D-BSP optimality.

 

We illustrate our framework by providing optimal wise network-oblivious algorithms for a few key problems, including martix multiplication and discrete Fouries transform.  We also show that some other key problems, such a broadcast and prefix computation, do not admit optimal network-oblivious algorithms and characterize exactly how close to optimal they can come.

 

(This is joint work with G. Pucci, A. Pietracaprina, and M. Scquizzato,

and F. Silvestri)

Back to Session IV

Distributed Dense Numerical Linear Algebra Algorithms on massively parallel heterogeneous architectures

 

George Bosilca

Innovative Computing Lab, University of Tennessee

Knoxville, USA

 

In this talk we overlook the drastic changes at the architectural level over the last few years, and their impact on the achieved performance for parallel applications. Several approaches to alleviate this problem have been proposed; unfortunately most of them are limited to share memory environments. New generation of algorithms and software are needed to bridge the gap between peak and sustained performance on massively parallel heterogeneous architectures. Based on a new generic distributed Direct Acyclic Graph engine for high performance computing (DAGuE), the DPLASMA project is capable of taking advantage of multicores and accelerators in a distributed environment. Through three common dense linear algebra algorithms, namely: Cholesky, LU and QR factorizations, I will demonstrate from our preliminary results that our DAG-based approach has the potential to temper this characteristic problem in the state-of-the-art distributed numerical software on current and emerging architectures.

Back to Session IV

Mixing and matching virtual and physical HPC clusters

 

Gianluigi Zanetti

CRS4 Center for Advanced Studies, Research and Development in Sardinia

Cagliari, ITALY

 

High Performance Computational Clusters are, in general, rather rigid objects that present to their user a limited number of degrees of freedom related, usually, only to the specification of the resources requested and to the selection of specific applications and libraries.

While in standard production environments this is reasonable and actually desirable, it can become an hindrance when one needs a dynamic and flexible computational environment, for instance for experiments and evaluation, where very different computational approaches, e.g., map-reduce, standard parallel jobs and virtual HPC clusters need to coexist on the same physical facility. We describe a new approach in the management of small to medium, general-purposes clusters based on a flexible High Performance Computing (HPC) software platform capable of partition physical clusters, allocate computing nodes and create and deploy virtual HPC clusters.

Back to Session IV

Distributed Resource Management: The Problem That Doesn’t Go Away

 

Miron Livny

Computer Sciences Dept., University of Wisconsin, USA

 

After more than 30 years of distributed computing research and more than a decade of grid computing, operators and users are still struggling with how to allocate (some refer to it as provision) resources in a distributed environment.  Community owned overly job managers (some refer to them as glide-ins or pilot jobs) are growing in popularity as a partial solution to this problem. We will report on our recent work on using such overlays to manage the allocation of CPUs and will discuss our plans to add support for allocation of storage resources.

Back to Session V

Service-Oriented Distributed Data Analysis in Grids and Clouds

 

Domenico Talia

Dept. of Electronics, Informatics and Systems

University of Calabria, Rende, Italy

 

Today a major challenge in data mining and knowledge discovery is the analysis

of distributed data, information and knowledge sources. New data analysis and knowledge discovery systems should be able to mine distributed and highly heterogeneous data found on Internet, Grids and Clouds. The Service-Oriented paradigm provides a viable approach to implement distributed knowledge discovery systems and applications that run on dispersed computers and analyse distributed data sets.

This talk introduces a general framework for service-oriented knowledge discovery and presents some experiments developed in the new service-oriented version of the Knowledge Grid framework.

Back to Session V

Integrating service and desktop grids at middleware and application level

 

Peter Kacsuk

MTA SZTAKI, Budapest, HUNGARY

 

Current Grid systems can be divided into two main categories: service grids (SG) and desktop grids (DG). Service grids are typically organized from managed clusters and provide a 24/7 service for a large number of users who can submit their applications into the grid. The service grid middleware is quite complex and hence relatively few managed clusters take the responsibility of providing grid services. As a result the number of processors in SGs is moderate typically in the range of 1.000-50.000. Even the largest SG system, EGEE has collected less than 200.000 computers.

Desktop grids are collecting large number of volunteer desktop machines to exploit their spare cycles. These desktops have no SLA requirement, their client middleware code is extremely simple and hence typical number of volunteer desktops in desktop grids is in the range of 10.000-1.000.000. However, their drawback is that they can execute only some very limited number of pre-registered applications, typically compute-intensive bag-of-task applications. The most well-known volunteer desktop grid is SETI@home that collected over 2 Million CPUs.

Comparing the price/performance ratio of SGs and DGs the creation and maintenance of DGs is much cheaper than the one of SGs. Therefore it would be most economical if the compute-intensive bag-of-task applications could be transferred from the expensive SG systems into the cheap DG systems and executed there. On the other hand when an SG system is underloaded its resources could execute WUs coming from a DG system and in this way existing SG systems could support the solution of grand-challenge scientific applications.

The recognition of these mutual advantages of integrating SGs and DGs led to the initiation of the EDGeS (Enabling Desktop Grids for e-Science) EU project that was launched in January 2008 with the objective of integrating these two kinds of grid systems into a joint infrastructure in order to merge their advantages into one system. The EDGeS project integrated gLite-based service grids with BOINC and XtremWeb DG systems.

To make these systems interoperate there are two main options. At the level of the middleware EDGeS has created the 3G Bridge (Generic Grid-Grid Bridge) solution that enables the interconnection of any service and desktop grids. This bridge was used in EDGeS to create the gLite<->BOINC and gLite->XtremWeb bridges. The concept of 3G Bridge is so generic that it was successfully applied in the EELA-2 project, too in order to interconnect the OurGrid P2P desktop grid with gLite service grids. EDGeS has also created a BES plugin for the 3G Bridge based on which desktop grids can submit workunits to any service grid (e.g. Unicore and ARC) that deploy the BES service interface. The newly launched EDGI  (European Desktop Grid Initiative) project will also fully integrate Unicore and ARC service grids with BOINC and XtremWeb DGs based on 3G Bridge.

The second option for grid users to exploit both SGs and DGs can be achieved at the application level. Particularly, in the case of complex workflow applications when bag-of-task jobs, data-intensive jobs, MPI jobs and others are used in a mixed way, a high-level grid portal can help the users to distribute the jobs of the workflow to the most appropriate SG or DG systems. This has also been solved in EDGeS in the framework of P-GRADE portal.

All these experiences will be explained in detail in the talk. At the end, some future plans of the EDGI project will be shown how to support QoS requirements even in the DG part of the integrated SG-DG infrastructure by supporting DG systems with some dedicated local academic clouds.

Back to Session V

Cross-cloud Computing

 

José Fortes

Advanced Computing and Information Systems (ACIS) Lab

and  NSF Center for Autonomic Computing (CAC)

University of Florida, Gainesville, FL, USA

 

This talk will make the case for, discuss requirements and challenges of, and introduce possible technologies and applications for cross-cloud computing.

 

Back to Session V

Dynamic workload balancing with user-level scheduling for parallel applications on heterogeneous Grid resources

 

Valeria Krzhizhanovskaya

St. Petersburg State Polytechnic University, RUSSIA

and University of Amsterdam, THE NETHERLANDS

 

We present a hybrid resource management environment, operating on both application and system levels, developed for minimizing the execution time of parallel applications on heterogeneous Grid resources. The system is based on the Adaptive WorkLoad Balancing algorithm (AWLB) incorporated into the DIANE User-Level Scheduling (ULS) environment. The AWLB ensures optimal workload distribution based on the discovered application requirements and measured resource parameters. The ULS maintains the user-level resource pool, enables resource selection and controls the execution. We present the results of performance comparison of default self-scheduling used in DIANE with AWLB-based scheduling, evaluate dynamic resource pool and resource selection mechanisms, and examine dependencies of application performance on aggregate characteristics of selected resources and application profile.

Back to Session V

Rethinking Privacy and Security: How Clouds and Social Networks Change the Rules

 

Charlie Catlett

Argonne National Laboratory and University of Chicago

 

Many of today's concepts, architectures, and policies related to privacy and security are based on the traditional information infrastructure in which data and assets are created, controlled, and contained within a protected perimeter.  Concepts such as "control" and "containment" are quite different in today's context of interconnected social networks, powerful multi-sensor mobile phones, ubiquitous wireless broadband, and networks of cloud services.  This presentation will discuss examples of new capabilities exposing flaws in traditional privacy and security assumptions, suggesting areas of computer science research and development that are needed to address these new challenges.

Back to Session VI

Innovations in Cloud Computing Architectures

 

Ignacio Llorente

Dpt. de Arquitectura de Computadores y Automática, Facultad de Informática

Universidad Complutense de Madrid

Madrid, SPAIN

 

The aim of the presentation is to describe the innovations in cloud management brought by the OpenNebula Cloud Toolkit. This widely used open-source Cloud manager fits into existing data centers to build private, public and hybrid  Infrastructure-as-a-Service (IaaS) Clouds. Most of its innovative features have been developed to address requirements from business use cases in RESERVOIR, flagship of European research initiatives in virtualized infrastructures and cloud computing. The innovations comprise support for elastic multi-tier services;  flexible and scalable back-end for virtualization, storage and networking management; and support for Cloud federation and interoperability. The presentation ends with an introduction of the community and ecosystem that are evolving around OpenNebula and the new European project on cloud computing infrastructures that are using this innovative cloud technology.

 

Back to Session VI

The OpenCirrus Project. Towards an Open-source Cloud Stack

 

Marcel Kunze

Karlsruhe Institute of Technology, Steinbuch Centre for Computing

Karlsruhe, GERMANY

 

OpenCirrus is a collaboration of industrial and academic organizations: HP, Intel, Yahoo!, CMU, ETRI, IDA Singapore, KIT, MIMOS, RAS and UIUC. OpenCirrus is an open cloud-computing research testbed designed to support research into the design, provisioning, and management of services at a global, multi-datacenter scale. The open nature of the testbed aims to encourage research into all aspects of service and datacenter management. In addition, the hope is to foster a collaborative community around the testbed, providing ways to share tools, lessons and best practices, and ways to benchmark and compare alternative approaches to service management at datacenter scale. The specific research interests of KIT are in the field of HPC as a Service (HPCaaS) and Big Data as a Service (BDaaS).

 

Back to Session VI

Orchestrating the Cloud: High Performance Elastic Computing

 

Matt Wood

Amazon Web Services, Amazon, UK

 

Constraints are everywhere when working with high throughput tools, be they data mining, indexing, machine learning, financial analysis or complex scientific simulations. As scale increases so network, disk I/O, CPU performance and utilization all become larger and larger barriers to to actually getting work done.

 

This talk introduces the use of elastic, scalable cloud approaches as a set of productivity tools for constrained, highly flexible domains. We’ll explore how Cloud Computing can play a central part in the orchestration, management and monitoring of scalable computation at petabyte scale. Over the course of the session, we’ll aim to cover:

 

+ The productivity tax of web scale applications

+ Hard constraints of big data: maintaining availability at scale

+ Architectures and models for high throughput systems

+ Addressing data as a programmable resource

+ Orchestrating cloud architectures

+ Managing data sets and workflows

+ Map/reduce for elastic infrastructures.

Back to Session VII

A Prototype implementation of Desktop Clouds

 

Manoj Devare, Mehdi Sheikhalishai, Lucio Grandinetti

Department of Electronics, Informatics and Systems, University of Calabria, Cosenza, ITALY

 

The cloud computing is a popular paradigm, for serving the software, platform and infrastructure as a service to the customer. It has been observed, seldom the highest capacity of the personal computers (PC) are utilized. This work provides the virtual infrastructure manager and scheduling framework to leverage the idle resource(s) of PCs. VirtualBox hypervisor is used as the best suited virtualization technology for this work.

In this talk we will discuss this novel architecture and the scheduling approach to launch a computation abstracted as a virtual machine or a virtual cluster using full virtualization approach. The scheduling framework balances both requirements of resource provider (PC owners) i.e. the permission of the PC owner to be taken into account and user of the cloud system who expects the best performance during the whole session.

One can submit lease requirement to the scheduler e.g. for running HPC applications. It is a bit tricky to work in such a non-dedicated heterogeneous environment, for yielding the power of the idle resources of Computers.

Back to Session VII

Mechanisms for cost-efficient execution of Bags of Tasks in hybrid cloud-grid environments

 

Mark Silberstein

Technion-Israel Institute of Technology

Haifa, ISRAEL

 

Pay-for-use execution environments (clouds) allow for substantial cost savings when coping with unexpected surges in computing demand, eliminating the need to over-provision the required capacity.   Their cost effectiveness for scientific computing, however,  appears to be far lower, in particular if one already has access to established grids and clusters. To enable the grid/cloud hybrids, the current state-of-the-art in the middleware systems (Condor/Sun Grid Engine) has been to allow demand-driven extension of the local resources into the cloud when the former are insufficient to accommodate the load.

In this talk we suggest an alternative,  more cost-effective way to use grids and clouds together, which appears to be exactly the opposite of the common "rent-when-insuffcient-resources" approach. The core observation is that the cloud resources exhibit lower failure rate as opposed to the shared ones in the grid, and can be employed only when the resource reliability becomes critical: when the bulk of the Bag of Tasks has been already completed, and there are only a few tasks left. We argue that the middleware requires a number of policy-driven runtime mechanisms, with the task replication being the most important one. On the example of the GridBot system we demonstrate that these mechanisms combined with the proper policy enable up to an order of magnitude savings in costs and twice as faster execution of short and long BoTs on a number of production grids and Amazon EC2, versus the standard demand-driven approach.

 

Back to Session VII

Cloud Computing and Enterprise HPC

 

Mathias Dalheimer

Fraunhofer Institute for Industrial Mathematics, GERMANY

 

Cloud computing provides a new model for accessing compute and storage capacity. Companies have to pay for operational expenses (OpEx) only, no capital expenses (CapEx) will be billed. This provides an attractive approach for most enterprises. On the other hand, the integration of external resources in enterprise IT environments holds some challenges. The talk demonstrates which HPC applications can take advantage of cloud computing offers and outlines the problems we experienced.

Back to Session VII

PAIMS: Precision Agriculture Information Monitoring System

 

Amy Wang

Institute for Theoretical Computer Science, Tsinghua University

Beijing, CHINA

 

This talk will introduce some of the considerations taken and results of research conducted for the large scale wireless sensor network system for agriculture information monitoring applications. The sensor networks and crucial national requirements for real-time farmland and crop information monitoring systems will be introduced first. Second, a Precision Agriculture Information Monitoring System (PAIMS) will be introduced, which is designed for the long-term monitoring of large scale farmland. The PAIMS system consists of a two-tiered sensor network and an information service platform. The sensor network contains a large number of energy-constrained low tier nodes (LNs) to capture and report information from their designated vicinity. Then, some powerful gateways in the high tier to organize the LNs to form clusters which report the aggregated information to the Internet. The information service platform logs information from the sensor network and provides value-created services to the user.  After giving an overview of the PAIMS system, selected research results in PAIMS will be introduced in detail, including, to name a few, a Multi-hop Joint Scheduling algorithm and a Distributed Dynamic Load balancing algorithm. The hardware and system implementation issues of PAIMS will also be introduced. Finally, we will discuss possible directions of future research.

Back to Session VIII

Design Patterns and the quest for General Purpose Parallel Programming

 

Tim Mattson

Principal Engineer Intel Labs

Hillsboro, OR, USA

 

In a many core world, all software should be parallel software.  It’s hard enough making our scientific applications parallel, but how are we going to parallelize the large mass of general purpose software?   In this talk, I suggest that software frameworks informed by design patterns and deployed on top of industry standard hardware abstraction layers (such as OpenCL) could solve this problem.  My goal, however, is not to present a tidy solution to the problem.  We have too many solutions chasing this problem already.  My goal is to establish a position and then engage in a vigorous debate leading to objective analysis and in the long run, a solution to this problem that actually works.

Back to Session VIII

A Multicore Processor Designed For Petaflops Computation

 

Weiwu Hu

Institute of Computing Technology

Chinese Academy of Sciences, Beijing, CHINA

 

Godson-3 is a multicore processor designed by Institute of Computing Technology, Chinese Academy of Sciences. As a processor targets Petaflops computation, Godson-3 concentrates on scalability: The two level interconnection network of Godson-3 supports up to 64 core; 1024-way Godson-3 can be connected together as a large SMP. On the other hand, Godson-3 has a high performance with low power consumption. Each core of Godson-3 adopts 512-bit vector processing unit which can complete 8  double-precision floating-point MADD in one clock cycle. Hence, an 8-core Godson-3 can achieve 128 Gflops peak performance at 1.0GHz prequency with about 20 Watt power consumption. Thus a petaflops supercomputer only needs less than 10,000 Godson-3 chips.

Back to Session VIII

Efficient Execution on Heterogeneous Systems

 

Leonel Sousa

INESC-ID/IST, TU Lisbon

 

Modern parallel and distributed systems rely on multi-core processors for improving performance. Furthermore, modern systems are configured with processors of different characteristics such as general-purpose multi-core processors and accelerators. While such systems provide a better match between the application requirements and the hardware provided, the efficient execution of applications on heterogeneous systems is currently an open issue. On the one hand, the user wants to write a single version of the program while, on the other hand, in order to exploit the available performance it is needed to tune the application to the target architecture.

Common programming models that allow the user to program once and execute on different architectures have recently been developed. Examples of such models include Rapidmind (recently Intel is integrating it with their own Ct programming model) and the OpenCL standard. Nevertheless, previous works have shown that there is a considerable penalty in using such an approach as the application is not tuned to the different target architectures in heterogeneous systems.

In this work we take a different approach and exploit the execution of multiple tuned versions of the same application on the heterogeneous system. While at this point we rely on hand-tuned versions, in the future dynamic compilation techniques may help tuning the code. The issue though is the coordination of the execution of these different tuned versions. As such, we focus on a high-level scheduler that coordinates the execution of these versions on the heterogeneous system.

To prove the proposed concept we selected a database workload as target application. We provide versions of the basic database algorithms, tuned for GPUs using CUDA, for Cell/BE using the Cell SDK and for general-purpose multi-cores using OpenMP. The application uses basic algorithms to process queries from the standard Decision-Support System (DSS) benchmark TPC-H. The scheduler distributes the work among the different processors and assigns the corresponding tuned versions of the code for each algorithm.

Back to Session VIII

High-Performance Computing with NVIDIA Tesla GPU

 

Tim Lanfear

NVIDIA Ltd, Reading, UNITED KINGDOM

 

GPU computing is the use of a GPU (graphics processing unit) for general purpose scientific and engineering computing. The architecture of a GPU matches well to many scientific computing algorithms where the same operation is applied to every element of a large data set. Many algorithms have been ported to GPUs and benefited from significant increases in performance when compared with implementations on traditional microprocessors. NVIDIA’s next generation of Tesla GPUs using the Fermi architecture offer all the features needed for deployment of GPUs in large-scale HPC systems, including enhanced double-precision performance, hierarchical caches, ECC protection of data on-chip and off-chip, and high memory bandwidth.

Back to Session VIII

Hybrid Computing for Solving High Performance Computing Problems

 

Janusz Kowalik and Piotr Arlukowicz

University of Gdansk, POLAND

 

Hybrid computing is a combination of sequential and highly parallel computing.Sequential computing is executed by CPU and parallel by GPU. This binomial mode of computation is becoming a practical tool for solving HPC problems. The paper presents the key architectural features of GPU and discusses performance issues. A large linear algebra example illustrates the benefits of the GPU acceleration.

Back to Session VIII

An Introduction to CUDA Programming: A Tutorial

 

Piotr Arlukowicz and Janusz Kowalik

University of Gdansk, POLAND

 

The tutorial is about the basics of NVIDIA accelerators architecture and the methods of programming with CUDA technology. It will deal with internal hardware structures, and will cover some simple, but low-level C code examples illustrating fundamental techniques used to obtain the processing speed and efficiency. This tutorial is intended for persons having programming experience but without CUDA programming expertise.

Back to Session VIII

Cyber Science Infrastructure in Japan - NAREGI Grid Middleware Version 1 and Beyond -

 

Kenichi Miura, Ph.D

Center for Grid Research and Development

National Institute of Informatics, Tokyo, Japan

 

The National Research Grid Initiative (NAREGI) Project was a research and development on the grid middleware from FY2003 to FY2007 under the auspices of the Ministry of Education, Culture, Sports, Science and Technology (MEXT). Now we are in the phase of deploying the NAREGI Version 1 grid middleware to the nine university-based national supercomputer centers and some domain specific research institutes such as the Institute for Molecular Science (IMS), National High Energy Accelerator Research Organization (KEK) and the National Astronomical Observatory (NAOJ). The National Institute of Informatics (NII) already initiated the realization of the computational research and educational environment called “Cyber Science Infrastructure” in 2005, based on the national academic backbone network, called SINET3, and the NAREGI grid middleware is a key component in it.

As a follow-on project to NAREGI, the “RENKEI (Resources Linkages for e-Science) Project” also started in September 2008. In this project, a new light-weight grid middleware and software tools are being developed in order to provide the connection between the NAREGI Grid environment and wider research communities. In particular, technology for the flexible and seamless accesses between the national computing center level and the departmental/laboratory level resources, such as computers, storage and databases is one of the key objectives. Also, this newly developed grid environment will be made interoperable with the major international grids along the line of OGF standardization activities.

 

http://www.naregi.org/index_e.html

http://www.e-sciren.org

 

Kenichi Miura is a professor in High-end Computing at the National Institute of Informatics (NII) and also the director of the Center for Grid Research and Development. He was the former project leader of the Japanese National Research Grid Initiative (NAREGI) project. He is also a fellow of Fujitsu Laboratories, Ltd and a visiting researcher of RIKEN in conjunction with the Next Generation Supercomputer Project. Dr, Miura is a member of the Engineering Academy of Japan.

Dr. Miura received his Ph.D. degree in Computer Science from the University of Illinois, Urbana-Champaign in 1973. Dr. Miura joined Fujitsu in 1973 and has since been engaged in high-end computing. From 1992 to January 1997, he was Vice President and General Manager of the Supercomputer Group at Fujitsu America, Inc., based in San Jose CA He is the recipient of the IEEE Computer Society’s Seymour Cray Award for 2009.

Back to Session IX

The road to sustainable eInfrastructures in Latin America

 

Raoul Ramos Pollan

CCETA-CIEMAT Computing Center, SPAIN

 

The EELA and EELA2 projects constitute the first effort to build a continental eInfrastructure in Latin America integrating both the diversities and commonalties of the different scientific communities, technicians and decision makers across the continent. Their legacy has been inherited by the GISELA project, starting in September 2010, which will consolidate the eInfrastructure organizational model devised by EELA2 in a sustainable manner. This sustainability is based on the integration of the continental Grid innitiative within the existing NREN's organization, profiting from existing structures and lead by CLARA, the academic research network linking NREN's in LA.

Back to Session IX

The South African National Grid : Blueprint for Sub-Saharan

e-Infrastructure

 

Bruce Becker

South African National Grid

Pretoria, SOUTH AFRICA

 

The Sub-Saharan Region of Africa has been isolated geographically and technologically from the rapid advances in e-Science due in part to a lack of infrastructures, such as high-performance networks, computing resources and skilled technicians. However, this situation is evolving very rapidly, due to the improved network connectivity of the region, and the significant investment made in South Africa in terms of e-Infrastructure. The South African National Grid aims to coordinate and improve the deployment of and access to resources for e-Science, enabling and catalysing scientific research and regional scientific collaboration. SAGrid is by design interoperable at the middleware layer with almost all EU-funded projects such as EUIndiaGrid, EUChinaGrid, EELA and of course EGEE/EGI. Several activities to unlock the collaborative potential of the network and distributed computing applications which exploit it, with a view to long-term developments towards a "Sub-Sahara AfricaGrid" along the lines of projects such as EUMedGrid will be presented.

Back to Session IX

GARUDA: Indian National Grid Computing Initiative

 

Dr. B.B. Prahlada Rao

Programme SSDG, C-DAC Knowledge Park

Bangalore, India

 

Nations are realizing the importance of new e-infrastructures to enable scientific progress and research competitiveness. Making grid / cloud infrastructures available to the research community is crucial and is important to the researchers and the development teams in India. C-DAC launched the Indian National Grid Computing Initiative-GARUDA with the support of DIT, Government of India.  GARUDA is a collaboration of science researchers & Application developers on a nationwide grid of computational nodes, mass storage and scientific instruments.  GARUDA aims to provide the technological developments required to enable data & compute intensive, and collaborative applications for the 21st century. In  previous  phase of  the project,  a nation-wide service oriented grid test bed  was enabled. It is based on  open standards. C-DAC, and GARUDA partners are enabling  applications of societal & national importance on this grid infrastructure. In the current Phase,  the focus of GARUDA is to build stable and reliable collaborative grid applications. The author focus on various research initiatives enabled on  GARUDA,  its evolution, the applications, and the deliverables of the project.  C-DAC is an active partner of the EU-IndiaGrid2  project, which is exploring various options of Interoperability between GARUDA in India and EGEE in Europe.

Back to Session IX

Building e-Science and HPC Collaboration in Asia

 

Simon Lin

Academia Sinica Grid Computing (ASGC)

Institute of Physics, Taipei, TAIWAN

 

Data deluge drives the evolution of new science paradigm and the Grid-based distributed computing infrastructure such as WLCG and EGEE.  Starting from WLCG since 2002, with the support of Asia Pacific Regional Operation Centre (APROC) running by the only WLCG Tier-1 Center in Asia – ASGC, Grid resource centers has growth from 6 sites in 2005 to 30 sites in 2010 and contributing to 16 virtual organizations, CPU utilization increases over 571 times in the past 5 years. Asia is geographically large and cultural diverse in nature. This region as a whole traditionally inexperienced in regional cooperation. With the emergence of new e-Science paradigm and e-infrastructure, we are leapfrogging forward together and undermining the barriers of different administration domains. LHC experiments and high energy physics are still the most substantial user communities in Asia, however, the world largest biomedical grid application on avian influenza drug discovery was initiated by Asia with close cooperation of the worldwide grid and user communities in 2006. Aiming for sustainability, foundation of the regional e-infrastructure was established by application and user communities, collaboration across institutes and disciplines.  EUAsiaGrid project is such a major driving force to extend the e-infrastructure to non-HEP applications in this region to expedite closer collaboration by regional focused topics.

 

Back to Session IX

PL-Grid: the first functioning National Grid Initiative in Europe

 

Marian Bubak

ICS/ACC Cyfronet AGH University of Science and Technology, Krakow, Poland

Informatics Institute, Universiteit van Amsterdam, The Netherlands

 

The goal of the PL-Grid Project [1] is to provide the Polish scientific community with an IT platform based on Grid computer clusters, enabling e-science research in various fields. This infrastructure will be both compatible and interoperable with existing European and worldwide Grid.

PL-Grid aims at significantly extending the amount of computing resources provided to the Polish scientific community (by approximately 215 TFlops of computing power and 2500 TB of storage capacity) and constructing a Grid system that will facilitate effective and innovative use of the available resources. PL-Grid will draw upon the experience of European initiatives, such as EGEE and DEISA, the scientific results attained by individual partners and the outcome of R&D activities carried out within the project.

An Operations Center is responsible for provision of the PL-Grid infrastructure, extending support to its users and ensuring consistent quality of service during and beyond the project's life cycle. One of environments already used is the PL-Grid virtual laboratory [2].

PL-Grid will engage in collaboration with end users from its inception, providing training services for approximately 700 participants and undertaking joint development activities on new applications and domain-specific services. The emergence of the PL-Grid framework is consistent with European Commission policies which actively encourage development and integration of computing Grids.

On 30 March 2010, Poland - as the first country in Europe - has initiated functioning of the National Grid Initiative (NGI).  The members of the PL-Grid elaborated the procedures concerning transition from EGEE ROC to NGI. These procedures pave the way for other countries.

 

Acknowledgements:

The Project is co-funded by the European Regional Development Fund as part of the Innovative Economy program under contract POIG.02.03.00-00-007/08-00.

 

References

 

[1] PL-Grid website: http://www.plgrid.pl/en
[2] PL-Grid virtual laboratory:
http://gs.cyfronet.pl/

Back to Session IX

DEISA and the European HPC Ecosystem

 

Wolfgang Gentzsch

The DEISA Project and Open Grid Forum, Germany

 

Over the last decade, the European HPC scenario has changed dramatically,

from a few scattered national HPC centers 10 years ago to an HPC ecosystem

of interoperating and collaborating HPC centers today serving virtual

organizations and communities tackling their  grand-challenge big-science

applications. The ever increasing complexity of science applications,

increasing demands of computational scientists, and the challenge of

building and deploying the fastest and most expensive HPC systems has

forced our HPC community to develop and implement one sound and joined HPC

ecosystem for Europe.

This widely agree ecosystem can best be described by a three-layer HPC

pyramid consisting, at its bottom level, of regional HPC centers and

national grid computing resources, and at its top level of the most

powerful HPC systems represented by the PRACE initiative. Central to this

HPC ecosystem is DEISA, the Distributed European Infrastructure for

Supercomputing Applications, connecting 15 of the most powerful

supercomputer centers in Europe, with seamless access to heterogeneous HPC

resources, identity management, security, and interoperability and

standardisation. In addition, we present high performance computing

applications on the DEISA infrastructure, as represented by the DEISA

Extreme Computing Initiative (DECI) or more recently, by so-called virtual

communities and their scientific endeavours.

Back to Session IX

HPC Infrastructure and Activity in Korea

 

Hongsuk Yi

KISTI (Korea Institute of Science and Technology Information)

 

The Supercomputing Center at Korea Institute of Science and Technology Information has deployed the largest open HPC architecture capable of scaling from departmental clusters all the way up to the largest supercomputer in Korea. Tachyon is a 300 teraflop SUN Constellation Cluster comprising 25,600 Intel Nehalem Quad-core processors. It provides an open petascale computing environment combining ultra-dense high performance computing, networking, storage and software into an integrated system. As the system is being optimized, understanding of performance and scalability characteristics is increasing. This talk will present the current status of HPC infrastructure and activities in Korea, and also provide research collaborations with both multi-core infrastructure and innovative GPU computing technology.

Back to Session IX

The Grid as Infrastructure for Sharing BioMedical Information: The Biomedical Informatics Research Network

 

Carl Kesselman

Information Sciences Institute, University of Southern California

Marina del Rey, Los Angeles, CA, USA

 

Increasingly, translational research and clinical practice is impeded by the ability to exchange diverse health related information between collaborating parties. The issue of data sharing in a health context is complicated by issues of privacy, heterogeneity of the underlying data types, diverse semantic models, and the fundamentally complex nature of the health-care ecosystem. In this talk, I will discuss the ramifications of the underlying systems complexity of the health care system and how Grids and the associated concept of virtual organizations can provide solutions to the problems that result from this complexity. I will illustrate how Grid infrastructure can be applied within a number of clinical and research applications, including as part of the Biomedical Informatics Research Network (BIRN), a national scale medical information sharing infrastructure supported by the NIH.

Back to Session X

System Level Acceleration for Multi-Scale Modelling in Physiological Systems

 

Tim David

Centre for Bioengineering, University of Canterbury

Christchurch, NEW ZEALAND

 

Computational models of human physiology require the integration of a number of physical scales from the nano-cellular to the human frame. Currently these models have relied on a single architecture on which to solve the resulting equation sets. In order to understand the important functions and relationships that make up the natural working human body, as well as pathological environments, scientists and engineers are required to solve integrated equations sets whose forms are mostly radically different. These are different essentially because of the variation in length scales that are required to be resolved. For instance a cell lining the human artery is of the order of a few hundred nanometers, where the artery itself can be of the order of a metre.

Several different computer architectures have appeared over the past twenty years, from the original Cray vector pipeline through the distributed memory systems of the Meiko computing surface to the new partially-mixed architectures of the IBM p-575 series and finally the massively parallel Blue Gene system. It is not difficult to understand how certain problems have evolved (and optimized) for certain architectures. With the present requirement for the solution multiple-scales problems it is pertinent to start to search for a more integrated solution algorithm that takes the best of both worlds.

A new computing paradigm - the System Level Accelerator - can be employed to solve complex multi-scale problems.

The presentation will show how we have mapped the cerebrovasculature on a p-575 SMP supercomputer, whilst simulating the autoregulation network via large binary trees by mapping on a fully distributed computing system, such as a Blue Gene supercomputer.

We will present load balancing data as well as full solutions for fluid flow (blood) throughout the entire cerebro-vasculature. a problem of this scale in unprecedented in computational physiology and bioengineering.

Back to Session X

Towards Scalable Online Interactive Applications on Grids and Clouds

 

Sergei Gorlatch

University of Muenster, Germany

 

We study a class of Internet-based environments with high interactivity of multiple users and soft real-time requirements.

Challenging examples of such applications are massively multi-player online games (MMOG) and high-performance e-Learning and training systems.

We describe the Real-Time Framework (RTF) developed at the University of Muenster. RTF provides a high-level application development and high-performance, distributed runtime platform. Our main focus is on application scalability, i.e. maintaining the real-time application constraints when the number of users increases. This is achieved by distributing and parallelizing computations during the application design, as well as by efficiently supporting computations and communication over multiple servers during runtime. We present the main design solutions of RTF and the experimental results of its use in modern Grid and Cloud systems.

Back to Session X

Simulation and Animation of Complex Flows Using 294912 Processor Cores

 

U. Ruede, J. Götz, K. Iglberger, C. Feichtinger, S. Donath

Lehrstuhl fuer Simulation, Universitaet Erlangen-Nuernberg

Erlangen, GERMANY

 

We employ the Jugene system at Jülich with almost 300,000 cores to simulate fluids with suspended particles based on a detailed microscopic fluid-structure-interaction. The largest runs involve more than a hundred billion fluid cells, resulting in one of the largest CFD computations to date. The simulations use novel parallel algorithms for the handling of moving obstacles that are fully resolved as individual geometric objects.

With these techniques, we can study flow effects resulting from particles of arbitrary, non-spherical shape. The efficient parallelization of the rigid-body dynamics algorithm and the two-way fluid-structure-interaction are major challenges, since they require a parallel collision detection and parallel collision response algorithms. We will present simulations with up to 264 million particulate objects that move and interact with the flow.

Despite its complexity, the overall simulation still runs at a parallel efficiency of better than 95%. The talk will focus on the design and analysis of the algorithms and their implementation for various supercomputer architectures, including the Blue Gene/P. Additionally, we will present experiments on heterogeneous architectures using GPUs as accelerator hardware.

Back to Session X

Asynchronous computing of irregular applications using the SVPN model and S-Net coordination

 

Alex Shafarenko1 and Alexei Kudryavtsev2

1) University of Hertfordshire, UK

2) Institute of Theoretical and Applied Mechanics, Novosibirsk, Russia

 

This talk will address a new design representation for irregular distributed applications: a Spinal Vector Petri Net (SVPN).

SVPN is a vector of identical graphs which have processing functions at the nodes and which have messages floating along the edges. It is a distant relative of Coloured Petri Nets in that it reveals the synchronisation structure of an asynchronous application and can be used for analysis of its behaviour. SVPN can naturally support irregular applications, such as Particles-in-Cells simulations of plasma and molecular dynamics, as message-driven processes, without the programmer being aware of any subtle concurrency issues. Our main contribution to knowledge is a demonstration that the mapping of SVPN to the coordination language S-Net is nearly mechanical, which makes it possible, at least in principle, to write asynchronous, irregular applications directly in S-Net and utilise its advance support for software engineering.

Back to Session X

Towards Collaborative Workbench for Science 2.0 Applications

 

Marian Bubak

ICS/ACC Cyfronet AGH University of Science and Technology, Krakow, Poland

Informatics Institute, Universiteit van Amsterdam, The Netherlands

 

This talk presents investigations aimed at building an universal problem solving environment which facilitates programming and execution of complex Science 2.0 applications running on distributed e-infrastructures.

Such applications involve experimentation and exploratory programming, with multiple steps which are not known in advance and often are selected ad-hoc depending on obtained results.

This research is based on experience gathered during development of the ViroLab Virtual Laboratory [1, 2, 3], the APPEA runtime environment for banking and media application in GREDIA project [4, 5], the GridSpace environment [6] for running in-silico experiments, on user requirements analysis during the initial phase of PL-Grid project [7], as well as on a concept of the Common Information Space which is developed in the framework of the UrbanFlood Project [8]. In this project, applications  have the form of reusable digital artifacts (such as scripts) that glue together loosely coupled components (plug ins) accessible via standardized programming interfaces and CIS enable distributed deployment of independent, loosely coupled information spaces, responsible for processing of data from separate sensor networks.

The workbench provide support for sharing and reuse of applications components, such as scripts, code snippets and whole experiments. It will be enabled to publish experiment as web application and obtained result data will be a subject to store, publish and share. As opposed to existing virtual laboratories, the workbench will support exploratory experimentation programming paradigm and multifaceted collaboration by sharing software, data and ready-to-use web applications.

 

Acknowledgements: This work is partially supported by the EU UrbanFlood

and PL-Grid projects.

References

[1] M. Bubak, M. Malawski, T. Gubala, M. Kasztelnik, P. Nowakowski, D. Harezlak, T. Bartynski, J. Kocot, E. Ciepiela, W. Funika, D. Krol, B. Balis, M. Assel, and A. Tirado Ramos. Virtual laboratory for collaborative applications. In M. Cannataro, editor, Handbook of Research on Computational GridTechnologies for Life Sciences, Biomedicine and Healthcare, chapter XXVII, pages 531-551. IGI Global, 2009.

[2] P.M.A Sloot, Peter V. Coveney, G. Ertayalan, V. Mueller, C.A. Boucher, and M. Bubak: HIV decision Support: from Molecule to Man. Philosophical Transactions ofthe Royal Society A, vol 367, pp 2691 - 2703, 2009, doi:10.1098/rsta.2009.0043.

[3] ViroLab Virtual Laboratory: http://virolab.cyfronet.pl

[4] D. Harezlak, P. Nowakowski, M. Bubak. Appea: A Framework for Design and Implementation of Business Applications on the Grid, Proceedings of Computational Science - ICCS 2008 , 8th International Conference Krakow, Poland, June 2008, volume III, LNCS 5103, Springer, 2008

[5] http://www.gredia.eu

[6] GridSpace http://gs.cyfronet.pl

[7] PL-Grid - Polish Infrastructure for Information Science Support in the European Research Space: http://www.plgrid.pl/en

[8] UrbanFlood EU Project: http://urbanflood.eu/

Back to Session X

On High Performance Software Component Models

 

Christian Perez

INRIA, FRANCE

 

Software component models appear as a solution to handle the complexity and the evolution of applications. It turns out to be a powerful abstraction mechanism for dealing with parallel and heterogeneous machines as it enable the structure of an application to be manipulated, and hence specialized. Challenges include the understanding and the definition of an adequate abstraction level of application description so as to enable machine specific optimization.

This talk will overview some works we have done in increasing the level of abstraction of component models for HPC applications, including component level extensions such as parallel components, or assembly level extensions such as dynamic assembly. It will also discuss the usage of model transformation techniques as a framework to optimize applications.

Back to Session X