High Performance Computing
High Performance Computing
COMPUTING
Who needs HPC?
What defines High Performance
Computing?
“We use HPC to solve problems that could
not be solved in a reasonable amount of time
using conventional mainframe machines.”
HISTORY
In 1994, Thomas Sterling and Don Becker, working at
the center of excellence in space data and information
sciences (CESDIS) under the sponsorship of the earth
and space sciences project, built a cluster computer
called “Beowulf” which consisted of 16 DX4 processors
connected by channel bonded 10Mbps Ethernet.
Applications
Programming Paradigms
Threads Interface
Microkernel Operating System
Multi-Processor Computing System
P P P P P .. P Hardware
of
Checkpointing, Automatic Failover, recovery
from failure,
fault-tolerant operating among all cluster
nodes.
Cluster Computing - Commercial
Software
Load Leveler - IBM Corp., USA
LSF (Load Sharing Facility) - Platform Computing,
Canada
NQE (Network Queuing Environment) - Craysoft Corp.,
USA
Open Frame - Centre for Development of Advanced
Computing, India
RWPC (Real World Computing Partnership), Japan
UnixWare (SCO-Santa Cruz Operations,), USA
Solaris-MC (Sun Microsystems), USA
Cluster Tools (A number for free HPC clusters tools
from Sun)
Computational Power
Improvement
Multiprocessor
C.P.I.
Uniprocessor
1 2. . . .
No. of Processors
Characteristics MPP SMP Cluster
Myrinet
160 MB/s
Fast comm.
AM, MPI, ...
Ether/ATM
switched
external net
Global OS
Self Config
HA Cluster: Server Cluster with
"Heartbeat" Connection
Clusters of Clusters
(HyperClusters)
Cluster 1
Scheduler
Master
Daemon
LAN/WAN
Submit
Graphical Cluster 3
Control Execution
Daemon Scheduler
Clients
Master
Daemon
Cluster 2
Scheduler Submit
Graphical
Master Control Execution
Daemon Daemon
Clients
Submit
Graphical
Control Execution
Daemon
Clients
MPI (Message Passing Interface)
A standard portable message-passing library definition.
Developed in 1993 by a group of parallel computer
vendors, software writers, and application scientists.
Target platform is a distributed memory system.
All inter-task communication is by message passing
Portable (once coded, it can run on virtually all HPC
platforms including clusters!)
Available for both Fortran and C programs.
Available on a wide variety of parallel machines.
All parallelism is explicit: the programmer is responsible
for parallelism the program and implementing the MPI
constructs.
PVM (PARALLEL VIRTUAL
MACHINE)
Originally developed in 1989 as a research tool to
explore heterogeneous network computing by Oak Ridge
National Laboratory. Now available as a public domain
software package.
Enables a collection of different computer systems to be
viewed as a single parallel machine.
All communication is accomplished by message passing.
PVM enables users to exploit their existing computer
hardware to solve much larger problems at minimal
additional cost.
The software is very portable. The source,available free
through netlib, has been compiled on everything from
laptops to CRAYs.
Cluster Computing - Research
Projects
Beowulf (CalTech and NASA) - USA
CCS (Computing Centre Software) - Paderborn, Germany
DQS (Distributed Queuing System) - Florida State University, US.
HPVM -(High Performance Virtual Machine),UIUC&now UCSB,US
MOSIX - Hebrew University of Jerusalem, Israel
MPI (MPI Forum, MPICH is one of the popular implementations)
NOW (Network of Workstations) - Berkeley, USA
NIMROD - Monash University, Australia
NetSolve - University of Tennessee, USA
PVM - Oak Ridge National Lab./UTK/Emory, USA
PARAM PADMA
PARAM Padma is C-DAC's next generation high
performance scalable computing cluster,
currently with a peak computing power of One
Teraflop. The hardware environment is powered
by the Compute Nodes based on the state-of-
the-art Power4 RISC processors. Technology.
These nodes are connected through a primary
high performance System Area Network,
PARAMNet-II, designed and developed by C-
DAC and a Gigabit Ethernet as a backup
network.
ONGC CLUSTERS
ONGC implements two LINUX cluster machines :
One is a 272 nodes dual core computing system
with each node equivalent to two CPU’s. The
master node has 12 nodes dual CPU and a
32terabyte SAN storage.
The second system has 48 nodes i.e 96 CPUs
code computing nodes and the master node has
4 nodes and 20 terabyte SAN storage.
APPLICATIONS OF CLUSTER-
TECHNOLOGY
Numerous Scientific & Engineering Apps.
Parametric Simulations
Business Applications
E-commerce Applications (Amazon. COM, eBay.com ….)
Database Applications (Oracle on cluster)
Decision Support Systems
Internet Applications
Web serving / surfing
Info wares (yahoo.com, AOL.com)
ASPs (application service providers)
Email, eChat, ePhone, eBanking, etc.
Computing Portals
Mission Critical Applications
command control systems, banks, nuclear reactor control and
handling life threatening situations.
WHAT IS GRID COMPUTING?
IT IS A TYPE OF PARALLEL AND
DISTRIBUTED SYSTEM THAT ENABLES THE
SHARING, SELECTION AND AGGREGATION
OF GEOGRAPHICALLY DISTRIBUTED
AUTONOMOUS RESOURCES, DYNAMICALLY
AT RUNTIME DEPENDING ON THEIR
AVAILABILITY, CAPABILITY, PERFORMANCE
AND COST.
IT F OLLOWS SE RVICE ORIENTED
ARCHITECTURE(SOA) AND PROVIDES
HARDWARE AND SOFTWARE SERVICES FOR
SECURE AND UNIFORM ACCESS TO
HETEROGENEOUS RESOURCES, ENABLES
FORMATION AND MANAGEMENT OF VIRTUAL
ORGANIZATIONS
/
HISTORY OF GRID COMPUTING
The term Grid computing originated in the
early 1990s as a metaphor for making
computer power as easy to access as an
electric power grid in Ian Foster and Carl
Kesselmans seminal work, "The Grid:
Blueprint for a new computing infrastructure".
They lead the effort to create the globus
toolkit,which remains the defacto standard for
grid computing.
FATHER OF THE GRID
“IAN FOSTER”
With Steven Tuecke and
Carl Kesselman he began
the Globus Project, a
software system for
international scientific
collaboration. They
envisioned Globus software
that would link sites into a
“virtual organization,” with
standardized methods to
authenticate identities,
authorize specific activities,
and control data movement.
Towards Grid Computing….
GRID COMPUTING IS NEEDED
FOR…
Harnessing idle CPU cycles from desktop
computers in the network (e.g.
SETI@home,Alchemi)
Sharing of data resources across multiple
organizations (e.g. SRB)
Replication and collaborative analysis of
distributed datasets (e.g. Globus, Gridbus)
Supporting market based allocation and
management of grid resources and turn grid
services into utilities (e.g. Gridbus).
TYPES OF GRID
Computational grids: focuses primarily on
computationally intensive operations.
Data grids: controlled sharing and
management of large amount of
distributed data.
Equipment grid: have a primary piece of
equipment where the surrounding grid is
used to control the equipment remotely
and to analyze the data produced.
GRID MIDDLEWARE
Software tools and services providing the
capability to link computing capability and data
sources in order to support distributed analysis
and collaboration are collectively known as grid
middleware.
Consists of 4 layers:
1. Fabric.
2. Core Middleware.
3. User-level middleware.
4. Applications and portal layers.
GRID COMPONENTS
Applications and Portals Grid
User-
Development Environments and Tools level
Languages Libraries Debuggers Monitoring Resource Brokers …Web tools middlew
are
Grid Forum
European Grid Forum
IEEE TFCC!
GRID’2000
CHALLENGES IN GRID
Heterogeneity – that results because of
the vast range of technologies, both
software and hardware, encompassed by
the grid.
Handling of grid resources – that spread
across political and geographical
boundaries.
GRID EVOLUTION FROM
RESEARCH TO INDUSTRY
• In April 2004,IT venders-EMC,HP, Intel
,Network Appliance, Oracle and Sun have
taken step towards developing enterprise
gird solutions by establishing the EGA.
SHREYA THAKUR
Laxmi Devi Institute of Engineering & Technology
4 Year, ECE.
th