Module 1 ppt
Module 1 ppt
& Security
By Prof. Shivakumar M
www.cambridge.edu.
in
Chapter 1 assesses the evolutional changes in computing and IT
trends in the past 30 years, driven by applications with variable
workloads and large data sets.
On the HTC side, peer-to-peer (P2P) networks are formed for distributed
file sharing and content delivery applications. A P2P system is built over
many client machines. Peer machines are globally distributed in nature.
16
Meeting these goals requires to yield the following design
objectives:
⚫ Efficiency measures the utilization rate of resources in an
execution model by exploiting massive parallelism in
HPC. For HTC, efficiency is more closely related
to job throughput, data access, storage, and power
efficiency.
⚫ Dependability measures the reliability and self-
management from the chip to the system and
application levels. The purpose is to provide high-
throughput service with Quality of Service (QoS)
assurance, even under failure conditions.
⚫ Adaptation in the programming model measures the ability
to support billions of job requests over massive data
sets and virtualized cloud resources under various
workload and service models.
⚫ Flexibility in application deployment measures the ability of
distributed systems to run well in both HPC (science 17
and engineering) and HTC (business) applications.
Scalable Computing Trends and
New Paradigms
⚫ Degrees of Parallelism
when hardware was bulky and expensive, most computers were designed
in a bit-serial fashion.
bit-level parallelism (BLP): converts bit-serial processing to word-level
processing gradually.
Over the years, users graduated from 4-bit microprocessors to 8-,16-, 32-,
and 64-bit CPUs. This led us to the next wave of improvement, known as
instruction-level parallelism (ILP) , in which the processor executes
multiple instructions simultaneously rather than only one instruction at a
time.
For the past 30 years, we have practiced ILP through pipelining, super-
scalar computing,
VLIW (very long instruction word) architectures, and multithreading.
ILP requires branch prediction, dynamic scheduling, speculation, and
compiler support to work efficiently.
18
Data-level parallelism (DLP): was made popular through SIMD (single
instruction, multiple data) and vector machines using vector or array types of
instructions.
DLP requires even more hardware support and compiler assistance to work
properly. Ever since the introduction of multicore processors and chip
multiprocessors (CMPs) ,
we have been exploring task-level parallelism (TLP) .
A modern processor explores all of the aforementioned parallelism types. In
fact, BLP, ILP, and DLP are well supported by advances in hardware and
compilers. However, TLP is far from being very successful due to difficulty in
programming and compilation of code for efficient execution on multicore
CMPs.
As we move from parallel processing to distributed processing, we will
see an increase in computing granularity to job-level parallelism (JLP)
. It is fair to say that coarse-grain parallelism is built on top of fine-grain
parallelism
19
⚫ Innovative Applications
Few key applications that have driven the development of
parallel and distributed systems over the years
2009 2011
2008
2007
Cyber-Physical Systems
A cyber-physical system (CPS) is the result of interaction between
computational processes and the physical world. A CPS integrates “
cyber ” (heterogeneous, asynchronous) with“ physical ” (concurrent
and information-dense) objects. A CPS merges the “ 3C ”
technologies of computation, communication , and control into an
intelligent closed feedback system between the physical world and
the information world, a concept which is actively explored in the
United States.
24
TECHNOLOGIES FOR NETWORK-BASED
SYSTEMS
Time to explore hardware, software, and network technologies for distributed
computing system design and applications.
Performance Metrics and Scalability
Analysis
• Performance metrics are needed to measure various distributed systems.
• When scaling a system design with new technology one must consider three
aspects: time, space, and heterogeneity.
(1) Time refers to generation scalability. When changing to new-generation
processors, one must consider the impact to the motherboard, power supply,
packaging and cooling, and so forth.
(2) Space is related to packaging and energy concerns.
(3) Heterogeneity refers to the use of hardware components or software
packages from different vendors.
Scalability versus OS Image Count
Scalable performance implies that the system can achieve higher speed by adding
more processors or servers, enlarging the physical node’s memory size, extending
the disk capacity, or adding more I/O channels.
• Many cluster nodes are SMP or multicore servers and hence the
total number of processors or cores in a cluster system is one or two
orders of magnitude greater than the number of OS images running
Amdahl’s Law
• Assume that a fraction α of the code must be executed
sequentially, called the sequential bottleneck. Therefore, (1 − α)
of the code can be compiled for parallel execution by n
processors.
• Very often the system efficiency is rather low, especially when the
cluster size is very large. This is because only a few processors
(say, 4) are kept busy, while the majority of the nodes are left