Module- 01 Cc(Bcs601) Search Creators
Module- 01 Cc(Bcs601) Search Creators
Created By:
Hanumanthu
Dedicated To.
📺 YouTube: https://www.youtube.com/@searchcreators7348
📸 Instagram : https://www.instagram.com/searchcreators/
📱 Telegram: https://t.me/SearchCreators
💬 WhatsApp:+917348878215
Module-01
Over the last 60 years, computing has evolved through multiple platforms and
environments.
Shift from centralized computing to parallel and distributed systems.
Modern computing relies on data-intensive and network-centric architectures.
Platform Evolution
First Generation (1950-1970): Mainframes like IBM 360 and CDC 6400.
Second Generation (1960-1980): Minicomputers like DEC PDP 11 and VAX.
Third Generation (1970-1990): Personal computers with VLSI microprocessors.
Fourth Generation (1980-2000): Portable and wireless computing devices.
Fifth Generation (1990-present): HPC and HTC systems in clusters, grids, and cloud
computing.
Transparency in data access, resource allocation, job execution, and failure recovery is
essential.
Application domains:
OR
IoT extends the Internet to everyday objects, interconnecting devices, tools, and
computers via sensors, RFID, and GPS.
History: Introduced in 1999 at MIT, IoT enables communication between objects and
people.
IPv6 Impact: With 2¹²⁸ IP addresses, IoT can assign unique addresses to all objects,
tracking up to 100 trillion static or moving objects.
Communication Models:
o H2H (Human-to-Human)
o H2T (Human-to-Thing)
o T2T (Thing-to-Thing)
Smart Earth Vision: IoT aims to create intelligent cities, clean energy, better healthcare,
and sustainable environments.
CPS integrates computation, communication, and control (3C) into a closed intelligent
feedback system between the physical and digital worlds.
Features:
o IoT vs. CPS: IoT focuses on networked objects, while CPS focuses on VR applications
in the real world.
o CPS enhances automation, intelligence, and interactivity in physical environments.
Development:
Modern CPUs use multicore architecture (dual, quad, six, or more cores).
Instruction-Level Parallelism (ILP) and Thread-Level Parallelism (TLP) improve
performance.
Processor speed evolution:
o 1 MIPS (VAX 780, 1978) → 1,800 MIPS (Intel Pentium 4, 2002) → 22,000
MIPS (Sun Niagara 2, 2008).
Moore’s Law holds for CPU growth, but clock rates are limited (~5 GHz max) due to
heat and power constraints.
Modern CPU technologies include:
o Superscalar architecture, dynamic branch prediction, speculative execution.
o Multithreaded CPUs (e.g., Intel i7, AMD Opteron, Sun Niagara, IBM Power 6).
CPUs may scale to hundreds of cores but face memory wall limitations.
GPUs (Graphics Processing Units) are designed for massive parallelism and data-
level parallelism (DLP).
x86-based processors dominate HPC and HTC systems.
Multithreading Technology
GPUs were initially graphics accelerators, now widely used for HPC and AI.
First GPU: NVIDIA GeForce 256 (1999).
Modern GPUs have hundreds of cores, e.g., NVIDIA CUDA Tesla.
GPGPU (General-Purpose GPU Computing) enables parallel processing beyond
graphics.
Example 1.1 the NVIDIA Fermi GPU Chip with 512 CUDA Cores
o Need for self-aware OS, locality-aware compilers, and auto-tuners for GPU-
based computing.
Memory Technology
Flash memory & SSDs are revolutionizing HPC (High-Performance Computing) and HTC
(High-Throughput Computing).
SSD lifespan: 300,000–1 million write cycles per block, making them durable for years.
Storage trends:
System-Area Interconnects
Wide-Area Networking
VM Architectures
VM Primitive Operations
Virtual Infrastructures
Uses commodity x86 servers and Ethernet networks instead of expensive hardware.
Data Deluge: Massive data from sensors, web, simulations, requiring advanced data
management.
E-Science Applications: Used in biology, chemistry, physics, and social sciences.
MapReduce & Iterative MapReduce: Enable parallel processing of big data.
Multicore & GPU Clusters: Boost computational power for scientific research.
Cloud Computing & Data Science Convergence: Revolutionizes computing
architecture and programming models.
Four major types: Clusters, P2P Networks, Computing Grids, and Internet Clouds.
Involves hundreds, thousands, or even millions of participating nodes.
Clusters: Popular in supercomputing applications.
P2P Networks: Used in business applications but face copyright concerns.
Grids: Underutilized due to middleware and application inefficiencies.
Cloud Computing: Cost-effective and simple for providers and users.
Cluster Architecture
Ideal cluster design merges multiple system images into a single-system image.
SSI makes a cluster appear as a single machine to users.
Achieved using middleware or OS extensions.
Computational Grids
Grid Families
Types:
Computational/Data Grids: Built at a national level for large-scale computing and data
sharing.
P2P Architecture: Decentralized, with each node acting as both a client and a server.
P2P Systems
Overlay Networks
Types:
It shifts computing and data storage from desktops to centralized data centers, offering
on-demand services.
Cloud Computing
Internet Clouds
Security Considerations
Security is shared among cloud providers, users, and third-party software providers.
This section explores the software environments used in distributed and cloud
computing, focusing on Service-Oriented Architecture (SOA), web services, REST, and
the evolving relationship between grids and clouds.
Entities in SOA:
SOA builds on the OSI networking model, using middleware like .NET, Apache Axis, and
Java Virtual Machine.
Entity interfaces include Web Services Description Language (WSDL), Java methods,
and CORBA IDL.
Middleware support: WebSphere MQ, Java Message Service (JMS) for messaging, fault
tolerance, and security.
Service discovery models: JNDI (Java), UDDI, LDAP, ebXML, CORBA Trading Service.
Management services: CORBA Life Cycle, Enterprise JavaBeans, Jini lifetime model, and
web services frameworks.
Evolution of SOA
Grid Computing:
Cloud Computing:
Hybrid Approach:
o Grids can be built out of multiple clouds for better resource allocation.
o Models include cloud of clouds, grid of clouds, and interclouds.
o Technologies like BPEL Web Services, Pegasus, Taverna, Kepler, Trident, and Swift
help manage distributed services.
Future cloud computing will allow users to switch OS and applications easily.
MapReduce
Hadoop
Globus Toolkits
Dimensions of Scalability
Scalability is affected by OS image count. SMP systems have a single OS image, limiting
scalability, whereas NUMA, clusters, and cloud environments support multiple OS
images, enabling higher scalability.
Amdahl’s Law states that system speedup is limited by the sequential portion of a
program. Even with infinite processors, speedup is constrained by non-parallelizable
code.
Gustafson’s Law addresses this limitation by scaling the workload along with system
size, resulting in better efficiency in large distributed systems.
High Availability (HA) is crucial in clusters, grids, P2P networks, and clouds. It is defined
as: System Availability=
As system size increases, availability decreases due to higher failure probability. Grids
and clusters have better fault isolation than SMP and MPP systems, while P2P
networks have the lowest availability.
Common Threats:
Collusive Piracy: Paid clients (colluders) share copyrighted content with unpaid clients
(pirates), affecting commercial content delivery.
Content Poisoning Scheme: Proactively detects and prevents piracy using identity-based
signatures and timestamped tokens, protecting legitimate clients while stopping
colluders and pirates.
Reputation Systems: Essential in detecting and addressing piracy in P2P networks and
digital content sharing.
Second Generation: Tools for detecting intrusions (e.g., firewalls, IDS, reputation
systems) and triggering remedial actions.
Third Generation: Intelligent systems that respond to intrusions and adapt to security
threats.
Cloud Security: Cloud service models (IaaS, PaaS, SaaS) divide security responsibilities:
Piracy Prevention: Measures against online piracy and copyright violations in digital
content.
2. Unused Servers: Many servers in data centers are left powered on without use,
leading to significant energy waste (e.g., 4.7 million idle servers globally).
o Potential Savings: Estimated savings of $3.8 billion in energy costs and $24.7
billion in operational costs from turning off unused servers.
Application Layer:
Middleware Layer:
Resource Layer:
o Manages hardware (e.g., CPU) and operating systems to optimize energy usage.
o Dynamic Power Management (DPM): Switches between idle and lower-power states.
o Dynamic Voltage-Frequency Scaling (DVFS): Controls power consumption by
adjusting voltage and frequency.
Network Layer: