Unit 2
Unit 2
Cloud Computing
Concepts
Cloud Concepts
• Cloud Computing Architecture
• Cloud Reference Model
• Cloud Layers
• Types of Clouds
• Services models
• Data centre Design and interconnection Network
• Architectural design of Compute Storage Clouds,
• "Cloud Programming and Software: Fractures of cloud programming"
Parallel and distributed programming paradigms
• Map Reduce, Hadoop
• High level Language for Cloud.,
CLOUD COMPUTING ARCHITECTURE
1. Physical Layer
2. Virtual Layer
• Deployed on the physical layer.
• Specifies entities that operate at this layer : Virtualization software,
resource pools, virtual resources.
• Functions of virtual layer : Abstracts physical resources and makes
them appear as virtual resources (enables multitenant environment).
Executes the requests generated by control layer.
The Concept of Virtualization:
• Developers require multiple operating systems to build different
systems in different environments. For testers, it will be an easier
option as they can check different systems in different environments.
3. Control Layer
• Deployed either on virtual layer or on physical layer
• Specifies entities that operate at this layer : control software
• Functions of control layer : Enables resource configuration,
resource pool configuration and resource provisioning. Executes
requests generated by service layer. Exposes resources to and
supports the service layer. Collaborates with the virtualization
software and enables resource pooling and creating virtual
resources, dynamic allocation and optimizing utilization of resources.
Control Software
• Software control is the process of the
physical storage in a definitive software
library of all software to ensure that only
correctly released, licensed, and authorized
versions of software are in use. Processes in
this category include the tracking of releases,
patches, service packs, and bug fixes.
4. Service Orchestration Layer
• Specifies the entites that operate at this layer : Orchestration
software.
• Functions of orchestration layer : Provides workflows for executing
automated tasks. Interacts with various entities to invoke
provisionning tasks.
What is Orchestration?
• Orchestration is the coordination and management of multiple
computer systems, applications and/or services, stringing together
multiple tasks in order to execute a larger workflow or process. These
processes can consist of multiple tasks that are automated and can
involve multiple systems.
• The goal of orchestration is to streamline and optimize the execution
of frequent, repeatable processes and thus to help data teams more
easily manage complex tasks and workflows. Anytime a process is
repeatable, and its tasks can be automated, orchestration can be used
to save time, increase efficiency, and eliminate redundancies. For
example, you can simplify data and machine learning with jobs
orchestration.
5. Service Layer
• Consumers interact and consume cloud resources via thos layer.
• Specifies the entities that operate at this layer : Service catalog
and self- service portal.
• Functions of service layer : Store information about cloud services
in service catalog and presents them to the consumers. Enables
consumers to access and manage cloud services via a self-service
portal.
Cross-layer function
1.Business continuity
• Specifies adoption of proactive and reactive measures to mitigate
the impact of downtime.
• Enables ensuring the availability of services in line with SLA.
• Supports all the layers to provide uninterrupted services.
2. Security
• Specifies the adoption of : Administrative mechanisms (security
and personnel policies, standard procedures to direct safe
execution of operations) and technical mechanisms (firewall,
intrusion detection and prevention systems, antivirus).
• Deploys security mechanisms to meet GRC requirements.
• Supports all the layers to provide secure services.
3. Service Management
• Specifies adoption of activities related to service portfolio management and
service operation Management.
A. Service portfolio management :
• Define the service roadmap, service features, and service levels
• Assess and prioritize where investments across the service portfolio are most
needed
• Establish budgeting and pricing
• Deal with consumers in supporting activities such as taking orders,
processing bills, and collecting payments
B. Service operation management :
• Enables infrastructure configuration and resource provisioning
• Enable problem resolution
• Enables capacity and availability management
• Enables compliance conformance
• Enables monitoring cloud services and their constituent elements
CLOUD DEPLOYMENT TYPES
Cloud Service Models
There are the following three types of cloud service models -
• Infrastructure as a Service (IaaS)
• Platform as a Service (PaaS)
• Software as a Service (SaaS)
Infrastructure as a Service (IaaS)
• IaaS is also known as Hardware as a Service (HaaS). It is a computing
infrastructure managed over the internet. The main advantage of
using IaaS is that it helps users to avoid the cost and complexity of
purchasing and managing the physical servers.
• It is one of the layers of the cloud computing platform. It allows
customers to outsource their IT infrastructures such as servers,
networking, processing, storage, virtual machines, and other
resources. Customers access these resources on the Internet using a
pay-as-per use model.
• IaaS cloud computing platform layer eliminates the need for every
organization to maintain the IT infrastructure.
• IaaS is offered in three models: public, private, and hybrid cloud.
Characteristics of IaaS
There are the following characteristics of IaaS -
• Resources are available as a service
• Services are highly scalable
• Dynamic and flexible
• GUI and API-based access
• Automated administrative tasks
• Calculation
• enterprise data storage
• networking
How are Datacenters managed?
• Facilities Management. Management of a physical data center facility may
include duties related to the facility's real estate, utilities, access control,
and personnel.
• Datacenter inventory or asset management. Datacenter features include
hardware assets and software licensing, and release management.
• Datacenter Infrastructure Management. DCIM lies at the intersection of IT
and facility management and is typically accomplished by monitoring data
center performance to optimize energy, equipment, and floor use.
• Technical support. The data center provides technical services to the
organization, and as such, it should also provide technical support to the
end-users of the enterprise.
A Generic Cloud Architecture
CLOUD COMPUTING MAPREDUCE IN HADOOP
• What is MapReduce in Hadoop?
• MapReduce is a software framework and programming model used
for processing huge amounts of data.
• MapReduce program work in two phases, namely, Map and Reduce.
• Map tasks deal with splitting and mapping of data while Reduce tasks
shuffle and reduce the data.
• The programs of Map Reduce in cloud computing are parallel in
nature, thus are very useful for performing large-scale data analysis
using multiple machines in the cluster.
• The input to each phase is key-value pairs. In addition, every
programmer needs to specify two functions: map function and
reduce function.
MAP and REDUCE
MapReduce Architecture in Big Data
phases of MapReduce in Big Data
• Input Splits:
An input to a MapReduce in Big Data job is divided into fixed-size
pieces called input splits Input split is a chunk of the input that is
consumed by a single map
• Mapping
This is the very first phase in the execution of map-reduce program. In
this phase data in each split is passed to a mapping function to produce
output values. In our example, a job of mapping phase is to count a
number of occurrences of each word from input splits (more details
about input-split is given below) and prepare a list in the form of
<word, frequency>
• Shuffling
This phase consumes the output of Mapping phase. Its task is to
consolidate the relevant records from Mapping phase output. In our
example, the same words are clubed together along with their
respective frequency.
• Reducing
• In this phase, output values from the Shuffling phase are aggregated.
This phase combines values from Shuffling phase and returns a single
output value. In short, this phase summarizes the complete dataset.
• In our example, this phase aggregates the values from Shuffling phase
i.e., calculates total occurrences of each word.
How MapReduce Organizes Work?
Hadoop divides the job into tasks. There are two types of tasks:
• Map tasks (Splits & Mapping)
• Reduce tasks (Shuffling, Reducing)
The complete execution process (execution of Map and Reduce tasks,
both) is controlled by two types of entities called a
• Jobtracker: Acts like a master (responsible for complete execution of
submitted job)
• Multiple Task Trackers: Acts like slaves, each of them performing the
job
• For every job submitted for execution in the system, there is
one Jobtracker that resides on Namenode and there are multiple
tasktrackers which reside on Datanode.
Difference between PC/CC
Difference between CC/DC