Data Protection and Management Participant Guide 1 PDF
Data Protection and Management Participant Guide 1 PDF
MANAGEMENT
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Data Protection and Management-SSP
Replication.............................................................................................................. 193
Replication ....................................................................................................................... 194
Replication ....................................................................................................................... 195
Data Replication Overview ............................................................................................... 196
Knowledge Check: Data Replication Overview ................................................................ 200
Local Replication.............................................................................................................. 201
Knowledge Check: Local Replication ............................................................................... 210
Remote Replication.......................................................................................................... 211
Knowledge Check: Remote Replication ........................................................................... 216
Concepts in Practice ........................................................................................................ 217
Exercise- Replication ....................................................................................................... 219
Summary................................................................................................................. 452
Summary ......................................................................................................................... 453
You Have Completed This eLearning............................................................................... 454
Course Objectives
Objectives
Data protection is one of the least glamorous yet important aspect in any
organization. The sensitive data of an organization must be safeguarded so that
miscreants can't use that data to demand a ransom, encrypt it to make it
unavailable to the organization, publicly release an organization's client data, and
many other crimes. So, protecting an organization's data is of the utmost
importance.
Sensitive data, if lost, may lead to significant financial, legal, and business loss
apart from serious damage to the organization’s reputation.
For more information about the need for data protection, click here.
The correlation between data protection and availability has been depicted in the
following illustration:
Process of safeguarding data from corruption and loss Ability of an IT infrastructure component/service to function
as required during its operating time
Involves technologies/solutions that can prevent data Involves technologies, strategy, procedure, and IT
loss and recover data resource readiness appropriate for application/service
MTBF is the average time available for a component or a service to perform its
normal operations between failures. It is calculated as the total uptime divided by
the number of failures.
Hardware Failure
Disaster
Loss of Power
IT Infrastructure Refresh
Software Failure
Ransomware
Data Loss
Click the highlighted boxes on the image for more information about the impacts.
1: The loss of productivity can be measured in terms of the salaries, wages, and
benefits of employees that are made idle by an outage. It can be calculated as -
Number of employees impacted x hours of outage x hourly rate.
2:
• Revenue recognition
• Cash flow
• Lost discounts
• Payment guarantees
• Credit rating
• Stock price
• Direct losses
• Compensatory payments
• Future revenue losses
• Investment losses
5: The other possible consequences of outage include the cost of additional rented
equipment, overtime, and extra shipping.
A data center usually stores large amounts of data and provides services to a vast
number of users. Therefore, data protection in a data center is vital for carrying out
business operations.
Data Center B
Server-to-server Data Copy (North America)
Inter-data
Center Data
Copy
Management Data Center
Servers Servers
and Cloud
Connectivity Copy
Storage
1. What is the availability of a computer with MTBF = 8000 hours and MTTR = 12
hours?
a. 99.5%
b. 98.9%
c. 90%
d. 99.8%
Data Center
Data Center
Objectives
Facility
It is the building and floor space where the data center is constructed. It typically
has a raised floor with ducts underneath holding power and network cables.
IT equipment
Support infrastructure
It includes power supply, fire and humidity detection systems; heating, ventilation
and air conditioning (HVAC) systems; and security systems such as biometrics,
badge readers, and video surveillance systems.
1 2 3
Storage devices
Storage
System
Storage systems are designed for high
capacity, scalability, performance, reliability,
and security .
Storage may be internal (for example, internal hard disk drives, SSDs), removable
(for example, memory cards), or external (for example, magnetic tape drive) to a
compute system.
Management
Virtualization
Compute
Storage
Network
Deploy a fully-virtualized
environment in just 20
minutes
Unlike Converged Infrastructure (CI), which relies on hardware and uses physical
building blocks, HCI is software-defined. Moreover, HCI is more flexible and
scalable than CI.
2 3
4 5
Objectives
Data protection and availability solutions assure that the data is safe and
accessible to the intended users at a required level of performance. Different
solutions may be used in the same data center environment.
This is the point-in-time to which data must be recovered after an outage. It defines
the amount of data loss that a business can endure. Based on the RPO,
organizations plan for the frequency with which a backup or replica must be made.
For example, if the RPO of a particular business application is 24 hours, then
backups are created every midnight. The corresponding recovery strategy is to
restore data from the set of last backups. An organization can plan for an
appropriate data protection solution on the basis of the RPO it sets.
This is the time within which systems and applications must be recovered after an
outage. It defines the amount of downtime that a business can endure and survive.
Based on the RTO, an organization can decide which data protection technology is
best suited. The more critical the application, the lower the RTO should be.
Both RPO and RTO are counted in minutes, hours, or days and are directly related
to the criticality of the IT service and data. Usually, the lower the RTO and RPO,
the higher is the cost of a data protection solution or technology.
Time
Fault-tolerant IT Infrastructure
Redundant Redundant
Links Ports
Redundant Switches
Compute System
Port
Link
Switch
Storage
System
Compute
Cluster Single Point of
Failure
Data Backup
Data backup is the process of making a copy of primary data for the purpose of
restoring the original data in the event of data loss or corruption.
The backup data should not be kept in the same storage device where the original
data is stored. Otherwise, both the original data and the backup data will be lost if
physical damage occurs to the storage device. Often, data backups are performed
both within and between sites or data centers. The local backup within a site
enables easy access to the backup data and quick recovery. The backup data at
the remote site (cloud) provides protection against a disaster, which could damage
or destroy the local backup data.
Data Replication
Data replication is the process of creating an exact copy (replica) of the data so
that the data copy may be used to restore the original data in the event of a data
loss or corruption, or to restart business operations in case the primary storage is
not operational.
A replica can also be used to perform other business operations such as backup,
reporting, and testing. Data replication is similar to data backup, but it provides
higher availability because the replica can be made operational immediately after
the primary storage failure. Replication can be performed both within and across
data centers or clouds.
Data Archiving
Data archiving is the process of identifying and moving inactive data from primary
storage systems to lower cost storage systems, called data archives, for long term
retention. A data archive stores older but important data that is less likely to be
accessed frequently.
Data Migration
Data migration is the process of moving data between storage systems or compute
systems. A change in data format due to a system upgrade is also considered a
data migration. Data migration has several use cases.
Data Security
Counter Measures
Security Threats
Data security refers to the countermeasures that are used to protect data against
unauthorized access, deletion, modification, or disruption. It provides protection
against security threats that can potentially destroy or corrupt data and cause data
and service unavailability.
There are solutions like Dell EMC™ Cyber Recovery, which offer protection to
organizations against ransomware and other devastating attacks. With such a
solution in place, the organization is equipped with immutable clean backups, kept
safely in their vault, even in the case of production or backup data infiltration. This
way the organization can protect itself from huge data and revenue losses and
minimize downtime because of data unavailability.
Many organizations use a range of data protection as service offerings from cloud
providers. These offerings serve as a means of outsourcing non-strategic activities
as well as improving data protection and availability levels for certain workloads. As
the data protection infrastructure is maintained by the cloud provider, the expenses,
tasks, and time associated with data protection management is reduced. The
reduction of management tasks can drive new business initiatives, discovery of
new markets, and innovation.
Click the highlighted boxes in the given image for more information about the data
protection management activities.
1
4
2 5
Concepts in Practice
Concepts in Practice
Click the right and left arrows to view all the concepts in practice.
VxBlock combines industry-leading technologies that include Dell EMC storage and
data protection options, Cisco UCS blade and rack servers, Cisco LAN and SAN
networking, and VMware virtualization and cloud management into one fully
integrated system. It leverages its deep VMware integration to simplify automation
of everything from daily infrastructure provisioning tasks to delivery of IaaS and
SaaS.
1. Present Scenario:
2. Expected Deliverables:
• What are the expected annual uptime and downtime of the storage system?
• What is the achieved availability of the data archiving service in the last
year?
Solution
= 0.9988 × 100
= 99.88 %
Expected annual uptime of the storage system = 8760 hours per year × (0.9988) ≈
8749.5 hours
Expected annual downtime of the storage system = 8760 hours per year × (1 −
0.9988) ≈ 10.5 hours
= 0.9918 × 100
= 99.18 %
• Data source
• Protection application and storage
• Data security and management
Click on each highlighted box label for detailed information about the components.
5
2 3
1 4
1: It is the source of the data that must be protected. The data source can be a
business application, a hypervisor, or primary storage.
4: Both the protection application and the protection storage interact with the data
sources. During interaction, they can identify the data that needs protection.
5: The data security and management component interacts with other components
of the data protection architecture to exchange data, command, and status.
Objectives
Business Applications
API
GUI CLI
Interaction
Interaction
Business Users
(Clients)
Protection/Management/Ot
her Business Applications
Backup
Storage
API_Routine_SendCopy()
Call to Start Backup
Data set
Database Backup list
Application Application
2A REST API is an application program interface (API) that uses HTTP requests to
Get, Put, Post, and Delete data. An API for a website is code that allows two
software programs to communicate with each other.
3The hypervisor provides a compute virtualization layer that abstracts the physical
hardware of a compute system from the OS and enables the creation of multiple
VMs.
• Each Virtual Machine (VM) is isolated from the other VMs on the same physical
compute system. Therefore, the application running on one VM does not
interfere with those running on other VMs.
• The isolation also provides fault tolerance so that if one VM crashes, the other
VMs remain unaffected.
• A VM appears as a physical compute system with its own CPU, memory,
network controller, and disks.
• The compute system on which a hypervisor is running is called a host machine
and each VM is called a guest machine.
Virtual Machine
A Virtual Machine (VM) is a logical compute system with virtual hardware on which
a supported guest OS and application run. From the perspective of the guest OS, a
VM appears as a physical compute system.
• Each VM has its own configuration for hardware, software, network, and
security.
− Hardware and software are configured to meet the application’s
requirements.
• The image shows the typical virtual hardware components of a VM.
− This includes virtual CPU(s), virtual motherboard, virtual RAM, virtual disk,
virtual network adapter, optical drives, serial and parallel ports, and
peripheral devices.
• VM is a discrete set of files such as - Configuration file8, virtual disk file9,
Memory state file10 and Log file11.
Virtual Appliance
9Stores the content of a VM’s disk drive. A VM can have multiple virtual disk files,
each of which appears as a separate disk drive to the VM.
11Used to keep a record of the VM’s activity and is often used for troubleshooting
purposes.
• Virtual appliances13 are not so different from the physical appliances which are
used in kitchen, office, and data centers to perform specific tasks.
• Created using Open Virtualization Format (OVF)14 and simplifies the
deployment of an application.
− Appliance should be planned in such a way that it can easily run on the
hypervisor that is used in the organization’s cloud environment.
− Performance is limited to the resources of the hypervisor and it may
compete for resources with other VMs running on the same hypervisor.
− When deploying a virtual appliance, VM attributes need to be described by
providing the virtual appliance’s metadata15.
13The virtual appliance is a software packaged into a virtual format that is quickly
and easily deployed on a hypervisor.
15Metadata contains attributes of virtual machine such as RAM size and number of
processors.
Containers
Containers VMs
Stateless Stateful
Objectives
Read Write
Read Write
Internal/External Drive
Storage System
Primary Storage Device
• A primary storage device is the persistent storage for data used by business
applications to perform transactions.
• Data from a primary storage device can be copied or moved directly to a
protection storage to run business applications and hypervisors.
Controller
Storage
(HDD/SSD)
A primary storage system has two key components – controller and storage.
Click on "Scale-up" or "Scale-out" in this image for more information about the
architecture.
1:
It provides the capability to scale the capacity and performance of a single storage
system based on requirements. Scaling up a storage system16 involves upgrading
or adding controllers and storage.
2:
16 Storage systems typically have a fixed capacity ceiling, which limits their
scalability. Performance may also start to degrade when reaching the capacity
limit.
Based on the supported level of data access, primary storage systems can be
classified as:
SAN-Attached Storage
• SAN connects block-based storage with each other and to the compute
systems.
• SAN-attached storage improves the utilization of storage resources compared
to a direct-attached storage (DAS) environment.
− This reduces the total amount of storage that an organization needs to
purchase and manage.
− Storage management becomes centralized and less complex, which further
reduces the cost of managing data.
18The replication over long-distances helps in protecting data against local and
regional disaster.
NAS Clients
NAS Device
• Administrators create file systems on NAS systems, create shares, and export
shares to NAS clients.
• Enables clients to share files over an IP-based network.
20 File sharing, as the name implies, enables users to share files with other users.
− It enables both UNIX and Microsoft Windows users to share the same data.
• Uses file-sharing protocols such as CIFS and NFS to provide access to the file
data.
• NAS device uses its own OS and integrated hardware and software
components to meet specific file-service needs.
OSD stores data in the form of objects on a flat address space21. All objects exist at
the same level and an object cannot be placed inside another object.
• Object stored in an OSD is identified by a unique identifier called the object ID22.
• OSD provides a metadata service that is responsible for generating object ID
from the content of a file.
− Metadata service maintains the mapping of the object IDs and the file
system namespace.
• When an application server sends a read request to the OSD, the metadata
service retrieves the object ID for the requested file.
− The object ID is used to retrieve and send the file to the application server.
21Unlike file systems that have restrictions on the number of files, directories and
levels of hierarchy, the flat address space has no hierarchy of directories and files.
As a result, billions of objects can be stored in a single namespace.
22The object ID allows easy access to objects without the need to specify their
storage locations.
Unified Storage
Objectives
4 1 2 3
1: A backup application is software that creates a copy of the data so that the
backup copy can be used to restore the original data in the event of data loss or
corruption.
2:
• Identifies and moves inactive data out of primary storage systems into lower
cost storage systems, called data archives, for long term retention.
• Creates a stub file26 on the primary storage after moving the original data to
archive storage.
• Performs retrieval of archived data when required by the client.
• Creates index27 archived data to facilitate user searches and data retrieval.
25
Typically a technology or a system upgrade requires the existing data to be
moved to a new system before withdrawing the old system to avoid downtime.
27By utilizing the index, users may also search and retrieve their data with the web
search tool.
Application Server
Primary
Storage
Device Tape
Data is copied to
Protection Storage Disk
Protection Storage
28Typically organizations have protection storage at the remote data center for DR
purpose.
Disk density has increased dramatically over the past few years, lowering the cost
per gigabyte to the point where disk is a viable protection storage option for
organizations.
• SAN-attached Storage
• Network-attached Storage (NAS)
• Object-based Storage
• Cloud-based Storage
• Supports replicate data to a remote site to help an organization comply with off-
site requirements.
− This avoids the need to ship tapes from the primary site to the remote site
and thus reduces the risk of losing tapes in transit.
• Includes features such as data deduplication, compression, and encryption to
support various business objectives.
1 3
1: A tape library contains one or more tape drives that record and retrieve data on
magnetic tape.
3: Robotic arms are used to move tapes around the library such as moving a tape
drive into a slot.
4: Used to add or remove tapes from the library without opening the access doors
because opening the access doors causes a library to go offline.
A tape library is a tape-based protection storage system that has tape drives and
tape cartridges, along with a robotic arm or picker mechanism as shown in the
image.
Click on each mechanism in the Tape Library image for more information.
• A low-cost, portable solution and can be used for long-term, off-site storage29.
• Must be stored in locations with a controlled environment to ensure preservation
of media and prevention of data corruption.
• Highly susceptible to wear and tear and may have a short shelf life.
• Traditional backup process using tapes is not optimized to recognize duplicate
content.
• Storing and retrieving the data takes more time with tape.
• Data integrity and recoverability are also major issues with tape-based media.
Emulation Engine
Storage
Objectives
31 Integrity ensures that unauthorized changes to data are not allowed. Also ensure
to detect and protect against unauthorized alteration or deletion of data.
32Availability ensures that authorized users have reliable and timely access to data
and services.
Data is one of the most important assets for any organization. Other assets include
hardware, software, and other infrastructure components required to access and
protect data.
Threat Agent
Gives rise to
Threat
Wish to abuse and/or
That exploits
may damage
Vulnerabilities
Leading to
Risk Control Owner
To Impose
To reduce s
Asset
Values
Data is the currency of the internet economy and a critical asset that must be
protected, kept confidential and made available at a moment’s notice. Global
business relies on the constant flow of data across interconnected networks, and
digital transformation means an increase of sensitive data. This presents ample
opportunity for cyber threats and exposure to leverage data for ransom, corporate
espionage or even cyber warfare. Ransomware is a malware method that:
• Encodes the targeted system or files. To decode the system or files the hacker
demands for some ransom mostly in form of cryptocurrency.
• Spreads through phishing emails that contain malicious attachments or through
drive-by downloading33.
To protect your dynamically huge data from cyber-attacks requires proven and
modern solutions. Here are some components for proven and modern solution:
Click the tabs below for more information about proven and modern solutions.
Create unchangeable data copies in a secure digital vault and processes that
create an operational air gap between the production / backup environment and the
vault.
Machine learning and full-content indexing with powerful analytics within the safety
of the vault. Automated integrity checks to determine whether data has been
impacted by malware and tools to support remediation if needed.
Workflows and tools to perform recovery after an incident using dynamic restore
processes and your existing data recovery procedures.
33
Occurs when an end-user by mistake visits an infected website and then
malware is downloaded and installed without the user’s information.
Expert guidance to select critical data sets, applications and other vital assets to
determine RTOs and RPOs and streamline recovery.
Data protection management functions that are necessary for the visibility and
control of data source and protection components, and data protection operations.
Each of these functions are:
Discovery
Protection
Protection Configuration ....
Storage
Applications Performance ....
Operations Management
Orchestration
Data protection services are provisioned to meet the availability and data protection
requirements of business applications and IT services.
Concepts in Practice
Concepts in Practice
Click the right and left arrows to view all the concepts in practice.
Dell PowerEdge Server family includes various types of servers that include Tower
servers36, Rack servers37, and Modular Servers38.
36Tower servers generally contain more disk drives bays and expansion card slots
than other server form factors. The advantages of a tower server lie in its compact
shape. Tower servers can be used in work areas which are not designed to contain
servers. Its simplicity and robustness make the tower server an ideal place for a
small company to begin using a server.
37A rack server is also called a rack-mounted server. Rack-mount servers are
designed to save space when there are several servers in a confined area. Rack
servers are generally more expensive. They are better suited to medium-sized
businesses or micro-businesses.
38Modular servers are the latest development in the history of the different server
types. Also defined as a server that is hosted with a dedicated chassis, including
network and storage components.
• Dell EMC Unity XT All-Flash and Hybrid Flash arrays set new standards for
storage with compelling simplicity, all-inclusive software, blazing speed,
optimized efficiency, multi-cloud enablement.
− All in a modern NVMe-ready design – to meet the needs of resource-
constrained IT professionals in large or small companies.
• Designed for performance, optimized for efficiency and built for hybrid cloud
environments.
• These systems are the perfect fit for supporting demanding virtualized
applications, deploying unified storage and addressing Remote-Office-Branch-
Office requirements.
• Dell EMC PowerScale is the next evolution of the OneFS – the operating
system powering the industry’s leading scale-out NAS platform.
− The software-defined architecture of OneFS give simplicity at scale,
intelligent insights and the ability to have any data anywhere it needs to be.
− Whether it is hosting file shares or home directories or delivering high
performance data access for applications like analytics, video rendering and
life sciences, PowerScale can seamlessly scale performance, capacity and
efficiency to handle any unstructured data workload.
• PowerScale brings a new version of OneFS to our Isilon nodes as well as two
new all-flash PowerScale nodes, that delivers application requirements like S3
protocol and performance needs like NVMe, from the edge to the cloud.
• The new PowerScale all flash platforms co-exist seamlessly in the same cluster
with existing Isilon nodes to drive your traditional and modern applications.
• Dell EMC ECS is the leading object storage platform boasts unmatched
scalability, performance, resilience, and economics.
• Dell EMC ECS has been purpose-built to store unstructured data at public cloud
scale with the reliability and control of a private cloud.
• Capable of scaling to exabytes and beyond, ECS empowers organizations to
manage a globally distributed storage infrastructure under a single global
namespace with anywhere access to content.
• Deployable as a turnkey appliance or in a software-defined model, ECS delivers
rich S3-compatibility on a globally distributed architecture, empowering
organizations to support enterprise workloads such as:
− Cloud-native, archive, IoT, AI, and big data analytics applications at scale.
• Modern data centers require storage that can consistently meet heavy
performance demands, provide un-paralleled support to manage the entire
virtual workload lifecycle, and deliver demanding SLAs.
VMware vSphere
− Improve performance and scale for Monster VMs to support your large scale
up environments.
− Scale up to 24TB memory and support up to 768 vCPUs through Monster
VMs.
• Compute systems host financial, email, and backup applications, and the
organization’s website.
− Three file servers for Windows users and the remaining three file servers for
UNIX users.
• Email application uses a SAN-attached (block-based) storage system as
primary storage.
2. Organization’s Challenges:
• UNIX users and Microsoft Windows users are unable to share files.
• Some of the file servers are overly utilized and therefore new file servers
must be deployed.
3. Organization’s Requirements:
4. Expected Deliverables:
Solution
Objectives
Impact of Fault
− Compute
− Network
− Storage
− Application
Reliability
Availability
• Improved by ensuring that the IT systems and services can perform their
required functions during their operating time.
• Dependent on the reliability of systems on which the services are created.
The ability of a system to continue functioning in the event of a fault within or failure
of some of its components.
• Hardware failure
• Software bugs
• Administrator/user errors
Transient Unavailability
• Example: An online transaction times out but works fine when a user retries the
operation.
Intermittent Unavailability
Permanent Unavailability
Elimination of SPOF
Clustered Compute
Systems
Redundant
Links Redundant Storage
Systems
Redundant
LAN/WAN
NICs
Redundant
HBAs
Redundant
SAN
Switches
Redundant
Networks
Clients
Redundant
Redundant Remote Site
Links
NICs
Fault Isolation
HBAs Storage
Port
=Points of Fault
• Limits the scope of a fault into local area so that the other areas of a system is
not impacted by the fault.
• Does not prevent failure of a component but ensures that the failure does not
impact the overall system.
• Requires a fault detection mechanism that identifies the location of a fault and a
contained system design (like sandbox) that prevents a faulty system
component from impacting other components.
Fault Recovery
Restores a system to the desired operating level after a fault has occurred in the
system.
Forward recovery
• Involves correcting the fault in a system to continue system operations from the
faulty state. It is useful only when the cause and the impact of a fault is
understood.
− Example, consider a group of two mirrored disk drives that store same data.
Each write I/O is written to both the disk drives. If one of the drives in the
mirrored pair fails and is replaced by a new drive, the surviving drive in the
mirrored pair will be used for data recovery and continuous operation.
Therefore, I/O operations can be continued from the fault condition.
Backward recovery
− For example, the memory state, settings state, and power state (on, off, or
suspended) of a virtual machine (VM) is saved at a specific recovery point
so that the VM can be restored to its previous state if anything goes wrong.
1. Which of the following are types of fault recovery? Choose all that apply.
a. Complete functional recovery
b. Functional recovery using an alternative logic
c. Degraded functional recovery
d. Backwards recovery
Objectives
− Compute clustering
− Virtual machine (VM) live shadow copy
− Link aggregation
− NIC teaming
− Switch aggregation
− Multipathing
− Configuring hot-swappable components
Compute Clustering
3
1
2 4
If the primary VM fails due to hardware failure, the technique enables failover to the
secondary VM immediately. After the failover occurs, a new secondary VM is
created and redundancy is reestablished.
Link Aggregation
1: Combines two or more parallel interswitch links (ISLs) into a single logical ISL,
called a link aggregation group
Optimizes network performance by distributing network traffic across the shared
bandwidth of all the ISLs in a link aggregation group
Enables network traffic failover in the event of a link failure. If a link in a link
aggregation group is lost, all network traffic on that link is redistributed across the
remaining links.
• Combines multiple ISLs into a single logical ISL (link aggregation group)
• Distributes network traffic over ISLs, ensuring even ISL utilization.
• Enables network traffic failover in the event of a link failure.
• Provides higher throughput than a single ISL could provide.
NIC Teaming
Switch Aggregation
Multipathing
I/O
Hot Swappable
Controller Blade
2. Which of the following statements are correct? Select all correct options.
a. Clustering enables service failover from a failed server to an active server.
b. VM live shadow copy balances client’s traffic across primary and
secondary VMs.
c. Switch aggregation creates a group of active and passive switches.
d. Link aggregation combines multiple logical ISLs to create a single physical
ISL .
e. Hot-swappable components can be replaced while a system remains
available.
Storage
Storage
Objectives
1: Key Functions:
Managing drive aggregations
Translation of I/O requests between logical and physical drives
Data regeneration in the event of drive failures
2: A logical unit that consists of multiple drives where the data is written in blocks
across the drives in the pool
• RAID is a technique that combines multiple disk drives into a logical unit and
provides protection, performance, or both.
Node Cluster
− A set of n disks is divided into m disks to hold data and k disks to hold
coding information.
− Coding information is calculated from data.
To understand the illustration, click here.
• Refers to a spare disk drive that replaces a failed drive by taking the identity of
the failed drive.
• With the hot spare, one of the following methods of data recovery is performed
depending on the RAID implementation:
− If parity RAID is used, the data is rebuilt onto the hot spare from the parity
and the data on the surviving disk drives in the RAID set.
− If mirroring is used, the data from the surviving mirror is used to copy the
data onto the hot spare.
• When a new disk drive is added to the system, data from the hot spare is
copied to it. The hot spare returns to its idle state, ready to replace the next
failed drive.
• A hot spare should be large enough to accommodate data from a failed drive.
Some systems implement multiple hot spares to improve data availability.
1: Each write is mirrored and stored in two independent cache memory cards
2: Even if one cache fails, the data is still available in the mirrored cache
Cache is a volatile memory. So, a power failure or any kind of cache failure will
cause loss of data that is not yet committed to the storage drive. This risk of losing
uncommitted data held in cache can be mitigated using cache mirroring and cache
vaulting.
Mirroring:
After power is restored, the data from the drive is written back to cache
Objectives
Graceful Degradation
1: In case of failure, when a module is down and when client is accessing the
application, the application is still available to the client with degraded functionality
and performance.
1: When failure happens, the retry logic sends a second request and the service
becomes available then.
1: State information can be accessed by the new server from the repository
• State information is stored out of the memory and stored in a data repository.
• If an instance fails, the state information will still be available in the repository.
Database Rollback
Checkpointing
1: In the event of a zone outage, services can fail over to another zone.
Availability zones, although isolated from each other, are connected through low-
latency network links.
• Availability zone is a location with its own set of resources and isolated from
other zones.
• Availability zones, although isolated from each other, are connected through
low-latency network links.
• In the event of a zone outage, services can fail over to another zone.
Concepts in Practice
Concepts in Practice
Dell EMC PowerPath is a host-based software that provides automated data path
management and load-balancing capabilities for heterogeneous server, network,
and storage deployed in physical and virtual environments. The PowerPath family
includes PowerPath Multipathing for physical environments, as well as Linux, AIX,
and Solaris virtual environments and PowerPath/VE Multipathing for VMware
vSphere and Microsoft Hyper-V virtual environments. It automates multipathing
policies and load balancing to provide predictable and consistent application
availability and performance across physical and virtual environment. PowerPath
improves service-level agreements by eliminating application impact from I/O
failures.
VMware vSphere HA
VMware vSphere FT
functioning ESXi host seamlessly becomes the Primary VM host without losing
network connections or in-progress transactions. With transparent failover, there is
no data loss and network connections are maintained. After a transparent failover
occurs, a new Secondary VM is respawned and redundancy is re-established. The
entire process is transparent and fully automated and occurs even if vCenter
Server is unavailable.
Exercise
2. Organization Challenges:
− Customers were unable to view product catalog, shopping cart, and order
status.
• Recently a power supply failure caused an entire zone outage and loss of in-
progress transactional data.
3. Organization requirements:
4. Expected Deliverables:
Solution
Data Backup
Data Backup
Data Backup
Data Backup
Introduction to Backup
Introduction to Backup
Objectives
For more details about need for data backup, click here
Backup Architecture
Component Role
Backup Server • Manages the backup operations and maintains the backup
catalog.
• Contains information about the backup configuration39 and
backup metadata40.
Storage Node • Responsible for organizing the client’s data and writing the
data to a backup device.
• Controls one or more backup devices41.
• Sends the tracking information about the data written to
the backup device to the backup server.
• Reads data from the backup device during recoveries .
• A wide range of backup targets are currently available such as tape, disk, and
virtual tape library.
39 The backup configuration contains information about when to run backups, which
client data to be backed up, and more.
Backup Operations
Backup Description
Operation
Backup-Type • File-level
One or more files are backed up on a client system.
• Block-level
Backup data at block-level instead of file-level.
• Image-level
Recovery Operations
After the data is backed up, it can be restored42 when required. A recovery
operation restores data to its original state at a specific PIT. Typically, backup
applications support restoring one or more individual files, directories, or VMs.
42 A restore process can be manually initiated from the client. It can also be initiated
from the server interface.
Types of Recovery
The various types of recoveries are data recovery, disaster recovery, bare metal
recovery and cloud disaster recovery.
Types of Description
Recovery
Operational Restores small numbers of files after they have been accidentally
Recovery deleted or corrupted.
or restore
Consistency43 is critical to ensure that a backup can restore a file, directory, file
system, or database to a specific point-in-time.
Offline Online
Backup Granularities
Backup granularity depends on business needs and the required RTO/RPO. Based
on the granularity, backups can be categorized as full, incremental, cumulative (or
differential), incremental forever, and synthetic full backup.
Most organizations use a combination of these backup types to meet their backup
and recovery requirements. Let us understand each of them in detail:
Full Backup
Full Backup-Restore
In the motion graphics shown below, a full backup is created on every Sunday.
When there is a data loss in the production on Monday, the recent full backup that
is created on the previous Sunday is used to restore the data in the production.
Incremental Backup
Incremental backup copies the data that has changed since the last backup.
• The main advantage of incremental backups is that less files are backed up
daily, allowing for shorter backup windows.
Cumulative Backup
Cumulative (differential) backup copies the data that has changed since the last full
backup.
48For example, the administrator created a full backup on Sunday and differential
backups for the rest of the week. Monday’s backup would contain all of the data
that has changed since Sunday. It would therefore be identical to an incremental
backup at this point. On Tuesday, however, the differential backup would backup
any data that had changed since Sunday (full backup).
Rather than scheduling periodic full backups, this backup solution requires only one
initial full backup.
49Also reduces the data growth because all incremental backups contain only the
blocks that have changed since the previous backup.
• Created from an existing full backup and is merged with the data from any
existing incremental backups.
• This backup is not created directly from production data.
Backup Multiplexing
One of the ways that backup software achieves backup efficiency with tapes is by
interleaving or multiplexing multiple backups onto a backup device. Multiplexing
allow:
Some of the backup software provides the ability to further manage and protect the
backup data through the use of cloning50 and staging51.
Task Description
50 Cloning improves data protection through redundancy, since each backup has a
clone at a geographically-dispersed location. Some backup software supports the
capability of performing copy operation at the same time as backups.
51Staging also allows data to be moved off the device outside the backup period,
ensuring that sufficient disk space is available for the next backup session.
1. From the list of steps provided - drag and drop each into the correct sequence
to perform a backup operation.
Backup Topologies
Backup Topologies
Objectives
Direct-Attached Backup
Metadata
Data
In the image shown, the client acts as a storage node that writes data on the
backup device.
− Backup device is not shared, which may lead to silos of backup device in the
environment.
− In a large data center environment, backup devices may be underutilized.
LAN-based Backup
Storage Node
Data
Metadata
Backup Device
Backup Server
• Advantage:
− Centralized backups reduce management complexity.
• Disadvantage:
SAN-based Backup
Storage Node
Metadata
Data
The SAN-based backup52 (as shown in the image) is also known as the LAN-free
backup. The high-speed and extended distance capabilities of Fiber Channel are
used for the backup data movement path.
• Advantage:
− Production LAN environment is not impacted.
− Backup device can be shared among the clients.
− Offers improved backup and restore performance due to FC SAN.
• Disadvantage:
− Impacts the application’s performance.
• In the shown image, clients read the data from the application servers in the
SAN and write to the SAN-attached backup device.
− The backup data traffic is restricted to the SAN and the backup metadata is
transported over the LAN.
− However, the volume of metadata is insignificant when compared to
production data.
52 The SAN-based backup topology is the most appropriate solution when a backup
device needs to be shared among the clients. In this case the backup device and
clients are attached to the SAN.
NAS-based Backup
Storage System
Metadata
Data
Backup Request
Backup Server
− In this approach, the NAS head retrieves data from storage over the network
and transfers it to the backup client running on the application server.
− The backup client sends this data to a storage node, which in turn writes the
data to the backup device.
Cloud-based Backup
Backup Client
Cloud
Backup Data from Cloud
Backup Client
• Monitor the health of the data protection environment and comply with
government and industry regulations.
• Manages the data backup with robust on-site, off-site and hybrid cloud–based
security.
• Advantages:
Backup Methods
Backup Methods
Objectives
Backup Agent
Backup Device
Backup Data
Application Servers
Backup Server/
Storage Node
• Advantage:
− Backup configurations and recovery options follow traditional methods that
administrators are already familiar with, so there are no added configuration
requirements.
− Supports a single file backup and restore.
• Disadvantage:
Image-Based Backup
Create Snapshot
VM Management Server
VM Snapshot
Proxy Server
Create
Snapshot Mount Snapshot
Backup Device
Backup Server
Image-level backup53 (as shown in the image) makes a copy of the virtual machine
disk and configuration associated with a particular VM. The backup is saved as a
single entity called VM image.
Application Server
VM Kernel creates an
additional file where it stores a
VM Kernel map of all the VM disk’s blocks
Virtual Machine FS
Volume
Backup Device
To further enhance the image-based backup some of the vendors support changed
block tracking54 mechanism.
For more details about the changed block tracking mechanism, click here.
54This feature identifies and tags any blocks that have changed since the last VM
snapshot. This enables the backup application to backup only the blocks that have
changed, rather than backing up every block.
VM Disk Files
FS Volume
(Production)
NDMP-Based Backup
Storage
Application Server
Backup Device
Backup Data
NDMP Client
Instruct NAS to Start Backup Backup Data
NAS Device
Backup Server
NDMP Server Running on
NAS Head
The key components of an NDMP infrastructure are NDMP client55 and NDMP
server.
• The NDMP server has two components- data server56 and media server57.
• Backup server uses NDMP client and instructs the NAS head to start the
backup.
• The NAS head uses its data server to read the data from the storage.
• The NAS head then uses its media server to send the data read by the data
server to the backup device.
Direct primary storage backup approach backs up data directly from a primary
storage system to a backup target without any additional backup software.
This data protection solution integrates primary storage and protection storage
(backup device).
56The data server is a component on a NAS system that has access to the file
systems containing the data to be backed up.
57The media server is a component on a NAS system that has access to the
backup device.
Concepts in Practice
Concepts in Practice
Click the right and left arrows to view all the concepts in practice.
Dell EMC NetWorker is a backup and recovery solution for mission-critical business
applications in physical and virtual environments for on-premises and cloud.
• Unified backup and recovery software for the enterprise: deduplication, backup
to disk and tape, snapshots, replication and NAS.
• Provides a robust cloud capability enabling long term retention to the cloud,
backup to the cloud and backup in the cloud.
• NetWorker Module for Databases and Applications (NMDA) provides a data
protection solution for DB2, Informix, Lotus Domino/Notes, MySQL, Oracle,
SAP IQ, and Sybase ASE data.
− NMDA also provides data protection for MongoDB, MySQL, and
PostgreSQL data through the Orchestrated Application Protection feature.
• NetWorker Snapshot Management (NSM) is integrated with Dell EMC storage
and enables end-to-end snapshot management and backup from within the
NetWorker UI.
• The next generation of Dell EMC Data Domain appliances, that are now setting
the bar for data management from edge to core to cloud.
• Integrates easily with existing infrastructures, enabling ease-of-use with leading
backup and archiving applications, and offers superior performance in
conjunction with PowerProtect Data Manager and Data Protection Suite.
• Natively tier deduplicated data to any supported cloud environment for long-
term retention with Dell EMC Cloud Tier.
• Provides fast disaster recovery with orchestrated DR and provides an efficient
architecture to extend on-premises data protection.
• The organization:
− Currently uses tape as their primary backup storage media for backing up
application data.
− Uses an agent-based backup solution for backing up data.
− Has a file-sharing environment in which multiple NAS systems serve all the
clients including application servers.
2. Organization’s Challenges:
3. Organization’s Requirements:
• Need to offload the backup workload from the compute system to avoid
performance impact on applications.
4. Expected Deliverables:
Solution
− In NDMP-based backup, data is sent directly from the NAS head to the
backup device without impacting the application servers.
Data Deduplication
Data Deduplication
Data Deduplication
Data Deduplication
Objectives
Backup device
consists of a series of daily incremental backups and weekly full backups. Daily
backups are usually retained for a few weeks and weekly full backups are retained
for several months. Due to this process, multiple copies of identical or slowly-
changing data are retained on backup media, leading to a high level of data
redundancy.
Data deduplication is the process of detecting and identifying the unique data
segments (chunk) within a given set of data to eliminate redundancy. Only one
copy of the data is stored; the subsequent copies are replaced with a pointer to the
original data. Deduplication addresses all the aforesaid challenges.
Deduplication Ratio
Retention period The longer the data retention period, the greater is the
chance of identical data existence in the backup.
Frequency of full The more frequently the full backups are conducted,
backup the greater is the advantage of deduplication.
Data type The more unique the data, the less intrinsic
duplication exists.
Click the icons shown below for information on the benefits of deduplication.
1 2 3 4
2: As data deduplication reduces the amount of content in the daily backup, users
can extend their retention policies. This can have a significant benefit to users who
currently require longer retention.
Data deduplication is easily performed inline with the help of software and general-
purpose CPUs, which are connected to the PowerProtect DD controller to transfer
deduplicated data to a backup appliance. Application Agents are installed on
application or database host servers to manage protection using PowerProtect
Data Manager. These Agents are commonly known as Data Domain Boost
Enterprise Agents (DDBEA) for databases and applications.
DD Boost is a software that improves the interactions of backup servers and clients
with a Data Domain backup appliance. The DD Boost makes the data deduplication
process distributed, so that there is faster data throughput and reduced server CPU
utilization. The File System agent allows an application administrator to protect and
recover data on the file system host. PowerProtect Data Manager integrates with
the File System agent to check and monitor backup compliance against protection
policies. PowerProtect Data Manager also allows central scheduling for backups.
If the system doesn't have DD Boost, the Data Domain performs inline data
deduplication. As and when the files and data are sent over the network, the DD
deduplicates the data using RAM and CPU, writing only the unique data chunks to
the backup target.
With DD Boost, a considerable portion of the deduplication can occur before the
data is sent across to the Data Domain. The backup source takes the data,
segments it out, compares it with segments already on the Data Domain, and only
sends over new, unique segments.
Step 1
Client agent checks the file system and determines if a file has been backed up
before.
Modified files
Step 2
Step 3
Hashes are compared with chunks already existing on the Data Domain.
Step 4
Only new and unique data chunks are backed up on the Data Domain.
Objectives
Deduplication Granularity
File-level Deduplication
Client 1 Client 2
File-level
Deduplication
Backup
device
Fixed-length
Deduplication
Client
Backup
The changed segment is now considered as a
unique data, so it is backed up. Remaining
unchanged segment is considered as a duplicate
data and not backed up.
Backup device
• Breaks files down to smaller segments and fixes the chunking at a specific size,
for example 8 KB or maybe 64 KB.
• Detects redundant data within and across files.
Fixed-length block may miss many opportunities to discover the redundant data
because the block boundary of similar data may be different.
Variable-length
Client Deduplication
Backup
• The length of the segments vary and provide greater storage efficiency for
redundant data regardless of where new data has been inserted.
• If there is a change in the block, then the boundary for that block is only
adjusted, leaving the remaining blocks unchanged.
• It yields a greater granularity in identifying duplicate data.
Object-level Deduplication
This object ID is unique to ensure that only one protected copy of the content is
stored (single instance storage), no matter how many times clients store the same
information. This significantly reduces the total number of data stored, and is a key
factor in lowering the cost of storing and managing content.
At write time, the object-based storage system is polled to see if it already has an
object with the same signature. If the object is already on the system, it is not
stored, rather only a pointer to that object is created.
Deduplication Methods
Source-based Deduplication
In this deduplication method, the data is deduplicated at the source (backup client).
The backup client sends only new, unique segments across the network. This is
suitable for environment where storage and network is a constraint. However, it
may require a change in the backup software if this option is not supported by the
existing backup software. Source-based deduplication consumes CPU cycles on
client and may impact the application performance. So, it is recommended for
remote office branch office environment for performing centralized backup.
Target-based Deduplication
Typically, organizations maintain a copy of data at the remote site (DR site or
cloud) for DR purpose. If the primary site goes down due to disaster or any other
reasons, the data at the remote site will enable restoring of services and data to the
primary site. Data deduplication can enhance DR because of the following reasons:
Deduplicated Deduplicated
data data
Concepts in Practice
Concepts in Practice
1. Present Scenario:
• Currently uses disk as their backup storage media for backing up application
data.
• Uses tapes for protecting data at the remote site for DR purpose.
2. Organization’s Challenges:
• Sending tapes to offsite locations would increase the risk of losing sensitive
data.
3. Organization’s Requirements:
4. Expected Deliverables:
Solution
Replication
Replication
Replication
Replication
Objectives
Data Replication
Replica
Data Center B
Servers
Connectivity
Storage
Data Replication to
Data Replication Cloud
Data Center A
A data replication solution is one of the key data protection solution that:
Replica
Replication
Can be used to restart business operations or to
recover the data
Replication
Data Used for running decision support activities
Source
Replication
Data migration
58 Enables restoration of data to the source if there is a data loss at the source.
59 Enables to restart the business operations on it, if the source is not available due
to some reasons.
Offline Online
Types of Replication
Replication can be classified into two major categories: local and remote
replication.
Local Replication
Local Replication
Objectives
Storage System
LUN A
LUN B
Local Replication
− Enables one to perform operational recovery in the event of data loss and
also provides the support for other business operations such as backup.
File system (FS) snapshot creates a copy of a file system at a specific point-in-time,
even when the original file system continues to be updated and used normally.
− To read from the snapshot FS, the bitmap is consulted. If the bit is 0, then
the read will be directed to the production FS.
− If the bit is 0, then the read will be directed to the production FS.
− If the bit is 1, then the block address will be obtained from the blockmap and
the data will be read from that address on the snapshot FS.
FS Snapshot 3
Wednesday View
Tuesday View
Monday View
FS Snapshot 2
FS Snapshot 1
VM Snapshot
VM Virtual Disk
Storage
60The state includes the VM’s power state (for example, powered-on, powered-off,
or suspended).
61 The data includes all of the files that make up the VM.
VM Clone
VM Copy of VM
Full
Configuration Configuration
Clone
VM Disk Copy of VM
Disk
Full volume local replication provides the ability to create fully populated point-in-
time copies of LUNs within a storage system.
62 Synchronization is the process of copying data from the source LUN to the clone.
Replication Session
Save
Location
Source Snapshot
• CDP provides the capability to restore data and VMs to any previous point-in-
time (PIT).
1: CDP provides continuous replication, tracks all the changes to the production
volumes that enable to recover to any point-in-time.
3: CDP supports both local and remote replication of data and VMs to meet
operational and disaster recovery respectively.
5: CDP supports multi-site replication, where the data can be replicated to more
than two sites using synchronous and asynchronous replication.
In this method, before the start of replication, the replica is synchronized with the
source and then the replication process starts. After the replication starts:
Write I/O
Write
Splitte
CDP
r
Appliance
Write I/O
Data is written to the journal volume
along with its timestamp
Data is written
to replica
Storage System
• All the writes to the source are split into two copies.
− One of the copies is sent to the CDP appliance and the other to the
production volume.
• CDP appliance writes the data to the journal volume.
• Data from the journal volume is sent to the replica at predefined intervals.
• While recovering data to the source, the appliance restores data from the
replica and applies journal entries up to the point-in-time chosen for recovery.
Virtual Appliance
Write Splitter
VM Disk VM Disk
Files Files
Source Local
Journal
Volume Replica
Hypervisor-based CDP –
Local Replication
Some vendors offer continuous data protection for VMs through hypervisor-based
CDP implementation. This deployment option:
Remote Replication
Remote Replication
Objectives
VM Replication
− One of the storage systems is in the source site and the other system is in
the remote site for DR purpose. Data can be transmitted from the source
storage system to the target system over a shared or a dedicated network.
Remote Replication
Synchronous
• Writes must be committed to the source and the target prior to acknowledging
“write complete” to the production compute system.
− Provides near zero RPO.
• The shown image, illustrates an example of synchronous remote replication. If
the source site is unavailable due to disaster, then the service can be restarted
immediately in the remote site to meet the required SLA.
Asynchronous
− Data is buffered at the source and sent to the remote site periodically.
− Replica will be behind the source by a finite amount (finite RPO).
To learn more about asynchronous remote replication, click here.
Multi-site Replication
Source Replica
Synchronous
Replication
Asynchronous with
Differential
Resynchronization
Asynchronous
Replication
Replica
Storage System at
Remote Site 2
In a two-site synchronous replication, the source and the target sites are usually
within a short distance.
• In synchronous replication, if a regional disaster occurs, both the source and the
target sites might become unavailable.
• In asynchronous replication, if the source site fails, production can be shifted to
the target site, but there will be no further remote protection of data until the
failure is resolved.
• Data from source site is replicated to multiple remote sites for DR purpose.
− Disaster recovery protection is always available if any one-site failure
occurs.
• Mitigates the risk in two-site replication.
Write
Splitter
Remote CDP
Appliance
5. Data is copied to the remote
replica
Local CDP
Appliance
In this method, the replica is synchronized with the source, and then the replication
process starts. After the replication starts:
• All the writes from the host to the source are split into two copies.
− Write splitter creates a copy of a write data and sends it to the CDP
appliance and production volume.
• Data is sequenced, compressed, and replicated to the remote appliance.
• Data is received, uncompressed, and sequenced.
• Data is written to the journal.
• Data is copied to the remote replica.
For more information about remote replication CDP operation, click here.
Concepts in Practice
Concepts in Practice
SnapVX
TimeFinder SnapVX is a local replication solution with cloud scalable snaps and
clones to protect your data. SnapVX solution:
• Provide space-efficient local snapshots that can be used for localized protection
and recovery and other use cases including development/test, analytics,
backups, and patching.
• Secure snapshots prevent accidental or malicious deletion, securing them for a
specified retention period.
SRDF
SRDF is Dell EMC’s Remote Replication technology that enables the remote
mirroring of a data center with minimal impact to the performance of the production
application. SRDF replication products:
• Provides disaster recovery and data mobility solutions for the PowerMax and
VMAX Family storage arrays.
• Copies process between the sites is accomplished independently without the
host.
− There are no limits to the distance between the source and the target
copies.
• Enables storage systems to be in the same room, different buildings, or
hundreds to thousands of kilometers apart.
• Offers the ability to maintain multiple, host-independent, remotely mirrored
copies of data.
For detailed information about key PowerMax and VMAX Family Remote
Replication options, click here.
RecoverPoint
Exercise- Replication
Exercise- Replication
1. Present Scenario:
• Every month-end the bank runs billing and reporting applications to generate
bills and statement of customer’s account.
• The bank has two data centers which are 100 miles apart.
2. Organization’s Challenges:
• The backup window is too long and is negatively impacting the application
performance.
• These billing and reporting applications have huge impact on the source
volume.
• In the past year, the top management has become extremely concerned
about DR because they do not have any DR plans in place.
3. Organization’s Requirements:
• During billing and reporting, the source volume should not have any impact.
• Bank cannot afford any data loss; therefore, needs a disaster recovery
solution with near zero RPO.
4. Expected Deliverables:
Solution
Data Archiving
Data Archiving
Data Archiving
Data Archiving
Objectives
Fixed Data
To learn more about the need for data archiving, click here.
Data archiving moves fixed data63 that is no longer actively accessed to a separate
low-cost archive storage system for long term retention and future reference:
Archiving Architecture
Application Server
Archive Server (Policy Primary Storage
Engine)
File Server
Archiving
Archive Storage Agent
Clients
• Archiving agent- It is software installed on the application and file servers. The
agent is responsible for scanning the files and archiving them, based on the
policy defined on the archive server (policy engine).
• Archive server- It is software installed on a server that enables administrators
to configure policies for archiving data. Organizations set their own policies for
qualifying data to be moved into archive storage. Policies can be defined based
on file size, file type, or creation/modification/access time. Once the files are
identified for archiving, the archive server creates an index for the files. By
utilizing the index, users may search and retrieve their data.
• Archive storage- It stores the fixed data.
Sarbanes-Oxley Act
• Passed in 2002 and it protects the shareholders and the general public from
accounting errors and fraudulent practices in the enterprise.
• Created to protect investors by improving the accuracy and reliability of
corporate disclosures.
• Applies to all public companies and accounting firms.
• Not a set of business practices and does not specify how a business should
store records.
• Defines which records are to be retained and for how long.
1. Which of the following statements are correct? Choose all that apply
a. Archiving fixed data before taking backup reduces the backup window
b. Primary objectives of archiving are compliance adherence and lower cost
c. Nearline archive makes the data immediately accessible
d. Data archiving must occur outside the application operating time
e. Archiving agent indexes and moves fixed data to high-performance storage
Objectives
Primary Storage
Stub File
Communication
Archive Server
Clients
Archive Server
• Archiving agent scans primary storage to find files that meet the archiving
policy. The archive server indexes the files.
• Once the files have been indexed, they are moved to archive storage and small
stub files are left on the primary storage.
When a client attempts to access the By utilizing the index for archived files,
files through an application or file server, users may also search and retrieve
the stub file is used to retrieve the file files. The retrieval of files from the
from archive storage. archive storage is transparent to the
clients.
Tier 1
Performance Primary
Tiers Storage
Primary Tier 2
Storage
Archive
Tier Tier 3
Archive
Storage
To learn more about Correlating Storage Tiering and Archive, click here.
Tier 2
Tier 3
Data Movement
Example: If a policy states “move the files from tier 2 to tier 3 storage that are not
accessed for the last six months,” then all the files in tier 2 storage that match this
condition are moved to tier 3 storage. Multiple rules may also be combined to
create a policy as shown in the image.
The image illustrates an example of file-level storage tiering, where files are moved
from a NAS device (primary storage system) to an archive storage system. The
environment includes a policy engine, where tiering policies are configured.
Government Compliance
• Meets all requirements to produce emails from every individual involved in stock
sales or transfers.
Legal Dispute
Client
CAS
Application
Server
CAS API
Feature Description
Content integrity Provides assurance that the stored data has not
been altered.
Feature Description
To get a little more detail about the key features of CAS, click here.
Primary
Storage
Archive
Servers
Clients
Archive Storage
Concepts in Practice
Concepts in Practice
Dell EMC Cloud Tier provides a solution for long-term retention. Using advanced
deduplication technology that significantly reduces storage footprints, and with Dell
EMC Cloud Tier (Cloud Tier), DDOS (DD Operating System) can natively tier data
to a public, private or hybrid cloud for long-term retention. Only unique data is sent
to the cloud and data lands on the cloud object storage already deduplicated.
Cloud Tier supports a broad ecosystem of backup and enterprise applications and
a variety of public and private clouds. Cloud Tier enables:
• VMs host health care, email, and backup applications; and file servers.
• Physical compute systems are also connected to a tape library that is used
as backup storage system.
• One of the storage systems has mostly SSDs while another has only HDDs.
• Old records are needed when patients revisit the health care organization.
• Each backup copy is retained in the tape library for one month and then the
tapes are moved and maintained in a vault.
2. Organization Challenges:
• Storage systems have only 10% storage capacity available for storing new
data.
• Last year, some of the old records were altered resulting in a delay in
treatment.
− Old records were retrieved by bringing the old tapes from the vault and
making them online.
• A long backup window impacts application performance during peak hours.
• Maintaining a large number of tapes poses risks of labeling errors and lost
tapes.
3. Organization Requirements:
• Need to ensure that the old records are authentic and are not altered.
• Need to reduce the backup window and the associated costs and risks.
4. Expected Deliverables:
Solution
Data Migration
Data Migration
Data Migration
Data Migration
Data Migration
Organizations deploys data migration solutions for the following reasons. Click
each sub-heading for more information.
2. Disaster avoidance
3. Technology refresh
- VM live migration
- VM storage migration
Compute
Data moves (push) from Compute
system
the old system to the new system
system
Remote
device
Data can be moved from or to devices in the control storage system to or from a
remote storage system. Data migration solutions perform push and pull operations
for data movement. These terms are defined from the perspective of control
storage system.
Non-disruptive data
migration from
storage system A to B
The LUNs created at the storage systems are assigned to the appliance. The
appliance abstracts the identity of these LUNs and creates a storage pool by
aggregating LUNs from the storage systems. A virtual volume is created from the
storage pool and assigned to the compute system. When an I/O is sent to a virtual
volume, it is redirected through the virtualization layer at the SAN to the mapped
LUNs.
Clients
Migration
software runs
In a NAS to NAS direct data migration, file-level data is migrated from one NAS
system to another directly over the LAN without the involvement of any external
server.
Compute system
LAN
In a NAS to NAS data migration through intermediary compute system, all the data
is transferred through the compute system from the old NAS system to the new
NAS system. In this method of migration -
Clients
In this type pf NAS migration, the virtualization appliance facilitates the movement
of files from old NAS system to new NAS system. While the files are being moved,
clients can access their files non-disruptively. Clients can also read their files from
the old location and write them back to the new location without realizing that the
physical location has changed.
Host-based Migration
• It uses host operating system to migrate data from one storage to another. This
approach uses host resources to move data non-disruptively from a source to a
target.
• It works in conjunction with storage system-based replication and migration
solutions to migrate data from one storage to another.
Migrated VMs
VM Migration
Network
Storage System
In this type of migration, virtual machines (VMs) are moved from one physical
compute system to another without any downtime. This enables -
Compute System
Network
VM Storage
Migration
Storage Storage
system system
In a VM storage migration, VM files are moved from one storage system to another
system without any downtime or service disruption.
Application Migration
Application
Migration
Physical
compute system
VM disk
Network
Application migration typically involves moving the application from one data center
environment to another. Typically, the organization can move the application from
physical to virtual environment. In a virtualized environment, the application can
also be moved from one hypervisor to another for various business reasons such
as balancing workload for improving performance and availability.
Compute
system
Network
Storage system
Storage system
Concepts in Practice
Concepts in Practice
Dell EMC VPLEX provides continuous data availability, transparent data mobility
and non-disruptive data migration for mission critical applications. VPLEX delivers
high performance for the latest flash storage technology in combination with
reduced latency to ensure business critical applications are never down and
VPLEX delivers greater than five 9’s availability. VPLEX enables data and workload
mobility across arrays and datacenters without host disruption. Ansible modules for
VPLEX enable operational teams to rapidly provision storage infrastructure with
accuracy to respond to the fast-paced needs of application developers. VPLEX
requires no compute resources from the application hosts or on the underlying
array to maximize data availability.
Dell EMC Intelligent Data Mobility services enable organizations to reduce the time,
cost and complexity of data migration. Dell EMC Intelligent Data Mobility enables
fast and simple data migration to storage solutions like Dell EMC Unity, a simple,
modern, flexible and affordable flash storage solution for midrange storage. It
provides customers with the flexibility, simplicity and efficiency to seamlessly move
data and workloads by using technology, automation and Dell EMC expertise.
Intelligent Data Mobility follows a standardized methodology to minimize the time
and expense of onboarding new storage.
active memory and precise execution state over a high-speed network, allowing the
virtual machine to switch from running on the source vSphere host to the
destination vSphere host.
1. Present Scenario:
• Has a file-sharing environment in which multiple NAS systems serve all the
clients including application servers.
2. Organization’s Challenges:
• It is also identified that some of the NAS systems are over utilized and some
of the NAS systems are underutilized - –Clients are impacted when
accessing the over utilized NAS systems.
3. Organization’s Requirements:
4. Expected Deliverables:
Solution
• Allows to move files from over utilized NAS system to underutilized NAS system
without impacting the client
Objectives
Click on the name of each plane on the image for more information.
3: Provides the programming logic and policies that the data plane follows to
perform its operations.
The key functions of the control plane include asset discovery, resource abstraction
and pooling, provisioning resources for services.
Benefits Description
Objectives
Software- Defined
Compute
• Hypervisor decouples the application and the OS from the hardware and
encapsulates them in an isolated virtual container called a virtual machine (VM).
• Hypervisor controls the allocation of hardware resources to the VMs based on
policies, which means the hardware configuration of a VM is maintained using a
software.
SDS Software/Controller
Commodity
• Discovery64
• Resource abstraction and pooling65
• Service provisioning66
Physical storage systems are separated into two parts- virtual storage system and
virtual storage pool. Let us understand each of them.
64SDS controller discovers physical storage systems to gather data and bring them
under its control and management.
65SDS controller abstracts physical storage systems into virtual storage systems
and virtual storage pools as per policies and also enables an administrator to
define storage services.
66SDS controller automates the storage provisioning tasks and delivers virtual
storage resources based on the service request issued through a service catalog.
• A virtual storage pool is a logical entity that maps to the storage pools in the
virtual storage systems.
• Administrator may configure multiple virtual storage pools of different capacity,
performance, and protection characteristics based on the policy.
• A virtual storage pool may include storage pools from multiple virtual storage
systems.
Switch
Switch Switch
• Controls the switching and routing of the network traffic independent of the
underlying network.
• Abstracts the physical details of the network components and separates the
control plane functions from the data plane functions.
67The function of the data plane is to transfer the network traffic from one physical
port to another by following rules that are programmed into the component.
68The function of the control plane is to provide the programming logic that the
data plane follows for switching or routing of the network traffic.
• Provides instructions for data plane to handle network traffic based on policies.
• Provides CLI and GUI for administrators to manage the network infrastructure
and configure policies and APIs for external management tools and application
to interact with the SDN controller.
• Discovery69
• Network component management70
• Network flow management71
Virtual Network
71SDN controller controls the network traffic flow between the components and
chooses the optimal path for network traffic.
72A virtual network appears as a physical network to the compute and storage
systems (called nodes) connected to it, because the existing network services are
reproduced in a virtual network.
Virtual Switch
Virtual Router
Clients
Virtual Router
Physical
Switch
Virtual Switch Virtual Switch
Physical
NIC
Compute-based SAN
C S C S C S C S
C
Client Program
− Owns the local storage and performs I/O operations as requested by a client
from a compute system within the cluster.
Objectives
Management Tools
IT Infrastructure
− User may request for a protection service from the self-service portal and
the software controller will fulfill the requests automatically.
• Software controller leverages the protection technologies that are either natively
built into the underlying IT infrastructure components or provided by a separate
protection applications.
− Controls and manages the protection applications, storage, and operations
according to predefined policies.
• The data protection process in an SDDC consists of three key phases. These
are:
74 Data protection services are defined by administrators using the service catalog.
Service
Delivery
• Users request for services from a service catalog on self-service portal. The
portal interacts with the orchestrator and transfers service requests.
• Orchestrator interacts with appropriate components to orchestrate execution of
component-related tasks based on pre-defined workflows.
• Components that may be considered for integration are shown in the image.
Click on each integration components for detailed information.
5 3 1
4 2
5: It collects and records the usage of services per user group or consumer in
number of units consumed of a service.
Examples of a service unit are: per GB of storage, per transaction, and per hour of
application usage.
• It also generates billing report76 based on price per unit and number of units
consumed of a service.
76 The billing report is visible to the user through the cloud portal.
Concepts in Practice
Concepts in Practice
Click the right and left arrows to view all the concepts in practice.
ECS, the leading object storage platform from Dell EMC, provides unmatched
scalability, performance, resilience, and economics.
Dell EMC Unity Cloud Edition lets you deploy Dell EMC Unity unified storage as a
virtual storage appliance directly in an AWS cloud.
• Dell EMC Unity Cloud Edition is software-defined storage that runs on industry-
standard hardware and VMware ESXi.
• Enterprise Capabilities such as Snapshots, Quotas and Tiering are delivered
with Common Unity experience.
• With Unity Cloud Edition, File Services are consumed within each Customer
SDDC so there is no need for an External File Appliance or File Service.
• Dell EMC Unity Cloud Edition enables Cloud Synch Disaster Recovery between
on premises-deployed Dell EMC Unity systems and VMware Cloud-based
appliances.
− This block and file solution is ideal for a variety of use cases in the cloud
including home directory for running a VDI environment in VMware Cloud,
Test/Dev operations, or replication services to a third site.
VMware vSAN
• Reduce the cost and complexity of traditional storage and take the easiest path
to future ready hyper-converged infrastructure and hybrid cloud.
• Improve business agility, all while speeding operations and lowering costs when
integrated with hyper-converged infrastructure (HCI) solution.
• Modernize the infrastructure by leveraging existing tools, skillsets and software
solutions.
• Simplify the extension from on premises to the public cloud.
vSAN is integrated with vSphere, optimizing the data I/O path to provide the
highest levels of performance with minimal impact on CPU and memory.
vSAN minimizes storage latency with built-in caching on server side flash devices,
delivering up to 50 percent more IOPS than previously possible.
VMware NSX Data Center is the network virtualization and security platform that
enables the virtual cloud network.
• NSX Data Center reproduces the entire network model in software, enabling
any network topology—from simple to complex multitier networks—to be
created and provisioned in seconds.
• Users can create multiple virtual networks with diverse requirements, leveraging
a combination of the services offered via NSX.
• SDS controller provides a single virtual storage pool for all the VMs to store
email data.
2. Organization’s Requirement:
• Both the data centers must have capability to failover services automatically
in the event of a disaster.
3. Expected Deliverables:
Solution
Objectives
Cloud Infrastructure
Compute Platform
Network Storage Applications
Systems Software
Desktop Mobile
Thin Client Devices
computing resources that can be rapidly provisioned and released with minimal
management effort or service provider interaction.”
• Consumers pay only for the services that they use, either based on a
subscription or based on resource consumption.
On-demand
self service
enables Broad network access
consumers to provides capabilities that are
unilaterally available over the network
provision and accessed through
computing standard mechanisms
capabilities as
needed
The five essential characteristics or tenets of a cloud (as defined by NIST) are:
Rapid elasticity
On-Demand self-service
• The end user can provision computing capabilities themselves allowing them to
allocate things such as server time and network storage, as needed
automatically without requiring human interaction with each service provider.
Resource pooling
Measured service
• Capabilities are available over the network and accessed through standard
mechanisms that promote use by heterogeneous thin or thick client platforms
(e.g., mobile phones, tablets, laptops, and workstations).
To learn more about Cloud Service Models and Infrastructure as a Service, click
here.
• Cloud deployment models provide basis for how cloud infrastructure is built,
managed, and accessed.
• Each cloud deployment model may be used for any of the cloud service models:
IaaS, PaaS, and SaaS. The different deployment models present a number of
tradeoffs in terms of control, scale, cost, and availability of resources.
Public Cloud
Enterprise P Enterprise Q
Individual R
IT resources are made available to the general public or organizations and are
owned by the cloud service provider.
Private Cloud
Enterprise P
Resources of
Enterprise P
Cloud infrastructure is operated solely for one organization and is not shared with
other organizations. This cloud model offers the greatest level of security and
control.
Multi Cloud
Some application workloads run better on one cloud platform while other workloads
achieve higher performance and lower cost on another one.
The wide variety of business requirements result in a need for various cloud
offerings. For example, one might use Amazon EC2 for computing and Microsoft
Azure for data lake storage while leveraging Google Cloud SQL.
Enterprise P Enterprise Q
Resources of Resources of
Enterprise P Enterprise Q
Enterprise R
One or more participant organizations provide cloud services that are consumed by
the community.
Community Users
Enterprise R
Enterprise P
Enterprise Q
IT resources are hosted on the premises of the external cloud service provider and
not within the premises of any of the participant organizations.
Hybrid Cloud
Enterprise Q
Cloud Provider's
Resources
Enterprise P
Resources of
Public Cloud
Enterprise P
IT resources are consumed from two or more distinct cloud infrastructures (private,
community, or public).
Cloud Benefits
Objectives
On-demand, self-
Reduced CAPEX
service provisioning
Organizations need to regularly protect the data to avoid losses, stay compliant,
and preserve data integrity. Data explosion poses challenges such as strains on
the backup window, IT budget, and IT management. The growth and complexity of
data environment, added with proliferation of virtual machines and mobile devices,
constantly outpaces the existing data protection plans. Deployment of a new data
protection solution takes weeks of planning, justification, procurement, and setup.
Enterprises must also comply with regulatory and litigation requirements. These
challenges can be addressed with the emergence of cloud-based data protection.
Backup as a Service
Cloud Resources
The three common backup service deployment options that cloud service providers
offer to their consumers are:
Cloud Resources
Cloud Resources
Cloud Resources
Cloud Resources
Consumer organization
accesses cloud-hosted
applications (SaaS-based Backup data to the
application) third-party cloud
Consumer
Organization
Cloud
Disasters sometimes
happen at the consumer
production Data Center
Organization's Data
Center
• Requested data is gathered and sent to the server, running cloud backup agent.
• Received data is in an encrypted form. The agent software on the server
decrypts the files and restores it on the server.
• Considered if sufficient bandwidth is available to download large amounts of
data or if the restore data is small in size.
Cloud
Organization's Data
Center
Backup to Backup to
cloud cloud
Remote Remote
Office 1 Office 2
Backup to
cloud
Remote
Office 3
• Cloud backup services deploy disk-based backup solutions along with source-
based deduplication to eliminate the challenges associated with centrally
backing up remote-office data.
• Performing backup to the cloud, reduces the cost of managing the
organization’s ROBO backup environment.
Cloud
Replicating data to the cloud Resources
Backup
Storage
Clients
Cloud Resources
Backup is performed in
consumer's location Consumer
Organization
Objectives
Cloud-only Archiving
Cloud Archive
Storage
Email Servers Non-critical data on the
primary storage system is
Archive Server moved to the public cloud
(Policy Engine)
File Server
Inactive Data
Organization's Private Cloud
Hybrid Archiving
Cloud Archive
Archive server
determines which data Storage
needs to be archived
based on policies
Email Servers
Archive Server
(Policy Engine) Archive
Data
Active Data
Inactive Data
Organization's Data Center
• Organization’s inactive data (both critical and non-critical) that meets the
organization’s archiving policies is archived to the cloud.
− IaaS - Archiving server on its data center and the archiving storage will
reside on the cloud.
− SaaS - Both the archiving server and the archiving storage reside on cloud
infrastructure.
Cloud-based Storage-Tiering
Less frequently
accessed data is
moved to tier 2 Storage HDD Cloud
Storage SDD
Storage
Backup
Application
Application Data Migration
Servers
Archive Server
Cloud
Archive
Storage
Primary
Clients Storage
• An organization may decide to migrate from one cloud provider to another when
it identifies that the cloud service provider is not able to meet the SLAs, not
adhering to security best practices, not meeting acceptable performance, or not
able to fulfill its future requirements.
• Since different cloud vendors may have different protocols and architecture,
data migration between clouds requires integration tools that will migrate the
data from one cloud to another.
• Cloud integration tool should provide features such as simplicity, flexibility,
interoperability, data portability, data integrity, security, reliability, and ease of
management.
Application Servers
Cloud Gateway
Block Based Appliance
Interface REST
Data Center
Cloud
3. Which archiving method is most suitable and cost effective for a large
organization having both sensitive data and non-sensitive data?
a. Cloud-only Archiving
b. Hybrid Archiving
c. In-house Archiving
Concepts in Practice
Concepts in Practice
Click the right and left arrows to view all the concepts in practice.
Enables users to connect their file and block storage – Dell EMC Unity,
PowerStore, PowerMax and PowerScale - consumed as a service, directly to public
cloud(s) including VMware Cloud on Amazon Web Services (AWS), AWS,
Microsoft Azure and Google Cloud Platform. This is done through a high-speed,
low latency connection from Dell EMC storage at a managed service provider to
the cloud or clouds of choice. Organizations gain an on demand, cloud
consumption model for both compute workloads and storage combined with the
high performance, up to 6-9s availability, and scalability of Dell EMC storage. This
solution is ideal for securely moving or deploying demanding applications to the
public cloud for disaster recovery, analytics, test/dev and more.
VMware Cloud on Dell EMC is a fully managed hybrid cloud service that combines
the simplicity and agility of the public cloud with the security and control of on-
premises infrastructure. Delivered as a service to data center and edge locations,
VMware Cloud on Dell EMC and its hybrid cloud services provide simple, secure,
and scalable infrastructure. Enable intrinsic security, including encryption for data
at rest and in transit. For additional security, there are micro-segmentation
capabilities available through VMware NSX. VMware Cloud on Dell EMC simplifies
the management of your data center services and edge infrastructures with an
offering that is fully managed, subscription based, and delivered as-a-service.
2. Organization Challenges:
• Does not want to build and manage its own DR site due to budget
constraint.
• Increases the complexity and cost while managing the huge volume of
inactive data within its data center.
3. Organization Requirements:
4. Expected Deliverables:
Solution
Objectives
• Big Data represents the information assets whose high volume, high velocity,
and high variety require the use of new technical architectures and analytical
methods to gain insights and derive business value.
• The definition of Big Data has three principal aspects:
Characteristic of Data
• Apart from its considerable size (volume), the data is generated rapidly
(velocity) and is highly complex as it comes from diverse sources (variety).
Nearly 80-90 percent of the data getting generated is unstructured data.
Business Value
• Big Data has tremendous business importance to organizations and even to the
advancement on society.
• Proper analysis of big data will help to make better business decisions and add
value to business.
• Big Data analytics has many applications spanning numerous industry sectors
and scientific fields.
In 2001, Gartner analyst Douglas Laney specified volume, velocity, and variety as
the three dimensions of the challenges associated with data management. These
dimensions— popularly known as “the 3Vs"—are now widely accepted in the
industry as the three primary characteristics of Big Data. In addition to the 3Vs,
there are three other characteristics identified by the industry namely variability,
veracity, and value.
Volume
Velocity
• Refers to the rate at which data is produced and changes, and how fast the
data must be processed to meet business requirements.
• Real-time or near real-time analysis of the data is a challenge for many
organizations.
Variety
• Variety (also termed as “complexity”) refers to the diversity in the formats and
types of data.
• Data is generated by numerous sources in various structured and unstructured
forms. New insights are found when these various data types are correlated and
analyzed.
• Pertains to challenge of managing, merging, and analyzing different varieties of
data in a cost-effective manner.
• The combination of data from a variety of data sources and in a variety of
formats is a key requirement in Big Data analytics.
Variability
− For example, natural language search and analyzing social media posts
require interpretation of complex and highly-variable grammar. The
inconsistency in the meaning of data gives rise to challenges related to
gathering the data and in interpreting its context.
Veracity
Value
• Refers to both the cost-effectiveness and the business value derived from the
use of Big Data analytics technology.
• Many organizations have maintained large data repositories such as data
warehouses, managed non-structured data, and carried out real-time data
analytics for many years.
Storage
MapReduce
Query
HDFS is a distributed file system that provides access to data across nodes –
collectively called a “Hadoop Clusters”. HDFS architecture has two key
components:
Hadoop Cluster
Rack 1
Clients
Name Node
Rack 2
• Name Node:
− Acts as a Primary server and has in-memory maps of every file, file
locations, as well as all the blocks within the file and the Data Nodes to
which they reside on.
− Responsible for managing FS namespace and controlling the access of files
by the clients.
• Data Node:
Sources
Ingest Store
Analyze
Surface
Act
Objectives
Protecting a big data environment requires new strategies about how to use the
existing tools and adopting new technologies that help in protecting the data more
efficiently.
Sources
Sources Store
Analyze
Surface
Act
To learn more about Data Lake as a Repository for Big Data, click here.
Backup Data
HDFS Data Lake
DistCp Tool
Backup Device
Replication
Replication
Data is mirrored to multiple Method to protect striped data from disk drive failure
nodes. or node failure.
If the cluster is setup for Data is fragmented and encoded with parity data
3X mirroring, the original and stored across a set of different locations (drives
file will be stored along and nodes).
with two copies of the file
in various locations within
the cluster.
To learn more about Data Mirroring and Parity Protection, click here.
Service provider offers resources to enable the consumers to run big data analytics
workload in the cloud
• Allows the consumers to analyze and build analytics applications on top of huge
volume of data. The service provider offers platform (database, Hadoop) and
cloud infrastructure to run or build analytics applications.
2. Which native utility is built into HDFS to backup and restore data from the data
lake to a backup device?
a. HDFS Mirroring
b. Hadoop Distributed Copy
c. Erasure Coding
d. Hadoop Data Copy
Objectives
A compute system that is portable and typically a handheld device with a display,
and has either a keyboard and/or touch input.
• Data is protected (backed up) only when the mobile device is online.
• Data protection from mobile device to data center is impacted due to
intermittent network connectivity.
• Devices are not always connected to the corporate network, so it connects over
the Internet, which may rise to a security threat.
• Data protection software must support the mobile device OS.
• Network bandwidth limitations.
• Security features on the mobile devices restrict the access of the data stored on
the device.
To learn more about Challenges in Protecting Mobile Device Data, click here.
Mobile Backup
Clients
Backup Data
Enterprise Data
Center
Mobile Devices
File Sync-and-Share
Application Server
File Storage
Mobile Devices
Cloud
Resources
• Compute processing and storage are moved away from the mobile device and
takes place in a computing platform located in the cloud.
• Applications running in the cloud are accessed over wireless connection using a
thin client application/web browser on the mobile devices.
• Cloud services accessed over mobile devices.
Mobile
Devices Cloud
Resources
Backup Data to
the Cloud
• Backup client application (agent) that is installed on the device enables access
to perform backup to the cloud.
− Typically backs up only the changed blocks to the cloud storage.
• Some mobile applications have built-in backup feature that backs up the data to
the cloud.
• Most of the cloud backup solutions available today offer a self-service portal
that allows users to recover data without manual intervention.
• Ideal for larger businesses or multiple branch offices working together over
geographically dispersed environments.
• Backup solutions that span multiple types of devices (from servers to PCs to
mobile devices) will serve the organization well to provide business continuity
and reduce backup and data sharing complexities.
• Cloud-based backup easily institutes policies that govern backup processes and
access control.
• Establishing particular levels of service can be well defined through service-
level agreements (SLAs).
Concepts in Practice
Concepts in Practice
ECS, the leading object storage platform from Dell EMC, provides unmatched
scalability, performance, resilience, and economics.
• Plans to implement big data analytics for their business along with
necessary data protection solutions.
• Currently, it does not have infrastructure to support big data analytics and its
protection.
3. Organization Requirements:
• Need a solution to implement big data analytics but looking for OPEX cost
model.
4. Expected Deliverables:
Solution
Objectives
Orchestration
Operations Management
Discovery
Interaction Interaction
Backup
Applications
Replication
Applications
Interaction Archiving
Applications
Data security includes a set of practices that protect data and information systems
from unauthorized disclosure, access, use, destruction, deletion, modification, and
disruption.
• The two key drivers for organization’s data security are Confidentiality, Integrity,
and Availability (CIA); and Governance, Risk, and Compliance (GRC)
requirements.
Governance
Board of Directors
Finance
Risk
Steps Description
Compliance
Types of Compliance
Controls the nature of IT operations within an Includes legal requirements, legislation, and
organization industry regulations
• An act of adhering to, and demonstrating adherence to, external laws and
regulations as well as to corporate policies and procedures.
• There are primarily two types of policies controlling IT operations in an
enterprise that require compliance: internal policy compliance and external
policy compliance.
• Compliance management activities include:
1:
• A process to ensure that users or assets are who they claim to be by verifying
their identity credentials.
• A user may be authenticated by a single-factor77 or multi-factor78 method.
2:
• Refers to the logging of all transactions for the purpose of assessing the
effectiveness of security mechanisms.
• Helps to validate the behavior of the infrastructure components, and to perform
forensics, debugging, and monitoring activities.
Vulnerabilities
1 2 3
1:
• Refers to the various entry points that an attacker can use to launch an attack,
which include people, process, and technology.
• For example, each component of a storage infrastructure is a source of
potential vulnerability.
− An attacker can use all the external interfaces80 supported by that
component, such as the hardware and the management interfaces, to
execute various attacks.
• Unused network services, if enabled, can become a part of the attack surface.
2:
• Refers to the amount of time and effort required to exploit an attack vector.
Defense-in-depth
Objectives
Modification of system
configuration by unauthorized
Data Security and Management
access to management
application
Data Protection Services
Orchestration
Operations Management
Discovery
Restore data to
Unauthorized modification or
unauthorized destination via
deletion of data through
protection application
application, file system, or
database Interaction Interaction
Backup
Applications
Replication
Applications
Interaction Archiving
Applications
The threats in the data protection environment may exist at data source, protection
application and protection storage, and data management domain.
Compute
System A
Database Multimedia
Application Application
An attacker may gain
unauthorized access to
application by bypassing access
control
The protection applications are responsible for creating backups and replicas to
ensure business continuity.
An attacker is gaining
Primary Cloud access to user data
Storage Protection
Storage
Backup Agent
Backup
Agent
Compute
Backup Device
System B
(Protection
Storage) An attacker is stealing the
physical media by gaining
Protection
access to protection storage
Storage
An attacker is stealing the
physical media by gaining
access to protection storage
Backup Server/Storage
Node
The protection storage is exposed to various kinds of threats in both the backup
and the replication environment.
Storage Storage
System A System B
Attacker may gain unauthorized access
to management application to perform
unauthorized resource provisioning.
VSAN
Management Compute System A Compute System B
Applications
VLAN
Management VLAN
The management application provides visibility and control of the components and
protection operations.
Objectives
Security controls should involve all the three aspects of infrastructure: people,
process, and technology, and their relationships.
Physical Security
Implement biometric or
Physical Security Controls security badge-based
authentication
Install surveillance
cameras
81An attack that relies heavily on human interaction and often involves tricking
people into breaching security measures.
Network
Password
User
User Verified
Cloud
Verification
Biometric Request
• Click on the example to know how a user is validated for identity and privileges.
• Multi-factor authentication82 uses more than one factor to authenticate a user.
Security
Create, delete, and modify security settings
Administrator
Security Controls
Click the right and left arrows to view all security controls.
Firewall
84Clear separation of duties ensures that no individual can both specify an action
and carry it out.
85A rule may use various filtering parameters such as source address, destination
address, port numbers, and protocols.
86 A physical firewall is a device that has custom hardware and software on which
filtering rules can be configured. Physical firewalls are deployed at the network
level.
87A virtual firewall is a software that runs on a hypervisor to provide traffic filtering
service. Virtual firewalls give visibility and control over virtual machine traffic and
enforce policies at the virtual machine level.
IDPS
Block
anomalous
activity Servers
Attacke
r
Intrusion detection is the process of detecting events that can compromise the
confidentiality, integrity, or availability of IT resources.
• Intrusion Detection System (IDS)88 and Intrusion Prevention System (IPS)89 are
the two controls usually work together and are generally referred to as intrusion
detection and prevention system (IDPS).
• The key techniques used by an IDPS to identify intrusion in the environment
are:
− Signature-based detection90
88A security tool that automates the detection process. An IDS generates alerts, in
case anomalous activity is detected.
89A tool that has the capability to stop the events after they have been detected by
the IDS.
90IDPS relies on a database that contains known attack patterns or signatures, and
scans events against it.
− Anomaly-based detection91
VPN Clients
Remote user connects to the corporate
Remote User
network using VPN connection
91 IDPS scans and analyzes events to determine whether they are statistically
different from events normally occurring in the system.
VLAN
Compute
System E
Compute
VLAN configured on Ethernet Switch A
System C
provides traffic isolation and therefore
enhanced security
Ethernet Switch A Compute
System A
Compute
System F
IP Router
Compute
System D
Storage
System C
92 A remote client (typically client software installed on the user’s compute system)
initiates a remote VPN connection request. A VPN server authenticates and
provides the user access to the network.
93The remote site initiates a site-to-site VPN connection. The VPN server
authenticates and provides access to internal network.
VSAN
VSAN 10 VSAN 20
VSAN10 allows traffic between Compute (Engineering) (Finance)
System A and Storage System A. VSAN
10 also restricts traffic from VSAN 20 FC Switch A
Compute Compute
System A System B
FC Switch B
Storage Storage
System A System B
Zoning
Switch Domain ID = 15
Port 5
WWN 10:00:00:00:C9:20:DC:40
Port 12
Compute Port 9
System
Zone 3 (Mixed Zone) =10:00:00:00:C9:20:DC:56; 15,12
WWN 10:00:00:00:C9:20:DC:56
− WWN zoning
− Port zoning
− Mixed zoning
To learn more about types of zoning, click here.
LUN Masking
Compute system A can have access to LUN A and restricts the access to
LUN B
LUN A
Compute System A - HR
LUN B
Storage System
Compute system B can have access to LUN B and restricts the access to
LUN A
• LUN masking is the storage system-based security control that is used to:
− Protects against unauthorized access to LUNs of a Storage System.
− Grants LUN access only to the authorized hosts.
• Consider a Storage System with two LUNs that store data of the HR and
finance departments as shown in the image.
− Without LUN masking, both the departments can easily see and modify each
other’s data, posing a high risk to data integrity and security.
− With LUN masking, LUNs are accessible only to the designated hosts.
Discovery Domain
Storage System
Management A
iSNS can be part of a
Station network or a
management station
Compute
System A
• Internet Storage Name Service (iSNS) discovery domains work in the same way
as FC zones and primarily used in IP-based network.
• Provides functional groupings of devices in an IP-SAN.
• For devices to communicate with one another, they must be configured in the
same discovery domain.
− State change notifications inform the iSNS server when devices are added
or removed from a discovery domain.
• The image shows the discovery domains in an iSNS environment.
Objectives
Storage
Storage
System B
System A
•Restrict core functionality to selected
administrators
•Encrypt network traffic when managing remotely
•Design with proper architecture, threat modeling, and
•Deploy firewall between management system and secure coding
rest of the network
•Include process spawning control, executable file
protection, and system tampering protection
VSAN
VLAN
Management VLAN
• The hypervisor and the related management servers are critical components of
an IT infrastructure because they control the operation and management of the
virtualized compute environment.
Component Roles
For detailed information about IT infrastructure components and their roles against
security attacks, click here.
Viruses Spyware
Worms
Key Loggers
Trojans
Signature based detection scans the files to determine Malware protection software protects against these
signature attacks
94
In this technique, the malware protection software scans the files to identify a
malware signature.
96Installed on the mobile device that needs access to the organization’s resources.
The client receives commands from the server component which it executes on the
mobile device.
Data Encryption
Encryption Decryption
Appliance Replication Appliance
Data Shredding
Cyber Recovery
Cyber Recovery
Objectives
Cyber Attacks
Select the link on each title to learn more about the most common cyber attacks.
• Global crime damage is predicted to grow by 15 percent per year over the next
five years, reaching $10.5 trillion USD annually by 2025. This number is an
increase from 3 trillion USD in 2015.
Cyberwarfare-2021-Report.pdf (netdna-ssl.com)
− Keep the backup copies offline, where cyberattacks cannot access the
secure copies.
− Keep security software up to date on latest definitions of virus and malware.
− Keep operating systems and software updated with security patches.
− Educate employees to be aware of links or attachments in suspicious email
messages.
• True data protection emphasizes keeping an isolated copy of your critical data
such as essential applications and intellectual property off the network.
• Cyber recovery architecture:
1
4
5
6
2: Creates point-in-time copies that can serve as restore points in case production
backup data is subject to destructive cyberattack.
Synchronizes the latest data, creates a copy, and then secures it.
3: Immutable file locking and secure data retention to meet both corporate
governance and compliance standards.
5:
6: The data in a point-in-time copy can be re-orchestrated and then used to replace
the lost data in production.
5. Match the type of attack description with the name of the attack.
Concepts in Practice
Concepts in Practice
Click the right and left arrows to view all the concepts in practice.
• Dell Encryption Enterprise offers options with its flexible encryption technology
such as data-centric policy-based approach as well a Full Disk Encryption
approach to protect data.
• The solution is designed for:
− Ease of Deployment
− End-user transparency
− Hassle-free compliance
• Provides mobile banking to its customers that enables them to access the
application and data from any location.
• Currently performs remote replication between the primary site and the
secondary site for DR.
2. Organization’s Challenges:
• Sending tapes to offsite locations would increase the risk of losing sensitive
data in transit.
3. Organization’s Requirements:
4. Expected Deliverable:
Solution
Objectives
Data Protection Management includes all the protection-related functions that are
necessary for the management of data protection environment and services, and
for the maintenance of data throughout its lifecycle.
Scroll down to learn more about the need for data protection and management.
Overly complex
Manual operations
The traditional management processes and tools may not support a service-
oriented infrastructure, especially if the requirement is to provide cloud services.
Interoperability issues
They usually lack the ability to execute management operations in an agile manner,
scale resources rapidly, respond to adverse events quickly, orchestrate the
functions of distributed infrastructure components, and meet sustained service
levels. This component-specific, extremely manual, time consuming, and overly
complex management is simply not appropriate for modern-day data protection
management.
Click the right and left arrows to learn about the characteristics.
Service-focused approach
• Creating a disaster recovery plan to meet the recovery time objective (RTO) of
services.
• Ensuring that the management processes, management tools, and staffing are
appropriate to provide a data archiving service.
End-to-end visibility
Orchestrated operations
Data protection management performs two key functions, which are as follows:
• Discovery
• Operations management
Discovery
Discovery provides the visibility needed to monitor and manage data center infrastructure
Discovery tool
interacts and collects
information from the
components
Operations Management
Operations Management – 1
Operations Management – 1
Objectives
• Define monitoring.
• Explain alerting.
• Understand the concept of reporting.
Introduction to Monitoring
Monitoring provides visibility into the data protection environment and forms the
basis for performing management operations. It offers the following benefits:
Monitoring
Monitoring Parameters
Monitoring Parameters
Availability
Configuration
Capacity
Performance
Security
Monitoring Configuration
This table shows a list of backup clients (VMs), their type, CPU and memory
configurations, and compliance to a predefined backup policy. The VM
configurations are captured and reported by a monitoring tool.
Monitoring Availability
Application Servers/Backup
Clients No redundancy due to switch SW1
failure
Backup Storage
System
Storage Node
Monitoring Capacity
Pool expanded
Monitoring Performance
Backup Storage
System
Performance monitoring -
Monitoring Security
Workgroup 2 (WG2)
Workgroup 1 (WG1)
Notification: Attempted replication of WG2 devices by
WG1 user – Access denied
Alerting
Reporting
Reporting on the data protection environment involves keeping track and gathering
information from various components and protection operations that are monitored.
The gathered information is compiled to generate reports for trend analysis,
capacity planning, configuration changes, deduplication ratio, chargeback,
performance, and security breaches.
Click the report types on the given figure for more information about that report.
2
1
3
5
4
1: Capacity planning reports contain current and historic information about the
utilization of protection storage, file systems, ports, etc.
2: Configuration and asset management reports include details about the allocation
of protection storage, local or remote replicas, network topology, and unprotected
systems. This report also lists all the equipment, with details, such as their
purchase date, license, lease status, and maintenance records.
3: Chargeback reports contain information about the number of backup and restore
operations, amount of data backed up and restored, amount of data retained over a
period of time, and the number of tapes as archive storage media used by various
user groups or tenants along with the associated cost.
Operations Management - 2
Operations Management -2
Objectives
Configuration Management
SDDC Controller
The information about CIs include their attributes, used and available capacity,
history of issues, and inter-relationships.
Change Management
Capacity Management
Capacity Management ensures that the data protection environment is able to meet
the required capacity demands for protection operations and services in a cost
effective and timely manner.
Performance Management
Availability Management
Incident Management
The following table illustrates an example of an incident that was detected by the
Incident Management tool:
Severit Event Type Devic Priority Statu Last Updated Owne Escalatio
y Summar e s r n
y
Problem Management
Problem management:
• Managing user accounts and access policies that authorize users to use a
backup/replication service.
• Implementing controls at multiple levels (defense in depth) to access data and
services.
• Scanning applications and databases to identify vulnerabilities.
• Configuring zoning, LUN masking, and data encryption services.
With the flow of personal data across industries and on a global scale, data security
governance and data protection compliance requirements are becoming stronger
day by day. Organizations, which are dealing with personally identifiable
information (PII) must comply with stringent data protection regulations, including:
• Payment Card Industry Data Security Standard (PCI DSS) in the USA.
• Health Insurance Portability and Accountability Act (HIPAA) in the USA.
• General Data Protection Regulation (GDPR) in Europe.
• California Consumer Privacy Act (CCPA) in California.
• POPI in South Africa.
There are three primary software methods for DSG: classification, discovery, and
de-identification or masking. These methods have been successfully employed
by IRI customers for PII and other sensitive data.
Click each primary software method type on the given figure for more information.
1 2
2: Sensitive data can be found by using certain search functions, which may or
may not be associated with data classes. This function is known as discovery
technique. Examples of discovery include:
3: A great way to reduce or even eliminate data breach risks is masking of data, at
rest or in motion. This process masks or shields sensitive or confidential data, such
as names, addresses, credit card information, Social Security numbers etc. from
the risk of unintended exposure to prevent data breaches.
Concepts in Practice
Concepts in Practice
Dell EMC PowerProtect Data Manager provides software defined data protection,
automated discovery, deduplication, operational agility, self-service and IT
governance for physical, virtual and cloud environments.
With operational simplicity, agility and flexibility at its core, PowerProtect Data
Manager enables the protection, management and recovery of data in on-
premises, virtualized and cloud deployments, including protection of in-cloud
workloads.
Dell EMC PowerProtect Data Manager builds on top of project Velero to provide a
data protection solution that enables application-consistent backups and restores
and that is always available for Kubernetes in on-premises and in-cloud workloads,
VMware hybrid cloud environments and Tanzu modern applications.
Dell EMC Data Protection Advisor can automate and centralize the collection and
analysis of all data—and get a single, comprehensive view of organization’s data
protection environment and activities. With automated monitoring, analysis, and
reporting across backup and recovery infrastructure, replication technologies,
storage platforms, enterprise applications and virtual environment, organization will
be able to more effectively manage service levels while reducing costs and
complexity.
Data Protection Advisor’s analysis engine looks across the entire infrastructure to
provide end-to-end visibility into protection levels, performance, utilization and
more. This enables unified, cross-domain event correlation analysis – insight into
the entire data protection path to ensure each component is working correctly. And
that provides higher-level decision support based on defined policies. Built for cloud
infrastructure, Data Protection Advisor offers scalable, centralized multi-tenant data
protection management. With a SINGLE pane of glass view into your ENTIRE
infrastructure, every stakeholder has access to the information they need.
1. Present Scenario:
• The enterprise allows all its customer’s data to be stored, protected, and
accessed from worldwide location.
2. Organization’s Challenges:
3. Organization’s Requirements:
4. Expected Deliverables:
Solution
• Implement a capacity management process that will help in planning for current
and future resource requirements. This may include dynamic resource
consumption and seasonal spikes in resource demand.
• Deploy discovery tool that gathers and stores data in a configuration
management system.
• Deploy performance management tool that can proactively alert administrators
about potential performance issues.
• Orchestrate management operations that are common and repetitive to
reduce manual errors and administration cost.
• Implement an availability management process that will help in architecting the
new multi-site data protection solution.
Summary
Summary
Click the Save Progress and Exit button below to record this
eLearning as complete.
Go to the next eLearning or assessment, if applicable.
• Data Protection
• PowerProtect Data Manager • PowerProtect Data Manager • Implementing the NIST
- Data Protection Training Cybersecurity Framework
- PowerProtect Data Manager - PowerProtect Data Manager
Bundle
Training Bundle Training Bundle
• Avamar • Avamar
- Avamar Implementation and - Avamar Administration
Administration
• Data Domain
• Data Domain
- DataDomain System - DataDomain System
Administration Administration
• NetWorker • NetWorker
- NetWorker Implementation - NetWorker Implementation
and Administration and Administration
(C) - Classroom
Appendix
Data protection is one of the least glamorous yet important aspect in any
organization. In many respects it’s like being the goalkeeper in a soccer game—
when you do your job effectively, it’s easy to get overlooked. But if you fail, it
generally results in a loss. Data can exist in a variety of forms such as photographs
and drawings, alphanumeric text and images, and tabular results of a scientific
survey. In computing, digital data is a collection of facts that is transmitted and
stored in electronic form, and processed through software. Digital data is generated
by various devices such as desktops, laptops, tablets, mobile phones, and
electronic sensors. It is stored as strings of binary values (0s and 1s). In this
course, the word “data” implies the digital data. Most organizations use one or more
data protection methods to protect their digital data from disruption and disaster.
For example, backing up data creates a duplicate copy of data. The duplicate copy
or data backup is used to restore data in case the original data is corrupted or
destroyed. If a disaster occurs, an organization’s onsite data backup could be lost
along with the original data. Hence, it is a good practice to keep a copy of data in a
remote site. In addition, data archives are used to preserve older but important
files. Organizations also test data recovery operations periodically to examine the
readiness of their data protection mechanisms.
An organization’s data is its most valuable asset. An organization can leverage its
data to efficiently bill customers, advertise relevant products to the existing and
potential customers, launch new products and services, and perform trend analysis
to devise targeted marketing plans. These sensitive data, if lost, may lead to
significant financial, legal, and business loss apart from serious damage to the
organization’s reputation.
An organization seeks to reduce the risk of sensitive data loss to operate its
business successfully. It should focus its protection efforts where the need exists—
its high-risk data. Many government laws mandate that an organization must be
responsible for protecting its employee’s and customer’s personal data. The data
should be safe from unauthorized modification, loss, and unlawful processing.
Examples of such laws are U.S. Health Insurance Portability and Accountability Act
(HIPAA), U.S. Gramm-Leach-Bliley Act (GLBA), and U.K. Data Protection Act. An
organization must be adept at protecting and managing personal data in
compliance with legal requirements.
Data protection is the process of safeguarding data from corruption and loss. It
focuses on technologies or solutions that can prevent data loss and recover data in
the event of a failure or corruption. Data protection lays the foundation of improving
data availability.
Data protection technologies and solutions are used to meet data availability
requirements of business applications and IT services. Examples of IT services are
email service, data upload service, and video conferencing service. Data availability
refers to the ability of an IT infrastructure component or service to function
according to business requirements and end users’ expectations during its
operating time, ensuring that data is accessible at a required level of performance.
The operating time is the specified or agreed time of operation when a component
or service is supposed to be available.
For example, an application owner cares about the availability of their application,
and the application strategically requires 24x7 access to data. The backup
administrator is responsible for protecting the application data aptly using an
appropriate backup technology. In the event of a data corruption or loss, the
application owner relies on the backup administrator to restore data from a backup.
The goal of data availability is to ensure that users can access an application or a
service during its operating time. But failure of an infrastructure component or a
service might disrupt data availability and result in downtime. A failure is the
termination of a component’s or service’s ability to perform its required function.
The component’s or service’s ability can be restored by performing various external
corrective actions such as a manual reboot, a repair, or replacement of the failed
component(s). Therefore, both operating time and downtime of a component or a
service are factored in the measurement of data availability. Data availability is
usually calculated as a percentage of uptime, where uptime is equal to the
operating time minus the downtime. It is often measured by “Nines”. For example, a
service that is said to be “five 9s available” is available for 99.999 percent of the
agreed operating time in a year.
It is calculated as the total uptime divided by the number of failures. MTTR is the
average time required to repair a failed component or service. It is calculated as the
total downtime divided by the number of failures. These metrics are usually
expressed in hours.
For example, if the annual uptime of a component is 9609 hours, the annual
downtime of the component is 11 hours, and the component has failed thrice in an
year, then MTBF = 9609 hours / 3 = 3203 hours and MTTR = 11 hours / 3 = 3.66
Hours. Note: Mean Time to Restore Service (MTRS) is considered to be a better
metric than MTTR for measuring data availability. MTRS is the average time taken
to restore a failed component or a service.
The problem with MTTR is that while a component (or part of a service) may have
been repaired, the service itself is still not available to an end user. MTRS takes
care of the end user’s interest by encompassing the entire elapsed time after a
failure till the end user can get access to a service.
Natural disasters such as flood, earthquake, tornadoes, and volcanic eruptions can
affect businesses and availability of data in every part of the globe. In addition,
man-made disasters such as civil unrest, terrorist attacks, and accidents can
impact data availability.
Note: In general, the outages can be broadly categorized into planned and
unplanned outages. Planned outages may include installation and maintenance of
new hardware, software upgrades or patches, performing application and data
restores, facility operations (renovation and construction), and migration.
Unplanned outages include failure caused by human errors, database corruption,
failure of components, and natural or man-made disasters.
Data center usually stores large amounts of data and provides services to a vast
number of users. Therefore, data protection in a data center is vital for carrying out
business operations. There are several methods available to protect data in a data
center.
In an enterprise data center, data is typically stored on storage systems (or storage
“arrays”). A storage system is a hardware component that contains a group of
storage devices assembled within a cabinet. It is controlled and managed by one or
more storage controllers. These enterprise-class storage systems are designed for
providing high capacity, scalability, performance, reliability, and security to meet
business requirements. The compute systems that run business applications are
provided storage capacity from storage systems.
The commonly used cables are copper and optical fiber. A network adapter on a
compute or storage system provides a physical interface for communicating with
other systems.
Note: The OSI model defines a layered framework to categorize various functions
performed by the communication systems. The model has seven layers, and each
layer includes specific communication functions. If functions of a communication
protocol, a network switch, or a type of network traffic match with specific layer
characteristics, then they are often aliased by the layer number such as OSI layer-3
protocol, OSI layer-2 switch, and OSI layer-2 traffic.
• Pre-configured and optimized, which reduces the time to acquire and deploy the
infrastructure
• Less power and space requirements
• All hardware and software components can be managed from a single
management console
Notes
Organizations often keep their DR site ready to restart business operations if there
is an outage at the primary data center. This may require the maintenance of a
complete set of IT resources at the DR site that matches the IT resources at the
primary site. Organization can either build their own DR site, or they can use cloud
to build DR site.
Fault tolerance can be provided at the software level, or at the hardware level, or by
combining both of them. The fault-tolerant design can also be extended to include
multiple data centers or sites wherein redundant data centers are used to provide
site-level fault tolerance.
Business Applications
• Business applications run on compute systems102.
• Various types of business applications are enterprise resource planning (ERP)
applications, customer relationship management (CRM) applications, email
applications, ecommerce applications, database applications, and analytic
applications.
• A business application commonly provides a user interface such as a command
line interface (CLI) and graphical user interface (GUI).
− The user interface enables users to send requests and view responses.
− Also provide an application programming interface (API)103 that enables
other applications to interact with it.
• The protection applications and storage leverage these interfaces to track the
application data as it changes and also track the protection status of the data.
102Execute the requests from users or clients and pass back the generated
responses.
103Provides a flexible, easy-to-use means for integrating protection tools with the
business applications.
Hypervisors
• From a hypervisor’s perspective, each VM is a discrete set of files that store the
VM configuration, VM memory content, and guest OS and application data.
− Availability of these files is the key to run the VMs and continue business
operations. Therefore, protection of VMs should be included in the data
protection plan.
• Protection at the hypervisor level requires the hypervisor to function as the
source of all VM files managed by it.
Virtual Machine
• A VM does not have direct access to the hardware of the physical compute
system (host machine) on which it is created.
− The hypervisor translates the VM’s resource requests and maps the virtual
hardware of the VM to the hardware of the physical compute system.
− For example, a VM’s I/O requests to a virtual disk drive are translated by the
hypervisor and mapped to a file on the physical compute system’s disk drive.
A VM can be configured with one or more virtual CPUs. When a VM starts, its
virtual CPUs are scheduled by the hypervisor to run on the physical CPUs. Virtual
RAM is the amount of physical memory allocated to a VM and it can be configured
based on the requirements.
The virtual disk stores the VM’s OS, program files, and application data. A virtual
network adapter provides connectivity between VMs running on the same or
different compute systems, and between a VM and the physical compute systems.
Virtual optical drives and floppy drives can be configured to connect to either the
physical devices or to the image files, such as ISO and floppy images (.flp), on the
storage. SCSI/IDE virtual controllers provide a way for the VMs to connect to the
storage devices.
The virtual USB controller is used to connect to a physical USB controller and to
access the connected USB devices. Serial and parallel ports provide an interface
for connecting peripherals to the VM.
Containers
• Multiple containers can run on the same machine and share the Operating
System Kernel with other containers.
− For example, you might have one container on a system running Red Hat
Linux, serving a database, through a virtual network to another container
running Ubuntu Linux, running a web server that talks to that database, and
that web server might also be talking to a caching server that runs in a
SUSE Linux based container.
• Containers are lightweight in nature but running them in production environment
can quickly become a massive effort. Especially when used with microservices,
a containerized application might be translated into multiple containers. This can
introduce significant complexity if managed manually.
An entire storage system or some of its storage drives that store business
application data can also be the primary storage device. In addition to transactional
data, a primary storage device may also store OS and application software.
Fault Isolation
The example shown image represents two I/O paths between a compute system
and a storage system.
The compute system uses both the paths to send I/O requests to the storage
system. If an error or fault occurs on a path causing a path failure, the fault isolation
mechanism present in the environment automatically detects the failed path. It
isolates the failed path from the set of available paths and marks it as a dead path
to avoid sending the pending I/Os through it.
All pending I/Os are redirected to the live path. This helps avoiding the time-out and
the retry delays.
Compute Clustering
Compute clustering provides continuous availability of services even when a virtual
machine (VM), physical compute system, OS, or hypervisor fails. In compute
clustering technique, at least two compute systems or hypervisors work together
and are viewed as a single compute system to provide high availability and load
balancing. If one of the compute systems in a cluster fails, the service running on
the failed compute system moves to another compute system in the cluster to
minimize or avoid outage. Clustering uses a heartbeat mechanism to determine the
health of each compute system in the cluster. The exchange of heartbeat signals,
usually happens over a private network, allows participating cluster members to
monitor each other’s status. Clustering can be implemented among multiple
physical compute systems, or multiple VMs, or VM and physical compute system,
or multiple hypervisors.
Link Aggregation
Link aggregation combines two or more parallel interswitch links (ISLs) into a single
logical ISL, called a link aggregation group. It optimizes network performance by
distributing network traffic across the shared bandwidth of all the ISLs in a link
aggregation group. This allows the network traffic for a pair of node (compute
system and storage system) ports to flow through all the available ISLs in the group
rather than restricting the traffic to a specific, potentially congested ISL. The
number of ISLs in a link aggregation group can be scaled depending on
application’s performance requirement.
Link aggregation also enables network traffic failover in the event of a link failure. If
a link in a link aggregation group is lost, all network traffic on that link is
redistributed across the remaining links.
By combining ISLs, link aggregation also provides higher throughput than a single
ISL could provide. For example, the aggregation of three ISLs into a link
aggregation group provides up to 48 Gb/s throughput assuming the bandwidth of
an ISL is 16 Gb/s.
Multipathing
Enables a compute system to use multiple paths for transferring data to a storage
device on a storage system.
Can also perform load balancing by distributing I/Os across all available paths. The
figure on the slide shows a configuration where four paths between a compute
system (with dual-port HBAs) and a storage device enable multipathing.
RAID
Disk and solid state drives are susceptible to failures. A drive failure may result in
data loss. Today, a single storage system may support thousands of drives.
Greater the number of drives in a storage system, the greater is the probability of a
drive failure in the system.
RAID protects against data loss when a drive fails, through the use of redundant
drives and parity. Typically, in a RAID storage system, the data is distributed across
physical drives and these set of physical drives are viewed as single logical drive or
volume by operating system. RAID also helps in improving the storage system
performance as read and write operations are served simultaneously from multiple
drives.
Graceful Degradation
Graceful degradation of application functionality refers to the ability of an
application to maintain limited functionality even when some of the components,
modules, or supporting services are not available.
However, in this same scenario, it is possible that the product catalog module can
still be available to consumers to view the product catalog. The application can also
allow the consumers to place the order and move it into the shopping cart. This
provides the ability to process the orders when the payment gateway is available or
after failing over to a secondary payment gateway.
When applications use other services, errors can occur because of temporary
conditions such as intermittent service, infrastructure-level faults, or network issues.
Very often this form of problem can be solved by retrying the operation a few
milliseconds later, and the operation may succeed. The simplest form of transient
fault handling is to implement this retry logic in the application itself.
A retry strategy must also be defined to state how many retries can be attempted
before deciding that the fault is not transient and define what the intervals should
be between the retries. The logic will typically attempt to execute the action(s) a
certain number of times, registering an error, and utilizing a secondary service if the
fault continues.
In a persistent state model, the state information is stored out of the memory and is
usually stored in a repository (database). If a compute system (server) running the
application instance fails, the state information will still be available in the
repository.
A new application instance is created on another server which can access the state
information from the database and resume the processing.
Database Rollback
A rollback is the operation of restoring a database to a previous state by canceling
a specific transaction or transaction set. Rollbacks are important for database
integrity because they mean that the database can be restored to a consistent
previous state even after erroneous operations are performed.
Thus, a rollback occurs when a user begins to change data and realizes that the
wrong record is being updated and then cancels the operation to undo any pending
changes. Rollbacks may also be issued automatically after a server or database
crashes, e.g. after a sudden power loss. When the database restarts, all logged
transactions are reviewed; then all pending transactions are rolled back, allowing
users to reenter and save appropriate changes.
In the example shown in the image, transactions A,B, and C are performed and
committed to the database. Then, transactions D and E are performed and an issue
is identified. In such case, transactions D and E should be rolled back. After the
database is rolled back, transactions D and E are cancelled and database is
restored to the previous state with only committed data.
Recent world events including acts of terrorism, natural disasters, and large-scale
company fraud have resulted in a new raft of legislation designed to protect
company data from loss or corruption. Many government and regulatory laws
mandate that an organization must be responsible for protecting its employee’s and
customer’s personal data.
Backup Operations
The backup server initiates the backup process for different clients based on the
backup schedule configured for them. For example, the backup for a group of
clients may be scheduled to start at 3:00 a.m. every day. The backup server
coordinates the backup process with all the components in a backup environment.
The backup server maintains the information about backup clients to be backed up
and storage nodes to be used in a backup operation. The backup server retrieves
the backup-related information from the backup catalog. Based on this information,
the backup server instructs the storage node to load the appropriate backup media
into the backup devices.
The storage node then sends additional backup metadata (location of the data on
the backup device, time of backup, and so on) to the backup server. The backup
server updates the backup catalog with this information. The backup data from the
client can be sent to the backup device over a LAN or a SAN network.
Backup mode: Hot backup and cold backup are the two modes deployed for
backup. They are based on the state of the application when the backup is
performed. A cold backup requires the application to be shutdown during the
backup process. Hence, this method is also referred to as offline backup. The
disadvantage of a cold backup is that the application is inaccessible to users during
the backup process. In a hot backup, the application is up-and-running, with users
accessing their data during the backup process. This method of backup is also
referred to as online backup. The hot backup of online production data is
challenging because data is actively being used and changed. If a file is open, it is
normally not backed up during the backup process. In such situations, an open file
agent is required to back up the open file. These agents interact directly with the
operating system or application and enable the creation of consistent copies of
open files.
Recovery Operations
Upon receiving a restore request, an administrator opens the restore application to
view the list of clients that have been backed up. While selecting the client for
which a restore request has been made, the administrator also needs to identify the
client that will receive the restored data. Data can be restored on the same client
for whom the restore request has been made or on any other client.
The administrator then selects the data to be restored and the specific point-in-time
to which the data has to be restored based on the RPO. Because all this
information comes from the backup catalog, the restore application needs to
communicate with the backup server.
The backup server instructs the appropriate storage node to mount the specific
backup media onto the backup device. Data is then read and sent to the client that
has been identified to receive the restored data.
This ensures consistency for the restored data. In these cases, the RTO is
extended due to the additional steps in the restore operation. It is also important to
have security mechanisms on the backup and recovery applications to avoid
recovery of data by non-authorized users.
Types of Recovery
Operational recovery or restore typically involves the recovery of individual files
or directories after they have been accidentally deleted or corrupted.
Disaster recovery involves bringing a data center or a large part of a data center
to an operational state in case of a disaster affecting the production site location.
Data for recovery are located in offsite locations. Portable media, such as tapes,
sent to an offsite location could be used for recovery. In another example, data
backed up locally can be replicated to an offsite location by the backup application.
Recovery can be from the most recent point-in-time replicated backup data.
Full VM recovery permanently restore your VMs either to the same host or to a
different virtual host, it can be done through the Live Recovery to ESXi Server
option. The VMs will be restored into the data store that is present in the storage
repositories.
Cloud disaster recovery (Cloud DR) allows enterprises to copy backed-up VMs
from their on-premises environments to the public cloud to orchestrate DR testing,
failover and failback of cloud workloads in a disaster recovery scenario. These
workloads can be run directly in the public cloud, so full deployment of your on-
premises data protection solutions in the cloud is not required in order to protect
and recover your VMs. Organizations can manage, recover, failback and test DR
plans through the Cloud DR Server (CDRS) UI. Cloud DR takes advantage of the
agility and cost-effectiveness of cloud object storage (Dell EMC ECS, AWS S3 or
Azure Blob), requires minimal footprint in the public cloud, as well as minimal
compute cycles, delivering a highly efficient disaster recovery solution.
In case of file systems, consistency can be achieved by taking the file system
offline, i.e. by un-mounting the file system or by keeping the file system online and
flushing host buffers before creating the backup to ensure that all writes are
committed. No further writes are allowed to the data while the backup is being
created.
Backing up data while files are open becomes more challenging because data is
actively being used and changed. An open file is locked by the operating system
and is not copied during the backup process until the user closes it. The backup
application can back up open files by retrying the operation on files that were
opened earlier.
During the backup process, it may be possible that files opened earlier will be
closed and a retry will be successful. However, this method is not considered
robust because in some environments certain files are always open. In such
situations, the backup application or the operating system can provide open file
agents. These agents interact directly with the operating system and enable the
creation of copies of open files.
A database is composed of different files which may occupy several file systems.
Data in one file may be dependent upon data in another. A single transaction may
cause updates to several files and these updates may need to occur in a defined
order. A consistent backup of a database means that all files need to be backed up
at the same “point” or state. Consistent backups of databases can be done using a
cold (or offline) method which means that the database is shutdown while the
backup is running.
The downside is that the database will not be accessible by users. Hot backup is
used in situations where it is not possible to shutdown the database. Backup is
facilitated by database backup agents that can perform a backup while the
database is active. The disadvantage associated with a hot database backup is
that the agents can negatively affect the performance of the database application
server.
All subsequent increments use the created synthetic full backup as a new starting
point. A previously used full backup file remains on backup device until it is
automatically deleted according to the backup retention policy.
A synthetic full backup enables a full backup copy to be created offline without
disrupting the I/O operation on the production volume. This also frees up network
resources from the backup process, making them available for other production
uses. Synthetic backups enable to take advantage of reduced backup window.
Backup Multiplexing
Multiplexing allows backups of multiple client machines to send data to a single
tape drive simultaneously. Multiplexing is useful when your tape drive throughput is
faster than the rate at which data can be extracted from the source (client).
Multiplexing may decrease backup time for large numbers of clients over slow
networks, but it does so at the cost of recovery time. Restores from multiplexed
tapes must pass over all non-applicable data.
Note: Multistreaming
• Multistreaming is a process that divides the backup jobs into multiple sub-jobs
(streams) that run simultaneously and sends data to the destination backup
device.
• Multistreaming allows to use all of the available backup devices on the system
by splitting the backup jobs into multiple jobs using all available tape devices.
− It will increase the overall backup throughput compared to the sequential
method.
• Multistreaming is useful when performing large backup jobs, since it is more
efficient to divide multiple jobs between multiple backup devices.
Direct-Attached Backup
Direct-attached backups are generally better suited for smaller environments. The
key advantage of direct-attached backups is speed. The tape devices can operate
at the speed of the channels.
In a direct-attached backup, the backup device is not shared, which may lead to
silos of backup device in the environment. It might be difficult to determine if
everything is being backed up properly.
As the environment grows, however, there will be a need for central management
of all backup devices and to share the resources to optimize costs. An appropriate
solution is to share the backup devices among multiple servers.
LAN-Based Backup
In a LAN-based backup, the data to be backed up is transferred from the backup
client (source), to the backup device (destination) over the LAN, which may affect
network performance.
Streaming across the LAN also affects network performance of all systems
connected to the same segment as the backup server.
Network resources are severely constrained when multiple clients access and
share the same backup device. This impact can be minimized by adopting a
number of measures such as configuring separate networks for backup and
installing dedicated storage nodes for some application servers.
This approach doesn’t provide the ability to backup and restore the VM as a whole.
The agent running on the compute system consumes CPU cycles and memory
resources.
This may impact the performance of the services or applications running on the
VMs. To overcome these challenges, the backup process can be offloaded from
the VMs to a proxy server. This can be achieved by using the image-based backup
approach.
Image-Based Backup
Image-level backup makes a copy of the virtual machine disk and configuration
associated with a particular VM. The backup is saved as a single entity called VM
image. This type of backup is suitable for restoring an entire VM in the event of a
hardware failure or human error such as the accidental deletion of the VM. It is also
possible to restore individual files and folders/directories within a virtual machine.
In an image-level backup, the backup software can backup VMs without installing
backup agents inside the VMs or at the hypervisor-level. The backup processing is
performed by a proxy server that acts as the backup client, thereby offloading the
backup processing from the VMs.
The proxy server then performs backup by using the snapshot. Performing an
image-level backup of a virtual machine disk provides the ability to execute a bare
metal restore of a VM.
Given the scalability and sheer explosion in the size of virtualized and cloud
environments, the workload burden placed on one proxy server can quickly be built.
In this scenario, the recommendation is to provision multiple proxies to handle the
combined workload and increase the amount of parallelism.
This enables the backup application to backup only the blocks that have changed,
rather than backing up every block. If changed block tracking is enabled for a VM
disk, the virtual machine kernel will create an additional file where it stores a map of
all the VM disk's blocks.
Once a block is changed it will be recorded in this map file. This way the kernel can
easily communicate to a backup application about the blocks of a file that have
changed since a certain point-in-time.
The backup application can then perform a backup by copying only these changed
blocks. Changed block tracking technique dramatically reduces the amount of data
to be copied before additional data reduction technologies (deduplication) are
applied. It also reduces the backup windows and the amount of storage required for
protecting VMs.
This technique reduces the recovery time (RTO) compared to full image restores by
only restoring the delta of the changed VM blocks. During a restore process, it is
determined which blocks have changed since the last backup. For example, if a
large database is corrupted, a changed block recovery would just restore the parts
of the database that has changed since the last backup was made.
Recovery-in-Place
Recovery-in-place (Instant VM recovery) is a term that refers to running a VM
directly from the purpose-built backup appliance, using a backed up copy of the VM
image instead of restoring that image file to the production system. In the
meantime, the VM data is restored to the primary storage from the backup copy.
Once the recovery has been completed, the workload is redirected to the original
VM.
A large number of operating systems, application files and data files are common
across multiple systems in a data center environment. Identical files such as Word
documents, PowerPoint presentations and Excel spreadsheets, are stored by many
users across an environment. Backups of these systems will contain a large
number of identical files. Additionally, many users keep multiple versions of files
that they are currently working on. Many of these files differ only slightly from other
versions, but are seen by backup applications as new data that must be protected.
Due to this redundant data, the organizations are facing many challenges. Backing
up redundant data increases the amount of storage needed to protect the data and
subsequently increases the storage infrastructure cost. It is important for
organizations to protect the data within the limited budget. Organizations are
running out of backup window time and facing difficulties meeting recovery
objectives. Backing up large amount of duplicate data at the remote site or cloud for
DR purpose is also very cumbersome and requires lots of bandwidth.
of detecting and identifying the unique data segments (chunk) within a given set of
data to eliminate redundancy. Only one copy of the data is stored; the subsequent
copies are replaced with a pointer to the original data.
•Retention period: This is the period of time that defines how long the backup
copies are retained. The longer the retention, the greater is the chance of identical
data existence in the backup set which would increase the deduplication ratio and
storage space savings.
•Frequency of full backup: As more full backups are performed, it increases the
amount of same data being repeatedly backed up. So it results in high
deduplication ratio.
•Change rate: This is the rate at which the data received from the backup
application changes from backup to backup. Client data with a few changes
between backups produces higher deduplication ratios.
File-level deduplication (also called single instance storage) detects and removes
redundant copies of identical files in a backup environment. File-level deduplication
compares a file to be backed up with those already stored by checking its attributes
against an index. If the file is unique, it is stored and the index is updated; if not,
only a pointer to the existing file is stored. The result is that only one instance of the
file is saved and the subsequent copies are replaced with a pointer that points to
the original file. Indexes for file-level deduplication are significantly smaller, which
takes less computational time when duplicates are being determined. Backup
performance is, therefore, less affected by the deduplication process. File-level
deduplication is simple but does not address the problem of duplicate content
inside the files. A change in any part of a file results in classifying that as a new file
and saving it as a separate copy as shown in the figure. Typically, the file-level
deduplication is implemented in a NAS environment.
Fixed-length block deduplication fixes the chunking at a specific size, for example 8
KB or maybe 64 KB. The difference is that the smaller the chunk, the more likely is
the opportunity to identify it as redundant and results into greater reductions.
However, fixed-length block deduplication has challenge when a data is inserted or
deleted from a file. Inserting or deleting data causes a shift in all the data after the
point of insertion or deletion. This causes all the blocks after that point to be
different. The data is the same, but the blocks get cut at different points. So a small
insertion of data near the beginning of a file can cause the entire file to be backed
up and stored again.
Organizations with fast data growth, highly virtualized environments, and remote
offices greatly benefit from variable-length deduplication over a fixed-block
approach. Variable-length deduplication reduces backup storage and, when
performed at the client, also reduces network traffic, making it ideal for remote
backup.
If there is no match on the server, the client will send the hash and the
corresponding chunk to the deduplication server to store the backup data. If the
chunk has already been backed up, then the chunk will not be sent to the
deduplication server by the client, which ensures that the redundant backup data is
eliminated at the client. The deduplication server can be deployed in different ways.
The deduplication server software can be installed on a general purpose physical
server (as shown in the figure) or on VMs. Some vendors offer deduplication server
along with backup device as an appliance.
The deduplication server would support encryption for secure backup data
transmission and also would support replication for disaster recovery purpose.
Source-based deduplication reduces the amount of data that is transmitted over a
network from the source to the backup device, thus requiring less network
bandwidth. There is also a substantial reduction in the capacity required to store
the backup data. Backing up only unique data from clients reduces the backup
window. However, a deduplication agent running on the client may impact the
backup performance, especially when a large amount of data needs to be backed
up. When an image-level backup is implemented, the backup workload is moved to
a proxy server.
In post-processing deduplication, the backup data is first stored to the disk in its
native backup format and deduplicated after the backup is completed. In this
approach, the deduplication process is separated from the backup process and the
deduplication happens outside the backup window. However, the full backup data
set is transmitted across the network to the storage target before the redundancies
are eliminated. So, this approach requires adequate storage capacity and network
bandwidth to accommodate the full backup data set. Organizations can consider
implementing target-based deduplication when their backup application would not
have built-in deduplication capabilities.
Data Replication
Data is one of the most valuable assets of any organization. It is being stored,
mined, transformed, and utilized continuously. It is a critical component in the
operation and function of organizations. Outages, whatever may be the cause, are
extremely costly, and customers are concerned about data availability at all times.
Safeguarding and keeping the data highly available are some of the top priorities of
any organization. To avoid disruptions in business operations, it is necessary to
implement data protection technologies in a data center.
Replica Consistency
Consistency is a primary requirement to ensure the usability of replica device. In
case of file systems (FS), consistency can be achieved either by taking FS offline
i.e. by un-mounting FS or by keeping FS online by flushing compute system buffers
before creating replica.
File systems buffer the data in the compute system memory to improve the
application response time. Compute system memory buffers must be flushed to the
disks to ensure data consistency on the replica, prior to its creation. If the memory
buffers are not flushed to the disk, the data on the replica will not contain the
information that was buffered in the compute system.
Types of Replication
• Local replication helps to
− Replicate data within the same storage system (in case of remote
replication) or the same data center (in case of local replication).
− Restore the data in the event of data loss or enables restarting the
application immediately to ensure business continuity. Local replication can
be implemented at compute, storage, and network.
• Remote replication helps to
VM Clone
• When the cloning operation completes, the clone becomes a separate VM. The
changes made to a clone do not affect the parent VM. Changes made to the
parent VM do not appear in a clone.
• Installing a guest OS and applications on a VM is a time consuming task. With
clones, administrators can make many copies of a virtual machine from a single
installation and configuration process.
− For example, in an organization, the administrator can clone a VM for each
new employee, with a suite of preconfigured software applications.
• Snapshot is used to save the current state of the virtual machine, so that it can
allow to revert to that state in case of any error. But clone is used when a copy
of a VM is required for separate use.
− A full clone is an independent copy of a VM that shares nothing with the
parent VM. Because a full clone needs to have its own independent copy of
the virtual disks, the cloning process may take a relatively longer time.
− A linked clone is made from a snapshot of the parent VM. The snapshot is
given a separate network identity and assigned to the hypervisor to run as
an independent VM.
• All files available on the parent at the moment of the snapshot creation continue
to remain available to the linked clone VM in read-only mode.
− The ongoing changes (writes) to the virtual disk of the parent do not affect
the linked clone and the changes to the virtual disk of the linked clone do not
affect the parent. All the writes by the linked clone are captured in a delta
disk.
Snapshot – RoW
• Redirects new writes destined for the source LUN to a reserved LUN in the
storage pool.
• Replica (snapshot) still points to the source LUN.
− All reads from the replica are served from the source LUN.
To learn more about snapshot-RoW, click here.
Multi-Site Replication
• Multi-site replication mitigates the risks identified in two-site replication. In a
multi-site replication, data from the source site is replicated to two or more
remote sites. The example shown in the figure is a three-site remote replication
solution.
• In this approach, data at the source is replicated to two different storage
systems at two different sites. The source to remote site 1 (target 1) replication
is synchronous with a near-zero RPO. The source to remote site 2 (target 2)
replication is asynchronous with an RPO in the order of minutes.
− At any given instant, the data at the remote site 1 and the source is identical.
The data at the remote site 2 is behind the data at the source and the
remote site 1.
− The replication network links between the remote sites will be in place but
not in use.
− The difference in the data between the remote sites is tracked so that if a
source site disaster occurs, operations can be resumed at the remote site 1
or the remote site 2 with incremental resynchronization between these two
sites.
• The key benefit of this replication is the ability to failover to either of the two
remote sites in the case of source site failure, with disaster recovery
(asynchronous) protection between the remote sites.
Product Description
Thirdly, data that must be preserved over a long period for compliance reasons
may be modified or deleted by the users.
This poses a risk of a compliance breach. Finally, the backup of high-growth fixed
data results in an increased backup window and related backup storage cost. Data
archiving addresses these challenges.
Archiving fixed data before taking backup helps to reduce the backup window and
backup storage acquisition costs. Data archiving helps in preserving data that may
be needed for future reference and data that must be retained for regulatory
compliance. For example, new product innovation can be fostered if engineers can
access archived project materials such as designs, test results, and requirement
documents.
Similarly, both active and archived data can help data scientists drive new
innovations or help to improve current business processes. In addition, government
regulations and legal/contractual obligations mandate organizations to retain their
data for an extended period of time.
After the files are identified for archiving, the archive server creates the index for
the files. Once the files have been indexed, they are moved to the archive storage
and small stub files are left on the primary storage. Each archived file on primary
storage is replaced with a stub file. The stub file contains the address of the
archived file. As the size of the stub file is small, it significantly saves space on
primary storage.
From a client’s perspective, the data movement from primary storage to secondary
storage is completely transparent.
As the tier number reduces, the storage performance improves but the cost of
storage increases which limits the usage of storage capacity. The higher the tier
number, the higher can be the storage capacity due to its cost advantage.
Archive storage is typically configured as the final tier or highest numbered tier in
the storage tiering. Keeping frequently used data in lowered numbered tiers, called
performance tiers, improves application performance.
Moving less-frequently accessed data or fixed data to the highest numbered tier,
called the archive tier, can free up storage space in performance tiers and reduce
the cost of storage.
Before moving a file to archive storage, the policy engine scans the NAS device to
identify files that meet the predefined tiering policies. After identifying the candidate
files, the policy engine creates stub files on the NAS device and then moves the
candidate files to the destination archive storage.
The small, space-saving stub files point to the actual files in the archive storage.
When an application server (NAS client) tries to access a file from its original
location on the NAS device, the actual file is provided from the archive storage.
Content integrity: It provides assurance that the stored data has not been altered.
If the fixed data is altered, CAS generates a new content address for the altered
data, rather than overwriting the original fixed data.
NAS-based data migration involves migrating data at the file-level between NAS
systems. Even the file migration between NAS systems can happen by using
intermediary compute systems or virtualization appliance.
The best solution in one migration may not necessarily be the best solution for
another migration. No one-size-fits-all migration tool or solution exists. Each
migration solution has its own set of advantages and challenges. So it is important
to choose an appropriate solution to successfully perform migration operation.
In a NAS to NAS direct data migration, file-level data is migrated from one NAS
system to another directly over the LAN without the involvement of any external
server. The two primary options of performing NAS-based migration is either by
using NDMP protocol or software tool. In this example, the new NAS system
initiates the migration operation and pulls the data directly from the old NAS system
over the LAN. The key advantage of NAS to NAS direct data migration is that there
is no need for an external component (host or appliance) to perform or initiate the
migration process.
Application migration typically involves moving the application from one data center
environment to another. Typically, the organization can move the application from
physical to virtual environment. In a virtualized environment, the application can
also be moved from one hypervisor to another for various business reasons such
as balancing workload for improving performance and availability. In an application
migration from a physical to virtual environment, the physical server running the
application is converted into a virtual machine. This option usually requires a
converter software that clones the data on the hard disk of the physical compute
system and migrates the disk content (application, OS, and data) to an empty VM.
Hybrid Migration Strategy: In this strategy, some parts of the application are
moved to the new environment while leaving the other parts of the application in
place. Rather than moving the entire application at once, parts of it can be moved
and optimized, one at a time. This strategy is good for large systems that involve
several applications and those that are not tightly coupled.
Software Controller
• The control plane in software-defined data center is implemented by a software
controller. The controller is a software that:
• Discovers the available underlying resources and provides an aggregated view
of resources. It abstracts the underlying hardware resources (compute, storage,
and network) and pools them.
− This enables the rapid provisioning of resources from the pool, based on
pre-defined policies that align to the service level agreements for different
users.
− Enables storage management and provisioning.
• Enables organizations to dynamically, uniformly, and easily modify and manage
their infrastructure.
• Enables an administrator to manage the resources, node connectivity, and
traffic flow. It also controls the behavior of underlying components, allows
applying policies uniformly across the infrastructure components, and enforces
security, all from a software interface.
• Provides interfaces that enable software external to the controller to request
resources and access these resources as services.
• CLI and GUI are native management interfaces of the controller. API is used by
external software to interact with controller.
Architecture of SDDC
• The SDDC architecture decouples the control plane from the data plane.
− It separates the control functions from the underlying infrastructure
components and provides it to an external software controller.
− The centralized control plane provides policies for processing and
transmission of data, which can be uniformly applied across the multi-vendor
infrastructure components.
− The policies can also be upgraded centrally to add new features and to
address application requirements.
• The controller usually provides CLI and GUI for administrators to manage the IT
infrastructure and configure the policies. It also automates and orchestrates
many hardware-based or component-specific management operations.
− This reduces the need for manual operations that are repetitive, error-prone,
and time-consuming.
• The software controller provides APIs for external management tools and
orchestrators to manage data center infrastructure and orchestrate controller
operations.
• The SDDC architecture enables users to view and access IT resources as a
service from a self-service portal.
− The portal provides a service catalog that lists a standardized set of services
available to the users.
• The service catalog allows a user to request or order a service from the catalog
in a self-service way.
− Allows the end user to specify a compute system for which a virtual storage
must be provisioned and a virtual storage system and virtual storage pool
from which the storage has to be derived.
− Automates the storage provisioning tasks and delivers virtual storage
resources based on the requested services.
105API enables the external management tools and applications to interact with the
SDS controller for extracting data, monitoring SDS environment, and creating
logical storage resources.
Web DB Clients
App Server
Physical
Switch
The cloud model is similar to utility services such as electricity, water, and
telephone. When consumers use these utilities, they are typically unaware of how
the utilities are generated or distributed. The consumers periodically pay for the
utilities based on usage. Consumers simply hire IT resources as services from the
cloud without the risks and costs associated with owning the resources.
Cloud services are accessed from different types of client devices over wired and
wireless network connections. Consumers pay only for the services that they use,
either based on a subscription or based on resource consumption. The figure on
the slide illustrates a generic cloud computing environment, wherein various types
of cloud services are accessed by consumers from different client devices over
different network types.
As a result, they may take up a considerable amount of time. This can delay
operations and increase the time-to-market. Additionally, to the extent allowed by
the budget, the IT resources required for an application are sized based on peak
usage. This results in incurring high up-front capital expenditure (CAPEX) even
though the resources remain underutilized for a majority of the time.
As workloads continue to grow and new technologies emerge, businesses may not
afford for investments to increase proportionally. Further, a significant portion of the
IT budget goes to support and maintain the existing IT infrastructure, leaving a little
to provide innovative solutions to the business.
(CAPEX) rises, the risk associated with the investment also increases. For small
and medium sized businesses, this may be a big challenge, which eventually
restricts their ability to grow. As an individual, it may not be sensible or affordable to
purchase new applications every time, if you need them only for a brief period. This
image shows various requirements and constraints from a business perspective as
well as an individual perspective and also shows the way a cloud can address
these constraints and requirements.
Public Cloud
A cloud infrastructure deployed by a provider to offer cloud services to the general
public and/or organizations over the Internet.
There may be multiple tenants (consumers) who share common cloud resources. A
provider typically has default service levels for all consumers of the public cloud.
The provider may migrate a consumer’s workload at any time, to any location.
Private Cloud
Many organizations may not wish to adopt public clouds as they are accessed over
the open Internet and used by the general public. With a public cloud, an
organization may have concerns related to privacy, external threats, and lack of
control over the IT resources and data.
Since the costs are shared by lesser consumers than in a public cloud, this option
may be more expensive. However, a community cloud may offer a higher level of
control and protection against external threats than a public cloud. There are two
variants of a community cloud: on-premise and externally-hosted. In an on-premise
community cloud, one or more participant organizations provide cloud services that
are consumed by the community.
Many network configurations are possible in a community cloud. The figure on the
slide illustrates an on-premise community cloud, the services of which are
consumed by enterprises P, Q, and R. The community cloud comprises two cloud
infrastructures that are deployed on the premises of Enterprise P and Enterprise Q,
and combined to form a community cloud.
The cloud infrastructure is hosted on the premises of the external cloud service
provider and not within the premises of any of the participant organizations.
The provider:
Using an external provider’s cloud infrastructure for the community cloud may offer
access to a larger pool of resources as compared to an on-premise community
cloud.
Hybrid Cloud
Is composed of two or more individual clouds, each of which can be private,
community, or public clouds. There can be several possible compositions of a
hybrid cloud as each constituent cloud may be of one of the five variants as
discussed previously.
May change over the period of time as component clouds join and leave. In a
hybrid cloud environment, the component clouds are combined through the use of
open or proprietary technology such as interoperable standards, architectures,
protocols, data formats, application programming interfaces (APIs), and so on.
Enables the consumers to rent any required IT resources based on the pay-per-use
or subscription pricing. This reduces a consumer’s IT capital expenditure as
investment is required only for the resources needed to access the cloud services
Has the ability to ensure availability at varying levels, depending on the provider’s
policy towards service availability. Redundant infrastructure components enable
fault tolerance for cloud deployments.
Data in a cloud can be broken into small pieces and distributed across a large
cluster of nodes in such a manner that an entire data set can be reconstructed
even if there is failure of individual nodes.
A service availability zone is a location with its own set of resources. Each zone is
isolated from the other zone so that a failure in one zone does not impact the other.
If a service is distributed among several zones, consumers of that service can fail
over to other zones in the event of a zone failure.
Applications and data reside centrally and can be accessed from anywhere over a
network from any device such as desktop, mobile, and thin client.
Backup as a Service
• Enables organizations to procure backup services on-demand in the cloud.
Organizations can build their own cloud infrastructure and provide backup
services on demand to their employees/users. Some organizations prefer hybrid
cloud option for their backup strategy, keeping a local backup copy in their
private cloud and using public cloud for keeping their remote copy for DR
purpose. For providing backup as a service, the organizations and service
providers should have necessary backup technologies in place in order to meet
the required service levels.
• Enables individual consumers or organizations to reduce their backup
management overhead. It also enables the individual consumer/user to perform
backup and recovery anytime, from anywhere, using a network connection.
Consumers do not need to invest in capital equipment in order to implement and
manage their backup infrastructure. These infrastructure resources are rented
without obtaining ownership of the resources.
• Backups can be scheduled and infrastructure resources can be allocated with a
metering service. This will help to monitor and report resource consumption.
Many organizations’ remote and branch offices have limited or no backup in
place. Mobile workers represent a particular risk because of the increased
possibility of lost or stolen devices.
• Ensures regular and automated backup of data. Cloud computing gives
consumers the flexibility to select a backup technology, based on their
requirement, and quickly move to a different technology when their backup
requirement changes.
To perform backup to the cloud, typically the cloud backup agent software is
installed on the servers that need to be backed up. After installation, this software
establishes a connection between the server and the cloud where the data will be
stored.
The backup data transferred between the server and the cloud is typically
encrypted to make the data unreadable to an unauthorized person or system.
Replication to Cloud
Cloud-based replication helps organizations to mitigate the risk associated with
outages at the consumer production data center. Organization of all levels are
looking for the cloud to be a part of the business continuity. Replicating application
data and VM to the cloud enable organization to restart the application from the
cloud and also allow to restore the data from any location.
Data and the VM replicated to the cloud is hardware independent; this further
reduces the recovery time.
Having a DR site in the cloud reduces the need for data center space and IT
infrastructure, which leads to significant cost reduction, and eliminates the need for
upfront capital expenditure. Resources at the service provider can be dedicated to
the consumer or they can be shared. The service provider should design,
implement, and document a DRaaS solution specific to the customer’s
infrastructure.
They must conduct an initial recovery test with the consumer to validate complete
understanding of the requirements and documentation of the correct, expected
recovery procedures.
Typically VMs are allocated from a pool of compute resources located in the
provider’s location. Returning business operations back to the consumer’s
production environment is referred to as failback. This requires replicating the
updated data from the cloud repository back to the in-house production system
before resuming the normal business operations at consumer’s location.
Enables storing the right data to the right tier, based on service level requirements,
at a minimal cost. Each tier has different levels of protection, performance, and
cost.
For example, high performance solid-state drives (SSDs) can be configured as tier
1 storage to keep the frequently accessed data, lower cost HDDs as tier 2 storage
to keep the less frequently accessed data, and cloud as tier 3 storage to keep the
rarely used data.
Provides a local cache to reduce the latency associated with having the storage
capacity far away from the data center.
Compliance
• Internal policy compliance controls the nature of IT operations within an
organization. This requires clear assessment of the potential difficulties in
maintaining the compliance and processes to ensure that this is effectively
achieved.
• External policy compliance includes legal requirements, legislation, and
industry regulations. These external compliance policies control the nature of IT
operations related to the flow of data out of an organization.
− They may differ, based upon the type of information (for example, source
code versus employee records), and business (for example, medical
services versus financial services).
• Compliance management ensures that an organization adheres to relevant
policies and legal requirements. Policies and regulations can be based on
configuration best practices and security rules.
− These include administrator roles and responsibilities, physical infrastructure
maintenance timelines, information backup schedules, and change control
processes.
• Ensuring CIA and GRC are the primary objectives of any IT security
implementation.
106
Multi-tenancy is achieved by using mechanisms that provide separation of
computing resources such as memory and storage for each user.
107At the compute system level, security mechanisms are deployed to secure
hypervisors and hypervisor management systems, virtual machines, guest
operating systems, and applications.
109At the storage level, security mechanisms include LUN masking, data
shredding, and data encryption.
o The IAM controls prompt for the user’s credentials. Depending on the type of
IAM control deployed in this environment the user provides the necessary
credentials.
o Credentials are then verified against a system that has the ability to
authenticate and authorize the user.
o Upon successfully verifying the credentials, the authorized user is granted
access to the IT resources.
Firewall-Demilitarized Zone
• A demilitarized zone is a control to secure internal assets while allowing
Internet-based access to selected resources.
• In a demilitarized zone environment, servers that need Internet access are
placed between two sets of firewall.
• Servers in the demilitarized zone may or may not be allowed to
communicate with internal resources.
• Application-specific ports such as those designated for HTTP or FTP traffic are
allowed through the firewall to the demilitarized zone servers.
• No Internet-based traffic is allowed to go through the second set of firewall and
gain access to the internal network.
VLAN Example
• Consider the example with three VLANs: VLAN 10, VLAN 20, and VLAN 30.
VSAN Example
• Consider the example with two VSANs: VSAN 10 and VSAN 20.
Types of Zoning
• WWN zoning: It uses World Wide Names to define zones. The zone members
are the unique WWN addresses of the FC HBA and its targets (storage
systems).
− A major advantage of WWN zoning is its flexibility. If an administrator moves
a node to another SAN switch port, the node will maintain connectivity to its
zone partners without modifying the zone configuration. This is possible
because WWN is static to the node port.
− WWN zoning could run the risk of WWN spoofing, enabling a host to gain
access to resources from another host. Switches protect this by reviewing
WWN and FCID of the host match.
• Port zoning: It uses the switch port ID to define zones. In port zoning, access
to the node is determined by the physical switch port to which a node is
connected.
− The zone members are the port identifiers (switch domain ID and port
number) to which FC HBA and its targets (storage systems) are connected.
− If a node is moved to another switch port in the SAN, port zoning must be
modified to allow the node in its new port to participate in its original zone.
− If an FC HBA or a storage system port fails, an administrator just needs to
replace the failed device without changing the zoning configuration.
• Mixed zoning: It combines the qualities of both WWN zoning and port zoning.
Using mixed zoning enables a specific node port to be tied to the WWN of
another node.
o Maximum size of the log file. When this size is reached, the hypervisor
makes an archive copy of the log file and starts storing information in a
new log file.
o Maintain a specific number of old log files. When the configured limit is
reached, the hypervisor automatically deletes the oldest file.
Learn more about VM hardening, OS hardening and Application hardening.
Data Encryption
• Data should be encrypted as close to its origin as possible. Data encryption:
− Can be used for encrypting data at the point of entry into the storage
network.
− Can be implemented on the fabric to encrypt data between the compute
system and the storage media. These controls can protect both the data at-
rest on the destination device and the data in-transit.
− Can be deployed at the storage-level, which can encrypt data-at-rest.
• Another way to encrypt network traffic is by using cryptographic protocols such
as Transport Layer Security (TLS) which is a successor to Secure Socket Layer
(SSL).
Types of Attacks
Denial of Service
Types of Attacks
Digital Currency Mining
Types of Attacks
Spam
Types of Attacks
Adware
Types of Attacks
Malicious Web Scripts
Types of Attacks
Business Email Compromise
Types of Attacks
Banking Trojan
Types of Attacks
Ransomware
When a system is infected with ransomware malware, it asks the user to pay a fee
to unlock and reclaim the data, or else the data is lost or made public.
This example illustrates the importance of monitoring the capacity of a storage pool
in a NAS system. Monitoring tools can be configured to issue a notification when
thresholds are reached on the storage pool capacity. For example, notifications are
issued when the pool capacity reaches 66 percent and 80 percent so that the
administrator can take the right action. Proactively monitoring the storage pool can
prevent service outages caused due to lack of space in the storage pool.
This example shows a backup environment that includes three physical compute
systems—H1, H2, and H3—that host backup clients (VMs). Two SAN switches
(SW1 and SW2) connect the compute systems to a storage node and the storage
node to the backup storage system. A new compute system running backup clients
with a high workload must be deployed. The backup data from the new compute
system must be ingested through the same backup storage system port as H1, H2,
and H3. Monitoring backup storage system port utilization ensures that the new
compute system does not adversely affect the performance of the backup clients
running on other compute systems.
Here, utilization of the shared backup storage system port is shown by the solid
and dotted lines in the graph. If the port utilization prior to deploying the new
compute system is close to 100 percent, then deploying the new compute system is
not recommended because it might impact the performance of the backup clients
running on other compute systems. However, if the utilization of the port prior to
deploying the new compute system is closer to the dotted line, then there is room to
add a new compute system.
IT organizations typically comply with various data security policies that may be
specific to government regulations, organizational rules, or deployed services.
Monitoring detects all protection operations and data migration that deviate from
predefined security policies. Monitoring also detects unavailability of data and
In this example, the storage system is shared between two workgroups, WG1 and
WG2. The data of WG1 should not be accessible by WG2 and vice versa. A user
from WG1 might try to make a local replica of the data that belongs to WG2. If this
action is not monitored or recorded, it is difficult to track such a violation of security
protocols. Conversely, if this action is monitored, a notification can be sent to
prompt a corrective action or at least enable discovery as part of regular auditing
operations.
Examples of CI attribute are the CI’s name, manufacturer name, serial number,
license status, version, description of modification, location, and inventory status
(for example, on order, available, allocated, or retired). The inter-relationships
among CIs in a data protection environment commonly include service-to-user,
virtual storage pool-to-service, virtual storage system-to-virtual storage pool,
physical storage system-to-virtual storage system, and data center-to geographic
location.
All information about CIs is usually collected and stored by the discovery tools in a
single database or in multiple autonomous databases mapped into a federated
database called a configuration management system (CMS). Discovery tools also
update the CMS when new CIs are deployed or when attributes of CIs change.
CMS provides a consolidated view of CI attributes and relationships, which is used
by other management processes for their operations. For example, CMS helps the
security management process to examine the deployment of a security patch on
VMs, the problem management to resolve a remote replication issue, or the
capacity management to identify the CIs affected on expansion of a virtual storage
pool.
management team. The change management team assesses the potential risks of
the changes, prioritizes, and makes a decision on the requested changes.
The monitoring tools also help the administrators to identify the gap between the
required availability and the achieved availability. With this information, the
administrators can quickly identify errors or faults in the components that may
cause data unavailability in future. Based on the data availability requirements and
areas found for improvement, the availability management team may propose and
architect new data protection and availability solutions or changes in the existing
solutions.
For example, the availability management team may propose an NDMP backup
solution to support a data protection service or any critical business function that
requires high availability. The team may propose both component-level and site-
The example shown illustrates the resolution of a problem that impacts the
performance of a synchronous replication over a SAN recurrently. The problem is
detected by an integrated incident and problem management tool deployed in the
data protection environment. The problem is recognized by correlating multiple
incidents that pertain to the same performance-related issue. The integrated
incident and problem management tool performs root cause analysis, which reveals
that insufficient bandwidth of network links that carry replication traffic is the root
cause of the problem. The tool also logs the problem for administrative action.
Administrators of the problem management team can view the problem details
including the root cause recorded by the integrated incident and problem
management tool. They determine the remedial steps to correct the root cause. In
this case, the administrators decide to add a new network link to increase the
bandwidth for replication traffic. For that, they generate a request for change. Upon
obtaining approval from the change management, they ensure that the new link is
created by the implementation engineers. Thereafter, the problem management
team closes the problem.
Application Hardening
• Application hardening is a process followed during application development,
with the goal of preventing the exploitation of vulnerabilities that are typically
introduced during the development cycle.
• Application architects and developers must focus on various factors such as
proper application architecture, threat modeling, and secure coding while
designing and developing an application.
− Installing current application updates or patches provided by the application
developers can reduce some of the vulnerabilities identified after the
application is released.
• Application hardening process also includes process spawning control,
executable file protection, and system tampering protection.
• A common type of attack that can be imparted on applications is tampering with
executable files.