Dell Technologies PowerScale Administration
Dell Technologies PowerScale Administration
POWERSCALE
ADMINISTRATION - ON
DEMAND-SSP
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Dell Technologies PowerScale Administration - On Demand-SSP
PowerScale Administration 18
PowerScale 29
PowerScale 29
Scenario 29
PowerScale Nodes Overview 29
Flash Nodes 31
Hybrid Nodes 34
Archive Nodes 35
Gen6 Hardware Components 37
PowerScale F200 and F600 Hardware Components 39
PowerScale F900 Hardware Components 40
Accelerator Nodes 41
PowerScale Features 43
Node Interconnectivity 47
PowerScale Networking Architecture 49
Leaf-Spine Architecture 51
Leaf-Spine Network Components 53
Access Administration 90
Access Administration 91
Authentication Providers 92
Authentication Providers 92
Scenario 92
Authentication Providers Overview 92
Authentication Provider Structure 94
Active Directory Overview 95
Active Directory Configuration Videos 96
Network Time Protocol (NTP) Overview 97
NTP Configuration 98
LDAP Overview 99
LDAP WebUI and CLI Configuration Videos 101
Activity: Authentication-Active Directory 102
Challenge 102
Groupnets 115
Groupnets 115
Scenario 115
Network Configuration Planning 116
Groupnets and Access Zones Video 117
Multi-Tenancy Overview 118
Multi-tenancy Considerations 119
WebUI for Configuration 119
CLI for Configuration 121
Configure Groupnets with WebUI and CLI videos 122
Challenge 122
Authorization 178
Authorization 178
Permissions Scenario 178
Permissions Overview 178
Mixed data-access protocol environments 179
UNIX Permissions - POSIX Overview 181
POSIX in the WebUI 182
chmod Command 183
chown Command 183
Windows ACLs Overview 184
ACL Permission Policy Settings 185
Managing ACL Permissions 186
Synthetic vs Advanced ACLs 188
Permission Authority Video 189
Resources 191
Activity: Authorization 191
Challenge 191
Job Aid: Authorization 192
S3 Buckets 230
S3 Buckets 230
Scenario 230
S3 Overview 230
Implementation - Creating an S3 Bucket 232
Accessing the S3 Bucket 236
Considerations 236
Best Practices and Resources 237
Services 237
Challenge 237
HDFS 238
HDFS 238
Hadoop Introduction 238
SmartPools 321
SmartPools 321
CloudPools 331
CloudPools 331
Scenario 331
CloudPools Overview and Example Video 331
CloudPools Considerations 334
Cloud Providers and Storage 336
CloudPools Administration 338
File Pool Policies - CloudPools 341
CloudPools Settings 341
CLI for CloudPools 343
C2S Cloud Support 344
CloudPools Limitations 345
Activity: Cloudpools 346
Challenge 346
SmartQuotas 356
SmartQuotas 356
Scenario 356
SmartQuotas Overview Video 356
Quota Types 358
Enforcement Quotas 359
SmartQuotas Implementation 359
Default Quotas 360
Creating Default Directory Quotas 362
Quota Accounting 364
Overhead Calculations 365
Quotas and Thin Provisioning 366
Quota Nesting 367
Percent-Based Advisory and Soft Limits 369
Quota Notifications 370
Quota Notification Template 373
Template Variables 373
Considerations 375
Activity: SmartQuotas 376
Challenge 376
SmartDedupe 377
SmartDedupe 377
Scenario 377
SmartDedupe Overview 377
SmartDedupe Architecture 378
SmartDedupe Considerations 380
SmartDedupe Function 382
SmartDedupe Use Cases 383
SmartDedupe Jobs 385
SmartDedupe Administration 386
SnapshotIQ 390
SnapshotIQ 390
Scenario 390
SnapshotIQ Overview 390
Snapshot Operations 392
Copy on Write and Redirect on Write 393
Ordered and Unordered Deletions 394
Creating Snapshots 395
Accessing Snapshot Files 397
Preserving Permissions 398
Restoring Snapshots 399
Writable Snapshots 402
SnapshotIQ Considerations 403
Activity: SnapshotIQ 403
Challenge 404
SyncIQ 405
SyncIQ 405
Scenario 405
SyncIQ Overview Video 405
SyncIQ Deployment Topology 407
SyncIQ Considerations and Limits 410
SyncIQ Administrative Functions 414
SyncIQ Replication Policies 418
Creating the SyncIQ Policy 419
Copy vs Synchronize Policies 423
SyncIQ Configuration Video 424
Activity: SyncIQ 426
SmartSync 427
SmartSync 427
Smartsync Overview 427
SmartLock 429
SmartLock 429
Scenario 429
SmartLock Overview 430
SmartLock Concepts 431
SmartLock Operating Modes 431
SmartLock Directory Types 433
SmartLock Configuration 434
SmartLock CLI Example 436
Committing Files to WORM 436
SmartLock Considerations 437
Activity: SmartLock 438
Challenge 438
Appendix 467
PowerScale Administration
Scenario
NAS Overview
The two types of data used today are structured data and unstructured
data1. PowerScale specializes in storing unstructured data.
Scale-Up
Scale-Out
Scale-Out NAS
PowerScale architecture with different clients that are connected to the cluster through an
Ethernet switch. The backend layer allows for node-to-node communication.
With traditional NAS systems the file system4, volume manager5, and the
implementation of RAID6 are all separate entities.
7 As nodes are added, the file system grows dynamically, and content is
redistributed automatically.
8 OneFS performs the duties of the volume manager and applies
PowerScale
PowerScale
Scenario
Now that a scale-up or scale-out NAS and the OneFS operating system is
recognizable, move on to the differences of the PowerScale nodes. The IT
manager must ensure that everyone can distinguish between various
PowerScale nodes and determine what type of workflow works best for
each.
The PowerScale family has different offerings that are based on the need
for performance and capacity. OneFS powers all the nodes.
Gen6 4U chassis
• F200
• F600
• F900
• Minimum cluster is three nodes. Nodes can be added one at a time.
Isilon F800/810 flash nodes are EOL on April 30, 2024. End-of-Service
Support (EOSS) TBA.
Isilon Gen6 Hybrid and Archive nodes are EOL as of May 5, 2023. End-of-
Service-Support (EOSS) ends May 31, 2028.
Flash Nodes
The F-series nodes sit at the top of both performance and flash capacity
with All-Flash arrays for ultra-compute and high capacity. Even when the
cluster scales, the latency remains predictable.
F800/810 4U chassis
• F800
• F810
F900 2U chassis
• F900
111) Digital media: 4K, 8K, broadcast, real-time streaming, and post-
production
2) Electronic Design Automation: design, simulation, verification, and
analysis of electronic and mechanical systems design
3) Life Sciences: genomics DNA and RNA sequencing
F200 1U node
• F200
F600 1U node
• F600
Hybrid Nodes
The hybrid storage platforms are highly flexible and maintain a balance
between large capacity and high-performance storage providing support
for a broad range of enterprise workloads.
• H400
• H500
• H600
• H5600
• H700
• H7000
• Includes inline compression and deduplication
Archive Nodes
The A-series nodes are designed as highly efficient and resilient active
archive storage or long-term data retention for large-scale data archives.
The archive platforms can be combined with new or existing flash and
hybrid storage systems into a single cluster that provides an efficient
tiered storage solution.
• A200
• A2000
• A300
• A3000
• Includes inline compression and deduplication
Rear view and front view of an Isilon Gen6 and PowerScale Gen6 chassis.
1: The compute module bay of the two nodes make up one node pair.
Scaling out a cluster with Gen6 nodes is done by adding more node pairs.
You cannot mix node types in the same node pair.
2: Each Gen6 node provides two ports for front-end connectivity. The
connectivity options for clients and applications are 10GbE, 25GbE,
40GbE, and 100GbE.
3: Each node can have 1 or 2 SSDs that are used as L3 cache, global
namespace acceleration (GNA), or other SSD strategies.
4: Each Gen6 node provides two ports for back-end connectivity. A Gen6
node supports 10 GbE, 25 GbE, 40 GbE, and InfiniBand.
6: Each node has five drive sleds. Depending on the length of the chassis
and type of the drive, each node can handle up to 30 drives or as few as
15. A drive sled must always have the same type of disk drive.
7: You cannot mix 2.5" and 3.5" drive sleds in a node. Disks in a sled must
be the same type.
8: The sled can be either a short sled or a long sled. The types are:
9: The chassis comes in two different depths, the normal depth is about
37 inches and the deep chassis is about 40 inches.
1: Disks in a node are all the same type. Each F200 node has four SAS
SSDs.
2: The F200 and F600 nodes have two ports for backend connectivity
using PCIe slot 1.
3: The F200 front-end connectivity uses the rack network daughter card
(rNDC).
equally from both PSUs when the Hot Spare feature is disabled. Hot
Spare is configured using the iDRAC settings.
5: Scaling out an F200 or an F600 node pool only requires adding one
node. Both nodes support the use of iDRAC for remote access.
6: F600 nodes have eight NVMe SSDs. Slots 0 and 1 are not used for
drives.
V
Important: In the F600 nodes, the rNDC is not intended for
client network access, but it can provide 1 GbE
management port functionality.
1: Left control panel: Contains system health and system ID, and status
LED.
4: Right control panel: Contains the power button, VGA port, iDRAC Direct
micro-USB port, and USB 3.0 ports.
12: rNDC: The NIC ports that are integrated on the network daughter card
(NDC) provide front-end network connectivity.
Accelerator Nodes
OneFS 9.3 and later includes support for two new PowerScale accelerator
node models. The accelerator nodes do not contain any local storage
used as a part of the cluster and are optimized for CPU or memory
configurations. Based on the 1RU Dell PE R640 platform, accelerator
nodes include:
PowerScale Features
clustered usable
architecture Terabyte
that (TB)
combines • Enable
modular lower
storage. latency
• A cluster apps to
can scale leverage
up to 252 scale-out
nodes. NAS.
• Leveraging
standard
technology
to increase
focus on
scale.
Node Interconnectivity
18A simple and agile use case is a small start-up company growing at a
rapid pace. The company must start with limited capacity and then grow
on-demand for scale and new workloads.
1: Backend ports int-a and int-b. The int-b port is the upper port. Gen6
backend ports are identical for InfiniBand or Ethernet and cannot be
identified by looking at the node. If Gen6 nodes are integrated into a Gen5
or earlier cluster, the backend uses InfiniBand. There is a procedure to
convert an InfiniBand backend to Ethernet if the cluster no longer has pre-
Gen6 nodes.
3: Some nodes, such as archival nodes, might not use all 10 GbE port
bandwidth while other workflows might need the full utilization of the 40
GbE port bandwidth. The Ethernet performance is comparable to
InfiniBand so there should be no performance bottlenecks with mixed
performance nodes in a single cluster. Administrators should not see any
performance differences if moving from InfiniBand to Ethernet.
Network: There are two types of networks that are associated with a
cluster: internal and external.
network bandwidth available to the cluster scales linearly when new nodes
are added.
Leaf-Spine Architecture
A conceptual graphic of the leaf-spine topology for a PowerScale cluster using Z9100-ON
switches.
Spine Dell Z9264 64-port Back-end network with 64, 100 GbE
100 GbE switch ports that connects to the leaf
switches.
Resources
Challenge
1. What is the largest capacity SED hard drive available for the H700
and what would be the raw cluster size with forty nodes?
a. 10 TB hard drive, 6 PB cluster size.
b. 12 TB hard drive, 7.2 PB raw cluster size.
c. 16 TB hard drive, 9.6 PB raw cluster size
d. 20 TB hard drive, 12 PB raw cluster size
2. Select the statements that are true about Gen6 nodes and
PowerScale F200/600/900 nodes.
a. A minimum of four nodes are required to form a Gen6 cluster
whereas only three F200/600/900 nodes are required to form a
cluster.
b. Gen6 nodes and F200/600/900 nodes are based on a
server/blade type architecture that is adapted from PowerEdge
platforms.
c. You can add a minimum of one F200/600/900 node to an existing
cluster whereas Gen6 nodes can only be added as node-pairs.
d. A new Gen6 or PowerScale F200/600/900 cluster installation can
use either InfiniBand or Ethernet for the internal network. Use
caution as the ports look identical and damage occurs if incorrect
cables are inserted.
Scenario
• Serial Console23
• Web Administration Interface (WebUI)24
• Command Line Interface (CLI)25
• Platform Application Programming Interface (PAPI)26
• Front Panel Display27
• iDRAC/IPMI28
and manage the cluster. Access to the CLI is through a secure shell (SSH)
connection to any node in the cluster
26 The customer uses the OneFS application programming interface (API)
Movie:
External player for the video. The transcript of the video is in the course
guide.
Link:
https://edutube.dell.com/Player.aspx?vno=KjBgi9m8LmZLw58klDHmOA=
=&autoplay=true
Four options are available for managing the cluster. The web
administration interface (WebUI), the command-line interface (CLI), the
serial console, or the platform application programming interface (PAPI),
also called the OneFS API. The first management interface that you may
use is a serial console to node 1. A serial connection using a terminal
emulator, such as PuTTY, is used to initially configure the cluster. The
serial console gives you serial access when you cannot or do not want to
use the network. Other reasons for accessing using a serial connection
The WebUI is a graphical interface that is used to manage the cluster. The
WebUI requires at least one IP address that is configured to be on one
external Ethernet port of each node. To access the WebUI from another
system, an Internet browser is used to connect to port 8080.29
29 The user must log in using the root account, admin account, or an
account with log-on privileges. After opening the web administration
interface, there is a four-hour login timeout. The WebUI supports multiple
browsers including Microsoft Edge, Mozilla Firefox, Google Chrome, and
Apple Safari.
The CLI can be accessed in two ways: Out-of-band30 and In-band31. Both
methods are done using any SSH client such as OpenSSH or PuTTY.
Access to the interface changes based on the assigned privileges.
OneFS commands are code that is built on top of the UNIX environment
and are specific to OneFS management. You can use commands together
in compound command structures combining UNIX commands with
customer facing and internal commands.
The graphic shows the OneFS CLI access using one of the cluster nodes.
30 Accessed using a serial cable that is connected to the serial port on the
back of each node. As many laptops no longer have a serial port, a USB-
serial port adapter may be needed.
31 Accessed using an external IP address that is configured for the cluster.
5: The CLI command use includes the capability to customize the base
command with the use of options, also known as switches and flags. A
single command with multiple options results in many different
permutations, and each combination results in different actions performed.
6: The CLI is a scriptable interface. The UNIX shell enables scripting and
execution of many UNIX and OneFS commands.
CLI Usage
The example shows the help option used for the isi status command.
isi config
Use the console to change initial configuration settings. When in the isi
config console, other configuration commands are unavailable. The
exit command is used to go back to the default CLI.
Deep Dive: See the PowerScale OneFS Info Hubs and then
find the OneFS API Reference Guide under the proper
OneFS software version.
Access the Dell PowerScale API Developer Portal for full
API documentation.
• The Gen6 front panel display is an LCD screen with five buttons that
are used for basic administration tasks33.
• The interface consists of the LCD screen, a round ENTER button for
making selections, and four arrow buttons for navigating menus. There
are four LEDs across the bottom of the interface that indicate which
node is being communicated with. To change the communication to
another node, use the arrow buttons.
Column 2
Challenge
Lab Assignment: Launch the lab image and connect to the cluster using
the WebUI and the CLI.
34 You can join a node to a cluster and the panel displays the node name
after the node has joined the cluster. The LCD display provides system
information, status, and error messages to indicate that the system is
functioning correctly or requires attention.
Scenario
Licensing
A license file contains a record of the active cluster software licenses and
the cluster hardware. One copy of the license file is stored in the Dell
Software Licensing Central repository, and another copy is stored on the
cluster. Both license files must match.
License overview
Activation Wizard
After the activation file is generated, submit the file to Dell SLC to receive
back a signed license file for the cluster.
the activation file to Dell SLC, and then uploading an updated license file
to your cluster.
Trial Licenses
There are two different numbers that identify a node. The numbers are the
device ID and logical node number or LNN.
When adding new nodes to a cluster, the cluster gains more CPU,
memory, and disk space. The methods for adding a node are:
• Front panel
• Configuration Wizard using serial connection
• WebUI
• CLI
Join the nodes in the order that the nodes should be numbered.
Compatibility
Node series compatibility depends upon the amount of RAM, the SSD
size, number of hard drive, and the OneFS version.
Remove a Node
• Once the node has been reformatted, it can be added back39 into the
cluster or into another cluster.
• Use these commands40 to stop Smartfail.
Movie:
The web version of this content contains a movie.
The video shows how to remove a node from the cluster using the WebUI.
Reformat a Node
38 If the node is still connected to the back-end network when the process
completes, it automatically initiates a reformat - otherwise, it can be
reformatted manually.
39 The use cases are to remove older nodes during a life-cycle technology
refresh. First, to move a node from one cluster to another cluster. Second,
if PowerScale Support has identified a critical hardware failure.
40 1) 'isi devices node stopfail' command is used to discontinue the
Movie:
The web version of this content contains a movie.
The video shows the reformat option.
Reimage a Node
Download the reimaging bundle for USB flash drive from the PowerScale
OneFS Drivers and Downloads support site.
Export cluster
configuration
OneFS supports the backup and restore of http, quota, snapshot, nfs,
smb, s3, ndmp configurations.
44Copy the OneFS install file to the node and run the "isi_reimage"
command to reimage a node without using a flash drive.
Cluster Shutdown
Administrators can restart or shut down the cluster using the WebUI45 or
the CLI46.
Challenge
Lab Assignment: Launch the lab and add a node using the Configuration
Wizard and add a node using the WebUI.
45 The WebUI Hardware page has a tab for Nodes to shut down a specific
node, or the Cluster tab to shut down the cluster.
46 Native UNIX commands do not elegantly interact with OneFS, because
Scenario
At the core of OneFS, is the single file system across the cluster (/ifs).
The single file system in practice is a common directory structure.
• /ifs/data47
− Avoid modifying /ifs/data/Isilon_Support, which is created
the first time the isi_gather_info command is run to collect
cluster logs.
• /ifs/home48
• /ifs/.ifsvar49
Using a single file system starting with a newly created directory under
/ifs is recommended.
49SyncIQ report files are stored in this directory. Do not use for general
storage.
Defining the directory structure enables the creation of pools and group
data with common characteristics. A defined directory structure allows
servicing each group of data based on the class of service requirements.
Having a defined structure at the onset of an implementation avoids
repetitive work and streamlines configurations.
53For instance: compliance mode, backup systems, large file support, and
quotas.
Use case:
• A company named X-Attire plans to implement a single cluster for their
engineering team.
• The customer does not plan to have another cluster for remote disaster
recovery.
• The company name or authentication domain name is used as the
access zone name (x-attire).
Use case:
• X-Attire plans to implement a disaster recovery solution.
• X-Attire wants to replicate the Boston/homedirs directory to the
Seattle data center.
• From the Seattle DC, they plan to replicate the /groupdir directory to
Boston.
• Having the directory structure designed up front makes the
implementation easier.
On the /ifs directory, do not set inherited ACLs and do not propagate
ACL values.
ACL Type ACL Type ACL Type ACL Type ACL Type ACL Type
Challenge
Lab Assignment: Go to the lab and build the base directories. The base
directories are used throughout the implementation of the PowerScale
cluster.
Access Administration
Authentication Providers
Authentication Providers
Scenario
Before clients can access files that are stored on the cluster, they must be
authenticated.
54 Another local option is file provider, where a file is uploaded with user
information and can also contain UNIX user and group information from
other systems. The same file can be uploaded to other clusters, but the
file must be manually updated on each cluster.
55Most providers use UIDs (users ID), GIDs (group ID) and SIDs (security
ID).
5: The local provider provides authentication and lookup facilities for user
accounts added by an administrator.
Local authentication is useful when AD, LDAP, or NIS directory services
are not configured or when a specific user or application needs access to
the cluster.
Access control architectural components that show two configured access zones.
Active Directory can serve many functions, but the primary reason for
joining the cluster to an Active Directory domain is to perform user and
group authentication.
To join the cluster to AD, specify the fully qualified domain name, which
can be resolved to an IPv4 or an IPv6 address, and a username with join
permission. Areas to consider:
• Creates a single AD machine account
• Establishes trust relationship
• Supports NTLM and Microsoft Kerberos
• Each Active Directory provider must be associated with a groupnet
• Adding to an access zone
• Multiple AD instances
Movie:
Movie:
See the student guide for a transcript of the videos or view in eduTube
and download the transcripts.
• Synchronize issues58
• SMB time59
• Node time60
NTP Configuration
57 The cluster time property sets the date and time settings, either
manually or by synchronizing with an NTP server. After an NTP server is
established, setting the date or time manually is not allowed.
58 After a cluster is joined to an AD domain, adding an NTP server can
The list excludes nodes using their node numbers that are separated by a
space.
LDAP Overview
for clients connecting through the access zone. An access zone may
include at most one LDAP provider. The access zone and the LDAP
provider must reference the same groupnet.
Movie:
Movie:
See the student guide for a transcript of the videos or view in eduTube
and download the transcripts.
Link:
https://edutube.dell.com/Player.aspx?vno=JKBFLVJaUoqGz8DJmH4zqg=
=&autoplay=true
Select the Add LDAP Provider button. After the LDAP provider is
successfully added, the LDAP providers page displays a green status.
This means that the cluster can communicate with the LDAP server. Note
that AD and LDAP both use TCP port 389. Even though both services can
be installed on one Microsoft server, the cluster can only communicate
with one of services if they are both installed on the same server. This
concludes the demonstration.
Challenge
Lab Assignment:
Access Zones
Access Zones
Scenario
OneFS Access Zones provide secure, isolated storage pools for each
division within an organization, or even different tenants. Having their own
separate access zone allows consolidation of storage resources without
compromising security.
The cluster is configured for LDAP and Active Directory, and now it is time
to configure an access zone for two departments, finance, and
engineering. Finance is a Windows environment and engineering is a
Linux environment. Before configuring access zones, the IT manager
wants to ensure the administrator understands access zones and what
they do.
This video provides an overview for access zones. See the student guide
for a transcript of the video.
Movie:
https://edutube.dell.com/html5/videoPlayer.htm?vno=ju12CFZ0f+Z0Dv2lN
0VQRw
Although the default view of a cluster is that of one physical machine, you
can partition a cluster into multiple virtual containers called access zones.
Access zones enable you to isolate data and control who can access data
in each zone. Access zones support configuration settings for
authentication and identity management services on a cluster. Configure
authentication providers and provision protocol directories, such as SMB
shares and NFS exports, on a zone-by-zone basis. Creating an access
zone, automatically creates a local provider, which enables you to
configure each access zone with a list of local users and groups. You can
also authenticate through a different authentication provider in each
access zone.
The OneFS identity management maps users and groups from separate
directory services to provide a single combined identity. It also provides
uniform access control to files and directories, regardless of the incoming
protocol.
External Protocols
lsassd Daemon
External Providers
Internal Providers
Internal providers sit within the cluster operating system and are the Local,
or File Providers.
• File provider - authoritative third-party source of user and group
information.
• Local provider - provides authentication and lookup facilities for user
accounts added by an administrator.
• Local provider automatically created in access zone.
Once the client is at the front-end interface65, the associated access zone
then authenticates the client against the proper directory service. Once
authenticated to the cluster, mode bits and access control lists, or ACLs,
dictate the files, folders, and directories that clients can access.
A base or root directory defines the tree structure of the access zone.
The access zone cannot grant access to any files outside of the base
directory, creating a unique namespace.
4: The /ifs/eng base directory partitions data from the /ifs/dvt directory.
5: The base directory of the default System access zone is /ifs and cannot
be modified. Avoid using the OneFS built-in directories as base
directories.
Movie:
Movie:
You can avoid configuration problems on the cluster when creating access
zones by following best practices guidelines.
Challenge
Lab Assignment: Create the access zones for the environment and add
the authentication providers to the access zones using the WebUI and
CLI.
Just the Basics. The access zone job aid shows the basic requirements
and commands.
Create the access zone base directory and access zone for the Sales
organization. Sales uses LDAP to authenticate users. Although the
example shows adding an authentication provider, an authentication
provider is not required to create the access zone.
1. Create the base directory for the Sales access zone:
mkdir -p /ifs/Divgen/sales
Groupnets
Groupnets
Scenario
Groupnets reside at the top tier of the networking hierarchy and are the
configuration level for managing multiple tenants on your external network.
A subnet can also be called the SmartConnet zone and contain one or
more pools. Pools enable more granular network configuration.
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=b4A2l5FzF2na/Txqk2AUTA==&
attachments=true&autoplay=true
Because groupnets are the top networking configuration object, they have
a close relationship with access zones and the authentication providers.
Having multiple groupnets on the cluster means that you are configuring
access to separate and different networks, which are shown as org1 and
org2. Different groupnets enable portions of the cluster to have different
networking properties for name resolution. Configure another groupnet if
separate DNS settings are required. If necessary, but not required, you
can have a different groupnet for every access zone. The limitation of 50
access zones enables the creation of up to 50 groupnets.
When the cluster joins an Active Directory server, the cluster must know
which network to use for external communication to the external AD
domain. Because of this, if you have a groupnet, both the access zone
and authentication provider must exist within same groupnet. Access
zones and authentication providers must exist within only one groupnet.
Active Directory provider org2 must exist within the same groupnet as
access zone org2.
Multi-Tenancy Overview
In the scenario, the solution must treat each business unit as a separate
and unique tenant with access to the same cluster. The graphic shows
how each organization has its own groupnet and access zone.
Multi-tenancy Considerations
Groupnets are an option for those clusters that will host multiple
companies, departments, or clients that require their own DNS settings.
Some areas to consider are:
• DNS settings are per groupnet
• Create another groupnet only if separate DNS settings required.
• In a multiple tenant solution, a share can span access zones.
Combining namespaces and overlapping shares is an administrative
decision.
Like the WebUI, when using the CLI to create a groupnet with access
zones and providers in the same zone, create them in the proper order.
Movie:
Movie:
Challenge
SmartConnect Foundations
SmartConnect Foundations
Scenario
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=2xtwKNZ1xapC4+xrHKgV9w==
&attachments=true&autoplay=true
Architecture
For example, SmartConnect directs X-Attire users to F200 flash nodes for
their needed performance. GearItUp users access the H700 nodes for
general-purpose file sharing. The zones are transparent to the users and
can only be used with authorized access.
The SmartConnect Service IPs68 (SSIP or SIP) are addresses that are
part of the subnet.
Multiple Tiers
68 Do not put the SIPs in an address pool. The SIPs are a virtual IP within
the PowerScale configuration, it is not bound to any of the external
interfaces.
Use case69.
69 Clients use one DNS name to connect to the performance nodes and
another to connect to the general use nodes. The performance zone could
use CPU utilization as the basis for distributing client connections, while
the general use zone could use Round-robin.
70 The Marketing video group uses the F200 flash nodes. X-Attire uses a
subnet and/or pool that targets high-performance servers, giving the users
a higher level of performance. X-Attire can use a second subnet and/or
pool with a different zone name for general use, often desktops, without
the high-performance needs. Each group connects to a different name
and gets specific levels of performance. This way, whatever the desktop
users are doing, it does not affect the performance to the cluster.
SmartConnect Licensing
Failover Policies
Rebalance Policies
Rebalance Description
Policy
The SIPs, SmartConnect zone, and the DNS entries are the configuration
components for SmartConnect.
This demonstration shows the initial network configuration for the cluster.
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=4hL0i4iBe2BLqJzlT4dN/Q==&a
ttachments=true&autoplay=true
Enter the pool name and then select the access zone. For this
implementation the authentication providers and the access zones are
already created.
Next enter the range of IP address for this pool. Select the external node
interfaces that will carry the client traffic. The SmartConnect basic fully
qualified zone name is sales.dees.lab. We have the SmartConnect
SmartConnect Considerations
SMBv2/SMBv2.1 Stateful
S3 Stateless Dynamic
• Time-to-live value71.
Activity: SmartConnect
Challenge
Just the Basics: Configure the SSIP and create SmartConnect zone. The
example creates two SSIPs on subnet0.
1. Configure the SSIP:
isi network subnets modify subnet0 --scservice-addr
192.168.1.211-192.168.1.212
192.168.1.101-192.168.1.109 --sc-dns-zone
sales.delledu.lab
IP Address Pools
IP Address Pools
Scenario
After covering networking at the groupnet and subnet levels, it is now time
to examine IP address pools and then implement them on the cluster.
Determine the best allocation method and then configure IP address pools
and settings.
IP Address Pools
IP address pools are created within a subnet and consist of one or more
IP address ranges. IP address pools are allocated to external network
interfaces and associated with a node, a group of nodes, NIC ports or
aggregated ports.
Use case: Say that X-Attire adds four F800 nodes for a video media
group. X-Attire wants the video media team to connect directly to the F800
Link Aggregation
72The aggregated NICs must reside on the same node. You cannot
aggregate a NIC from node 1 and a NIC from node 2.
LACP
Configure LACP at the switch level and on the node. Enables the node to
negotiate interface aggregation with the switch.
Round Robin
Round robin balances outbound traffic across all active ports in the
aggregated link and accepts inbound traffic on any port.
Client requests are served one after the other based on their arrival.
Failover
Static aggregation mode that switches to the next active interface when
the primary interface becomes unavailable. The primary interface handles
traffic until there is an interruption in communication. At that point, one of
the secondary interfaces takes over the work of the primary.
Loadbalance (FEC)
IP Allocation Methods
Static
Dynamic
• Each interface in the pool gets exactly one IP (assuming enough IPs
as interfaces).
• Additional IPs will not be allocated to any interface.
• IPs do not move from one interface to another.
• If an interface goes down, then the IP also goes down.
Allocation Recommendations
• SMB Static
• HTTP
• HDFS
• sFTP
• FTPS
• SyncIQ
• SmartSync
• NFSv3 Dynamic
• NFSv4
• S3
Movie:
Link:
https://edutube.dell.com/html5/videoPlayer.htm?vno=xANsUhfq
pmyDEdwlJZIBDQ
In this six node Isilon cluster, an IP address pool provides a single static
node IP 10.126.90.140 through 145, to an interface on each cluster node.
Another pool of dynamic IPS have been created and distributed across the
cluster. When node one goes offline, the static node IP for node one is no
longer available.
The NFS failover IPS and the connected clients associated with node 1
failover to the remaining nodes based on the IP failover policy. If a node
with client connections established goes offline. The behavior is protocol
specific. The practice for NFS V3 and NFS V4 clients is to set the IP
allocation method to dynamic. NFS V3 automatically reestablishes an IP
connection as part of the NFS Failover.
reestablishes the connection with the IP on the new interface and retries
the last NFS operation.
Challenge
Scenario
Overview
By using roles, the root and admin users can assign others to integrated
or custom roles that have login and administrative privileges to perform
specific administrative tasks.
A user who is assigned to more than one role has the combined privileges
of those roles.
Roles
Integrated Roles
OneFS includes several integrated roles75 and are configured with the
most likely privileges necessary to perform common administrative
functions.
Custom roles
from /ifs.
80 The VMwareAdmin integrated role enables remote administration of
You can create custom roles81 and assign privileges that are mapped to
administrative areas in your PowerScale cluster environment.
With OneFS 8.2 and later, zRBAC enables the assigning of roles and a
subset of privileges that must be assigned on a per-access-zone basis.
Administrative tasks that the zone-aware privileges cover can be
delegated to an administrator of a specific access zone. Select each zone
integrated role to learn more.
• ZoneAdmin82
• ZoneSecurity Admin83
• BasicUserRole84
The following list describes what you can and cannot do through roles:
• Assign privileges to a role but not directly to users or groups.
− Data backup and restore privileges can be assigned to a user that
are explicitly for cluster data backup and restore actions.
• Create custom roles and assign privileges to those roles.
Movie:
Download the transcript from the player or see the student guide for a
transcript of the video.
Link:
https://edutube.dell.com/Player.aspx?vno=tQkWrNubtdORFBHxoRlMAg=
=&attachments=true&autoplay=true
Logout and then log in as Hayden, the AuditAdmin. The first indication is
the Access menu. Notice the options are missing. Navigating to Protocols,
Windows sharing, notice Hayden cannot create a share, only view. Also,
since added to a System zone role, Hayden can audit information in other
zones. System zone administrators are global.
Log out of the WebUI and login as Sai. You must login at an IP address or
netBios associated with the sales access zone. Viewing the Access
options, Sai does not have the privileges. Navigating to Protocols,
Windows sharing, notice Sai cannot switch to another access zone, but
can configure SMB shares. This demonstration stepped through
configuring RBAC and ZRBAC. This concludes the demonstration.
Role Management
Administrators can view, add, or remove members of any role, except for
integrated roles, whose privileges cannot be modified.
View Roles
Command Description
isi auth roles list A basic list of all roles on the cluster
isi auth roles list - Detailed information about each role on the
-verbose cluster, including member and privileged list
View Privileges
User Privileges are performed through the CLI. The table shows the
commands that display a list of your privileges or of another user.
Command Description
Create an empty custom role and then add users and privileges to the
role. Deleting a custom role does not affect the privileges or users that are
assigned to it.
The table shows the commands that are used to create, modify, and
delete a custom role.
Command Description
Privileges
• Using the CLI, run the command isi auth privileges for a
complete list of all privileges.
• For a list of privileges for a specific zone, run the command isi auth
privileges --zone=sales
------------------------------------------------------
--------------------------------------
Total: 51
Best Practices
Activity: RBAC-ZRBAC
Challenge
Lab Assignment: Go to the lab and create user accounts for RBAC and
ZRBAC.
The Basics: Create an administrative user and assign the user to a role.
1. Create a local user in the System zone:
isi auth users create DivAdmin -enabled yes --set-
password --password-expires no
Scenario
Layers of Access
Cluster connectivity has four layers of interaction. The third layer is identity
assignment. The layer is straightforward and based on the results of the
authentication layer.
There are some cases that need identity mediation within the cluster, or
where roles are assigned within the cluster that are based on user identity.
Identity Management
The video describes the access token generation. See the student guide
for the video transcript or download from the player.
Movie:
URL:
https://edutube.dell.com/Player.aspx?vno=MmSHIH1OvcP5nHsi0hd51g==
&attachments=true&autoplay=true
During authentication, OneFS creates an access token for the user. The
token contains the full identity of the user, including group memberships,
and OneFS uses the token later to check access to directories and files.
Steps Process
Primary Identities
OneFS supports three primary identity types, UIDs86, GIDs87, and SIDs88.
86 The user identifier, or UID, is a 32-bit string that uniquely identifies users
on the cluster. UNIX-based systems use UIDs for identity management.
87 The group identifier, or GID, for UNIX serves the same purpose for
domain identifier and ends with a 32-bit Relative Identifier (RID). Most
SIDs take the form S-1-5-21----, and are specific to a domain or system,
and denotes the object inside the domain. SID is the primary identifier for
users and groups in Active Directory.
89 A UID or GID is a 32-bit number with a maximum value of
4,294,967,295.
2: The security identifier, or SID, is a unique identifier that begins with the
domain identifier and ends with a 32-bit Relative Identifier (RID). Most
SIDs take the form S-1-5-21-<A>-<B>-<C>-<RID>, where <A>, <B>, and
<C> are specific to a domain or system, and <RID> denotes the object
inside the domain. SID is the primary identifier for users and groups in
Active Directory.
3: The group identifier, or GID, for UNIX serves the same purpose for
groups that UID does for users.
Secondary Identities
1: Windows provides a single namespace for all objects that is not case-
sensitive, but specifies a prefix that targets the DEES Active Directory
domain. UNIX assumes unique case-sensitive namespaces for users and
groups. For example, Sera and sera can represent different objects.
Multiple Identities
See the participant guide for information about mapping challenges and
considerations.
ID Mapper Database
1: The user mapper feature can apply rules to modify the user identity
OneFS uses, add supplemental user identities, and modify the group
membership of a user. The user mapping service combines user identities
from different directory services into a single access token. The mapping
service then modifies it according to any rules created.
On-Disk Identity
OneFS uses an on-disk identity store for a single identity for users and
groups.
On-disk identity types are Native, UNIX, and SID. Although you can
change the type of on-disk identity, the native identity is best for a network
with UNIX and Windows systems.
In native on-disk identity mode, setting the UID as the on-disk identity
improves NFS performance.
The use case for the default Native setting is an environment that has
NFS and SMB client and application access. With the Native on-disk
identity set, lsassd attempts to locate the correct identity to store on disk
by running through each ID-mapping method. The preferred object to
store is a real UNIX identifier. OneFS uses a real UNIX identifier when
found. If a user or group does not have a real UNIX identifier (UID or GID),
OneFS stores the real SID. Click on the highlighted icon to learn more.
Resources
For further documentation, see the PowerScale Product page for product
information, labs, demos, blogs, etc.
See the OneFS Info Hubs for user guides, hardware and software
compatibility, and other technical information.
Challenge
Lab assignment: Go to the lab and configure the on-disk identity type for
the cluster.
Authorization
Authorization
Permissions Scenario
The last topic to complete before creating Windows shares, NFS exports,
and S3 buckets is how OneFS handles permissions to the files and
directories. The IT manager must confirm each administrator can
configure POSIX mode bits, Windows ACLs, and how OneFS handles
both types of permissions.
Permissions Overview
OneFS supports two types of permissions data on files and directories that
control who has access: Windows-style access control lists (ACLs) and
the POSIX mode bits of UNIX systems. The individual files and folders
that clients access over NFS or SMB can have UNIX permissions and
Windows ACLs assigned.
• Supports NFS and SMB protocols and accesses the same directories
and files with different clients.
• OneFS generates a synthetic ACL, which is a direct representation of
the POSIX bits in ACL form.
The internal representation, which can contain information from either the
POSIX mode bits or the ACLs, is based on RFC 3530 (NFS version 4).
• Synthetic ACLs93
• Authority94
which approximates the mode bits of a UNIX file for an SMB client.
Because OneFS derives the synthetic ACL from mode bits, it can express
only as much permission information as mode bits can and not more.
94 OneFS must store an authoritative version of the original file
permissions for the file sharing protocol and map the authoritative
permissions for the other protocol. OneFS must do so while maintaining
the security settings for the file and meeting user expectations for access.
The result of the transformation preserves the intended security settings
on the files and ensures that users and applications continue to access
the files with the same behavior.
2: Group permissions
4: Configure permission flags to grant read (r), write (w), and execute (x)
permissions to users, groups, and others in the form of permission triplets.
The classes are not cumulative. OneFS uses the first class that matches.
Typically, grant permissions in decreasing order, giving the highest
permissions to the file owner and the lowest to users who are not the
owner or the owning group.
5: These permissions are saved in 16 bits, which are called mode bits.
6: The information in the upper 7 bits can also encode what the file can
do, although it has no bearing on file ownership. An example of such a
setting would be the “sticky bit.”
The graphic shows root user who is logged in and the /ifs/boston/hr
directory. Only root user can view and edit the owner and group of the
object.
chmod Command
Changes that are made using chmod can affect Windows ACLs.
chown Command
The chgrp command changes the group. View the man pages for
command definitions.
95A Windows ACL contains zero or more access control entries (ACEs),
each of which represents the security identifier (SID) of a user or a group
as a trustee. In OneFS, an ACL can contain ACEs with a UID, GID, or SID
as the trustee. Each ACE contains a set of rights that allow or deny
access to a file or folder.
are more complex than mode bits, ACLs can express much more granular
sets of access rules.
Windows includes many rights that you can assign individually, or you can
assign rights that are bundled together as permissions. For example, the
Read permission includes the rights to read and execute a file while the
Full Control permission assigns all user rights. Full Control includes the
right to change ownership and change the assigned permissions of a file
or folder.
When working with Windows, note the important rules that dictate the
behavior of Windows permissions. First, if a user has no permission that is
assigned in an ACL, then the user has no access to that file or folder.
Second, permissions can be explicitly assigned to a file or folder, and they
can be inherited from the parent folder. By default, when creating a file or
folder, it inherits the permissions of the parent folder. If moving a file or
folder, it retains the original permissions. On a Windows client, if the check
boxes in the Permissions dialog are not available, the permission are
inherited. You can explicitly assign permissions. Explicit permissions
override inherited permissions. The last rule to remember is that Deny
permissions take precedence over Allow permissions. However, an
explicit Allow permission overrides an inherited Deny permission.
The WebUI > Access > ACL policy settings page. Letter A-E in the WebUI
General ACL settings section translate in the CLI command output. Use
the "isi auth settings acls modify" command to configure the ACL settings.
2: The ls -len command shows numerical (n) owner and group SID or
UID/GID.
4: The long format includes file mode, number of links, owner, group, MAC
label, number of bytes, abbreviated month, day file last modified, hour file
last modified, minute file last modified, and the path name.
The list directory contents, ls, command provides file and directory
permissions information, when using an SSH session to the cluster.
PowerScale has added specific options to enable reporting on ACLs and
POSIX mode bits.
96 Advanced ACLs display a plus (+) sign when listed using an ls –l, or as
shown, the ls -led command. POSIX mode bits are present when a file has
a real ACL, however these bits are for protocol compatibility and are not
used for access checks.
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=EN8uMS3WuRwjY4Q0mIUaZw
==&attachments=true&autoplay=false
permission to the /dvt share then the user cannot write to the /linux
and /win directories or files within the directories.
A synthetic ACL does not exist on the file system and is not stored
anywhere. Instead, OneFS generates a synthetic ACL as needed, and
then discards it. OneFS creates the synthetic ACL in memory when a
client that only understands ACLs, such as Windows clients, queries the
permissions on a file that only has POSIX permissions.
With synthetic ACLs, POSIX mode bits are authoritative. POSIX mode bits
handle permissions in UNIX environments and govern the synthetic ACLs.
Permissions are applied to users, groups, and everyone, and allow or
deny file and directory access as needed. The read, write, and execute
bits form the permissions triplets for users, groups, and everyone. The
mode bits can be modified using the WebUI or the CLI standard UNIX
tools such as chmod and chown. Since POSIX governs the synthetic
ACLs, changes made using chmod change the synthetic ACLs. For
example, running chmod 775 on the /ifs/dvt directory changes the mode
bits to read-write-execute for group, changing the synthetic ACL for the
group. The same behavior happens when making the access more
restrictive, for example, running chmod 755, changes the synthetic ACL to
its corresponding permission. The chmod behavior is different when ACLs
are authoritative.
In the example, the directory /ifs/dvt/win has a real ACL. The POSIX
mode bits are 775. Running chmod 755 does not change to the POSIX
mode bits since merging 775 with 755 gives the combined value of 775.
Shown is an excerpt from the Isilon cluster WebUI page that shows the
different behaviors.
The first example shows that the share permission is everyone read-only
although the POSIX indicates read-write-execute. Windows users can
write to the share based on the synthetic ACLs. The second example
shows POSIX at 755. Although the ACL is set to a user with full control,
the user cannot write to the share—POSIX is authoritative.
The “+” indicates a real or native ACL that comes directly from Windows
and is applied to the file. Access control entries make up Windows ACLs.
An administrator can remove the real ACL permission using the chmod -
b command. ACLs are more complex than mode bits and can express a
richer set of access rules. However, not all POSIX mode bits can
represent Windows ACLs any more than Windows ACLs can represent
POSIX mode bits.
Once a file is given an ACL, its previous POSIX mode bits are no longer
enforced—the ACL is authoritative. The first example shows a real ACL
used, POSIX set for 777, and the share permissions for the user set to
read-only. Although the POSIX show read-write-execute for everyone, the
user cannot write because of the ACL. In contrast, the second example
shows the case where the user can write.
Resources
For further documentation, see the PowerScale Product page for product
information, labs, demos, blogs, etc.
See the OneFS Info Hubs for user guides, hardware and software
compatibility, and other technical information.
Activity: Authorization
Challenge
Lab Assignment: Log in to the cluster and verify the ACL policy setting.
Just the Basics: Commands are run using in an SSH session on a node.
• Change authorization to allow full control of a POSIX authoritative file:
chmod 777 /ifs/divgen/base/file1.foo
• Give the Windows domain users group full control with inheritance on a
directory:
chmod +ai group 'delledu\domain users' allow
generic_all /ifs/divgen/base
ls -led /ifs/divgen/base
OneFS Caching
OneFS Caching
Scenario
Caching maintains a copy of the metadata98 and/or the user data blocks in
a location other than primary storage.
Cache Levels
Both L1 cache and L2 cache are managed and maintained in RAM and
analogous to the cache used in processors (CPUs). These two cache
layers are present in all Dell PowerScale storage nodes. However, OneFS
is also capable of using SSDs as level 3, or L3 cache.
Each cache has its own specialized purpose and works together to
provide performance improvements across the entire cluster.
L1 Cache
L1 cache is the client-side cache. L1 is the buffer on the node that the
client connects and is involved in any immediate client data transaction.
Client-side cache.
1: L1 cache allows all blocks for immediate read requests. Read cache is
flushed after a successful read transaction and write cache is flushed after
a successful write transaction. L1 cache collects the requested data from
the L2 cache of the nodes that contain the data.
L2 Cache
L2 cache.
L3 Cache
Good for random, read heavy workflows accessing the same data sets.
L3 cache.
1: Extension of L2 cache.
2: SSD access is slower than access to RAM and is relatively slower than
L2 cache but faster than access to data on HDDs. L3 cache is an
extension of the L2 read cache functionality. Because SSDs are larger
than RAM, SSDs can store more cached metadata and user data blocks
than RAM. When L3 cache becomes full and new metadata or user data
blocks are loaded into L3 cache, the oldest existing blocks are flushed
from L3 cache. Flushing is based on first in first out, or FIFO. L3 cache
should be filled with blocks being rotated as node use requires.
SmartCache
When clients write to the cluster, OneFS uses write buffering to aggregate,
or coalesce, multiple write operations to the NVRAM file systems journals.
This data can then be written to disk safely and more efficiently, instead of
immediately writing to disk100. OneFS can then flush these cached writes
to disk at a later, more convenient time.
The graphic shows an eight-node cluster that is divided into two node
pools with a detailed view of one of the nodes.
2: The L2 cache on the node connects to the disk storage on the same
node. The L3 cache is connected to the L2 cache and serves as a read-
only buffer. The L2 cache on the node connects to the disk storage on the
same node.
5: Backend network.
Anatomy of a Read
When a client requests a file, the client-connected node uses the isi
get command to determine where the blocks that make up the file are
located.
1: The first file inode is loaded, and the file blocks are read from disk on all
other nodes. If the data is not present in the L2 cache, data blocks are
copied into the L2. The blocks are sent from other nodes through the
backend network.
2: If the data is already present in L2 cache, it is not loaded from the hard
disks. OneFS waits for the data blocks from the other nodes to arrive.
Otherwise, the node gets the data load from the local hard disks.
3: Data blocks are reconstructed in L1, and then the file is sent to the
client.
When a client requests a file write to the cluster, the client-connected node
receives and processes the file.
1: Writes are cached in the write coalescer, which is a portion of the RAM.
The write coalescer caches writes until becoming full, reaching a time
limit, or the client requests a confirmation that the blocks are committed to
stable storage.
2: The node then creates a write plan for the file, which includes
calculating Forward Error Correction, or FEC. Also, this node determines
where and how OneFS stores the file and corresponding metadata that
the file pool policy defines.
The write plan consists of a set of participant nodes that take part in
writing the chunk of data. Each participant node is responsible for
choosing the best place to locally put the data that it is sent.
4: Data blocks assigned to other nodes travel through the internal network
to their L2 cache, and then to their journal.
Once all the nodes have the data and parity blocks journaled, confirmation
from all the nodes is sent to the client-connected node.
5: Once complete, OneFS runs this write plan and guarantees its
successful completion. OneFS writes data at the highest protection level
that is possible given the cluster configuration and the protection policy.
Data is written to storage drives.
L3 Cache Settings
L3 cache is enabled by default for all new node pools that are added to a
cluster.
File system > Storage pools > SmartPools settings. Enabling and disabling L3 at the
global level and at the node pool level.
CLI Commands
The following commands are used to disable globally and to enable at the
node pool level.
• Global setting:
isi storagepool settings modify --ssd-13-cache-
default-enabled no
L3 Cache Considerations
substantially unchanged.
103 On Gen6.5 and PowerScale nodes F200, F600, F810, and F900, all
SMB Shares
SMB Shares
SMB Scenario
Share name
Share path
Protoocols
Enable the protocols intended for use for file sharing. Admins can
configure the OneFS cluster to use SMB or NFS exclusively, or both.
Administrators can also enable HTTP, FTP, and SSH, and configure
default shares and exports for each enabled protocol.
SMB Clients
105 SMB Multichannel is a feature of the SMB 3.0 protocol and supports
establishing a single SMB session over multiple network connections.
Multichannel provides: 1 - Increased throughput. 2 - Connection failure
tolerance. 3 - Automatic discovery.
Admins can configure access zones and connect to a zone through the
MMC Shared Folders snap-in to directly manage all shares in that zone.
For MMC connection requirements, see the OneFS Web Admin Guide.
File data no longer traverses the network for copy operations that the
server can perform.
The SMB server settings page contains the global settings that determine
how the SMB file sharing service operates.
These settings include enabling or disabling support for the SMB service.
A case107 for disabling the SMB service is when testing disaster readiness.
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=aMwue+nqUbFdOFoqKa98Fg=
=&attachments=true&autoplay=false
Share Creation
Settings Section
Type the full path of the share in the path field, beginning with /ifs.
You can also browse to the share. If the directory does not exist, the
Create SMB share directory if it does not exist creates the required
directory.
Directory ACLs
Use caution when applying the default ACL settings as it may overwrite
existing permissions in cases where the data has been migrated onto the
cluster.
When a cluster is set up, the default permissions on /ifs may or may not
be appropriate for the permissions on your directories.
Summary108
OneFS supports the automatic creation of SMB home directory paths for
users.
Variables:
• %D109
• %U110
• %Z111
• %L112
You can make access zone global changes to the default values in the
Default share settings tab. Changing the default share settings is not
recommended.
The share name can contain up to 80 characters, and can only contain
alphanumeric characters, hyphens, and spaces. The description field
contains basic information about the share. There is a 255-character limit.
Description is optional but is helpful when managing multiple shares.
Example for directory ACLs: Say that /ifs/eng is a new directory that was
created using the CLI. Windows users can create and delete files in the
directory. When creating the share, if the Do not change existing
permissions is set and then users attempt to save files to the share, an
access denied occurs because Everyone has read access. Even as an
administrator you cannot modify the security tab of the directory to add
Windows users because the mode bits limit access to only Root.As an
example, /ifs/eng is and NFS export and you explicitly want the /ifs/eng
Challenge
Lab Assignment: Log in to the cluster and create home directories and a
general-purpose share.
NFS Exports
NFS Exports
Scenario
Now that Windows users can access the cluster with SMB
shares, it is time to configure access for the Linux users. Create
an export that the Linux users can access. Begin with getting
familiar with NFS, creating the export, mounting the export, and
verifying client access.
NFS Overview
For the NFS protocol, OneFS supports NFSv3, NFSv4, plus NFSv4.1/2 in
OneFS 9.3. Also, OneFS 9.2 and later include support for NFSv3 over
RDMA.
With NFSv3 over RDMA support, direct memory access between OneFS
and NFSv3 clients is available with consuming less client CPU resource.
Also, NFSv3oRDMA improves OneFS network performance with lower
latency, lower CPU load, and higher throughput.
All in-flight reads and writes are handed off to another node in the cluster
to finish its operation without any user or application interruption.
If NFSv4 is enabled, specify the name for the NFSv4 domain in the NFSv4
domain field on the Zone setting page.
The NFS global settings determine how the NFS file sharing service
operates. The settings include enabling or disabling support for different
versions of NFS. Enabling NFSv4 is nondisruptive, and it runs
concurrently with NFSv3. Enabling NFSv4 does not impact any existing
NFSv3 clients.
Create and manage NFS exports using either the WebUI or the CLI.
OneFS can have multiple exports with different rules that apply the same
directory. A network hostname, an IP address, a subnet, or a netgroup
name can be used for reference. The same export settings and rules that
are created here apply to all the listed directory paths. If no clients are
listed in any entries, no client restrictions apply to attempted mounts.
When multiple exports are created for the same path, the more specific
rule takes precedence. For example, if the 192.168.3 subnet has read-
only access and 192.168.3.3 client has read/write access.
114The "root user mapping" default is to map root users to nobody, and
group is none. The default Security type is "UNIX (system)". Scrolling
down in the "Create an export" window shows the "Advanced settings".
because of the need for larger packet payloads than UDP supports. File
NFS Considerations
caching can be delegated to the client. The server can grant a read or
write file delegation to the clients, which enables the clients to
aggressively cache file data.
118 NFSv4.x servers present all the exported file systems within a single
Challenge
Lab Assignment: Create the NFS directory, export the directory, and
mount it to the Centos client.
S3 Buckets
S3 Buckets
Scenario
S3 Overview
OneFS supports the Amazon Web Services Simple Storage Service 120
(AWS S3) protocol for reading data from and writing data to the OneFS
platform.
120
AWS (S3) is an AWS service that provides object storage through a
web interface. OneFS 9.0.x and later support S3 as a tier 1 protocol.
OneFS S3 value:
• Multi-protocol access121
• Multi-tenancy - access zone aware
• Latency and IOPs equivalent to other OneFS protocols
• The data resides under a single namespace
Enable S3 Service
Zone Settings
CLI command to set the root path: isi s3 settings global modify
Create buckets using the Object storage (S3) page or using the isi s3
buckets create command.
Create Bucket
The graphic shows the Create a Bucket fields entry and the CLI
command to view an existing bucket.
S3 Bucket Table
Key Management
Considerations
Services
Challenge
HDFS
HDFS
Hadoop Introduction
File Striping
File Striping
Scenario
The IT manager wants to know about the process of striping and how the
operating system stripes a file. Describe in detail how files are broken up
for file stripes with diagrams for high-level file striping steps.
OneFS protects files as the data is being written. Striping protects the
cluster data and improves performance. To understand OneFS data
protection, the first step is grasping the concept of data and forward error
correction or FEC stripes.
• File Stripes - files are logically segmented into 128 KB stripe units to
calculate protection.
• FEC stripe unit - FEC stripe unit is the calculated piece of data
protection.
• Data stripe units(DSU) + FEC stripe units = Stripe width. In the
graphic, the stripe width is 12 (six DSU [1 MB file data] + 2 FEC).
• 16 data stripe units + 4 FEC = Maximum Stripe width of 20.
• 16 data stripe units = 2 MB. Files larger than 128 KB will have more
than one data stripe units.
The data stripe units and protection stripe units are calculated for each file
stripe by the Block Allocation Manager (BAM) process123.
3: The combined 128 KB stripe units are called the Stripe Width. A single
file stripe width can contain up to 16, 128 KB data stripe units for a
maximum size of 2 MB as the files data portion. A large file has thousands
of file stripes per file that is distributed across the node pool.
123The BAM process calculates 128 KB FEC stripe units to meet the
protection level for each file stripe. The higher the protection level, the
more FEC stripes units are calculated.
16 X 8 K = 128 KB
The steps show a simple example of the write process. The client saves a
file to the node that it is connected to. The file is divided into data stripe
units. The data stripe units are assembled into the maximum stripe widths
for the file. FEC stripe units are calculated to meet the Requested
Protection level. Then the DSU and FEC stripe units are striped across
nodes.
Step 1
OneFS stripes the data stripe units and FEC stripe units across the nodes
that make up the node pool. Some protection schemes124 use more than
one drive per node.
124 OneFS uses advanced data layout algorithms to determine data layout
for maximum efficiency and performance. Data is evenly distributed
across nodes in the node pool as it is written. The system can
continuously reallocate where the data is stored and make storage space
more usable and efficient. Depending on the file size and the stripe width,
as the cluster size increases, the system stores large files more efficiently.
Every disk within each node is assigned both a unique GUID (global
unique identifier) and logical drive number. The disks are subdivided into
32-MB cylinder groups that are composed of 8-KB blocks. Each cylinder
group is responsible for tracking, using a bitmap, whether its blocks are
used for data, inodes or other metadata constructs. The combination of
Graphic shows Gen6 cluster with a simple example of the write process.
Step 2
If the file is greater than 128 KB, then the file is divided into data stripe
units.
node number, logical drive number, and block offset make the block or
inode address, which the Block Allocation Manager controls.
Step 3
The node that the client connects to is the node that performs the FEC
calculation.
Step 4
The data stripe units are assembled to maximum stripe width for the file.
Also, here the protection level that is configured is N+1n125.
Step 5
Depending on the write pattern, the data and FEC stripes might be written
to one drive per node or two drives per node. The important take away is
that files segment into stripes of data, FEC is calculated, and this data
distributes across the cluster.
Challenge
Data Protection
Data Protection
Scenario
Data protection is one of the variables that are used to determine how
data is laid out. OneFS is designed to withstand multiple simultaneous
component failures while still affording access to the entire file system and
dataset.
• OneFS uses the Reed-Solomon algorithm
• The data can be protected up to an N+4n scheme.
• In OneFS, protection is calculated per individual file.
126 Smaller neighborhoods improve efficiency by the fact that the fewer
devices you have within a neighborhood, the less chance that multiple
devices will simultaneously fail.
127 The F200, F600, and F900 use a 16 GB NVDIMM for the journal. The
N+Mn
• N128
• M129
• Mn130
• N+Mn131
• N=M132
• N>M133
The number of sustainable drive failures are per disk pool. Multiple drive
failures on a single node are equivalent to a single node failure. The drive
loss protection level is applied per disk pool.
131 The available N+Mn Requested Protection levels are plus one, two,
three, or four “n” (+1n, +2n, +3n, and +4n). With N+Mn protection, only
one stripe unit is written to a single drive on the node.
132 If N equals M, the protection overhead is 50 percent. For example, with
N+Md: Bn Protection
The “d” is the number of drives and “n” is the number of nodes. So
N+3d:1n reads as N+3 drives or one node.
Unlike N+Mn, N+Md: Bn has different values for the number of drive loss
and node losses that are tolerated before data loss may occur. When a
node loss occurs, multiple stripe units are unavailable from each
protection stripe, and the tolerable drive loss limit is reached.
• M134
• d135
• Colon (:)136
• B137
• n138
134 In this protection level, M is the number of drives per node onto which a
stripe unit is written.
135 The number of drives.
137 The B value represents the number of tolerated node losses without
data loss.
138 “n” is the number of nodes.
With Gen 6x, for better reliability, better efficiency, and simplified
protection, using +2d:1n, +3d:1n1d, or +4d:2n is recommended.
N is replaced in the actual protection with the number of data stripe units
for each protection stripe. If there is no / in the output, it implies a single
drive per node. Mirrored file protection is represented as 2x to 8x in the
output.
The graphic shows the output showing Actual protection on a file from the isi get
command. The output displays the number of data stripe units plus the number of FEC
stripe units that are divided by the number of disks per node the stripe is written to.
The protection overhead for each protection level depends on the file size
and the number of nodes in the cluster. The percentage of protection
overhead declines as the cluster gets larger. In general, N+1n protection
has a protection overhead equal to the capacity of one node, N+2n to the
capacity of two nodes, N+3n to the capacity of three nodes, and so on.
Data mirroring requires significant storage overhead and may not always
be the best data-protection method. Example140
The table shows the relative protection overhead associated with each FEC requested
protection level. Indicators include when the FEC protection would result in mirroring.
MTTDL
MTTDL deals with how long you can go without losing data. MTTDL is
used to calculate the OneFS suggested protection.
• Accommodate failures141
• Disk pools142
• MTBF143
Quorum
the data lies on the cluster but for the safety of new data, no new
information will be written to the cluster. So, if a cluster loses its quorum,
the OneFS file system becomes read-only and will allow clients to access
data but not to write to the cluster.
146 Each protection level requires a minimum number of nodes. For
example, N+2d:1n needs a minimum of four Gen 6 nodes. Why? You can
lose one node and still have three nodes up and running, greater than
50%. You must keep quorum to keep the cluster writable.
There are six data stripe units to write a 768 KB file. The desired
protection includes the ability to sustain the loss of two hard drives.
1: Using N+2n protection, the 768-KB file will be placed into three
separate data stripes, each with two protection stripe units. Six protection
stripe units are required to deliver the requested protection level for the six
data stripe units. The protection overhead is 50 percent.
2: Using N+2d:1n protection the same 768-KB file requires one data
stripe, two drives wide per node and only two protection stripe units. The
eight stripe units are written to two different drives per node. The
protection overhead is the same as the eight node cluster at 25 percent.
3: If there is a eight node cluster, two FEC stripe units would be calculated
on the six data stripe units using an N+2n protection level. The protection
overhead in this case is 25 percent.
Mirroring is used to protect the file metadata and some system files that
exist under /ifs in hidden directories. Mirroring can be explicitly147 set as
the requested protection level in all available locations.
Use Case148
147 Mirroring is set as the actual protection on a file even though another
requested protection level is specified under certain conditions. If the files
are small, the FEC protection for the file results in a mirroring. The loss
protection requirements of the requested protection determine the number
of mirrored copies. Mirroring is also used if the node pool is not large
enough to support the requested protection level. For example, five nodes
in a node pool with N+3n Requested Protection, saves the file at 4X mirror
level, the actual protection.
148 One particular use case is where the system is used to only store small
Some protection schemes use a single drive per node per protection
stripe. The graphic shows only a single data stripe unit, or a single FEC
stripe unit is written to each node. These protection levels are N+Mn.
The table shows each requested N+Mn Requested Protection level over
the minimum number of required nodes for each level. The data stripe
units and protection stripe units149 can be placed on any node pool and in
any order.
FEC Node 8
149 The number of data stripe units depends on the size of the file and the
size of the node pool up to the maximum stripe width. N+1n has one FEC
stripe unit per protection stripe, N+2n has two, N+3n has three, and N+4n
has four. N+2n and N+3n are the two most widely used Requested
Protection levels for larger node pools, node pools with around 15 nodes
or more. The ability to sustain both drive or node loss drives the use when
possible.
FEC Node 9
The number of data stripe units depends on the size of the file and the
size of the node pool up to the maximum stripe width. As illustrated, N+1n
has one FEC stripe unit per protection stripe, N+2n has two, N+3n has
three, and N+4n has four. N+2n and N+3n are the two most widely used
Requested Protection levels for larger node pools, node pools with around
15 nodes or more. The ability to sustain both drive or node loss drives the
use when possible.
Advanced Protection
151Like other protection levels, the data stripe units and FEC stripe units
are placed on any node in the node pool and on any drive. N+3d:1n1d is
the minimum protection for node pools containing 6-TB drives. The use of
N+4d:2n is expected to increase especially for smaller to middle sized
node pools as larger drives are introduced.
Protection Overhead
The protection overhead for each protection level depends on the file size
and the number of nodes in the cluster. The percentage of protection
overhead declines as the node pool gets larger.
• N+1n152
• N+2n153
• N+3n154
• Data Mirroring155
152 N+1n protection has a protection overhead equal to the capacity of one
node.
153 N+2n protection has a protection overhead equal to the capacity two
nodes.
154 N+3n is equal to the capacity of three nodes, and so on. OneFS also
supports optional data mirroring from 2x-8x, enabling from two to eight
mirrors of the specified content.
155 Data mirroring requires significant storage overhead and may not
The table shows the relative protection overhead that is associated with each FEC
requested protection level available in OneFS. Indicators include when the FEC
protection would result in mirroring.
Considerations
As the cluster scales, the default protection may need adjusting. You may
not want to apply a higher protection to the entire cluster. Although you get
better protection, it is less efficient. Listed are areas to consider.
Challenge
Protection Management
Protection Management
Scenario
Requested Protection
Cluster-wide settings
The cluster-wide default data protection setting is made using the default
file pool161 policy.
To view or edit the default setting, go to File system > Storage pools >
File pool policies, and click View / Edit on the Default policy.
161The View default policy details window displays the current default file
pool policy settings. The current protection is displayed under requested
protection. The default setting is to use the requested protection setting at
the node pool level as highlighted in the Edit default policy details window.
The default file pool policy protection setting uses the node pool or tier
setting. When a node pool is created, the default requested protection162
that is applied to the node pool is +2d:1n.
The current requested protection for each node pool is displayed in the
Tiers and node pools section.
To view and edit the requested protection setting for the node pools in the WebUI, go to
the File system > Storage pools > SmartPools page. isi storagepool
nodepools modify v200_25gb_2gb --protection-policy +2n, sets the
requested protection of a node pool to +2n.
OneFS stores the properties for each file. To view the files and the next
level subdirectories, click the specific directory.
Manual settings163
To view directories and files on the cluster, go to File System > File system explorer.
The graphic shows a workflow that moves data to an archive tier of storage.
Suggested Protection
Not using the suggested protection does not mean that data loss occurs,
but it does indicate that the data is at risk. Avoid anything that puts data at
risk. What commonly occurs is a node pool starts small and then grows
beyond the configured requested protection level. The once adequate
+2d:1n requested protection level becomes no longer appropriate but is
never modified to meet the increased protection rrequirements. Not using
the suggested protection does not mean that data loss occurs, but it does
indicate that the data is at risk. Avoid anything that puts data at risk. What
commonly occurs is a node pool starts small and then grows beyond the
configured requested protection level.
loss occurs, but it does indicate that the data is at risk. Avoid anything that
puts data at risk. What commonly occurs is a node pool starts small and
then grows beyond the configured requested protection level. The once
adequate +2d:1n requested protection level becomes no longer
appropriate but is never modified to meet the increased protection
requirements.
The notification shows the suggested setting and node pools that are within suggested
protection levels are not displayed.
Actual Protection
The chart indicates the actual protection that is applied to a file according
to the number of nodes in the node pool. If actual protection does not
match the requested protection level, it may change to be more efficient
given the file or number of nodes in the node pool.
165 The actual protection level is the protection level OneFS sets. Actual
protection is not necessarily the same as the requested protection level.
166 A requested protection of +2d:1n and there is a 2-MB file and a node
minimum drive loss protection of two drives and node loss protection of
one node. The exception to meeting the minimum requested protection is
if the node pool is too small and unable to support the requested
protection minimums. For example, a node pool with four nodes and set to
+4n requested protection. The maximum supported protection is 4x
mirroring in this scenario.
isi get
The graphic shows the isi get –DD output. The output has three
primary locations containing file protection. The locations are a summary
in the header, line item detail settings in the body, and detailed per stripe
layout per drive at the bottom.
Challenge
Data Layout
Data Layout
Scenario
Now, examine how OneFS lays out the data on disks. The IT manager
wants to understand the data layout. Describe the different data access
pattern, illustrate an access pattern using concurrency and streaming.
1: The number of nodes in a node pool affects the data layout because
data spreads across all nodes in the pool. The number of nodes in a node
pool determines how wide the stripe can be.
3: The file size also affects data layout because the system employs
different layout options for larger files than for smaller files to maximize
efficiency and performance. Files smaller than 128 KB are treated as
small files. Due to the way that OneFS applies protection, small files are
triple mirrored.
4: The access pattern modifies both prefetching and data layout settings
that are associated with the node pool. Disk access pattern can be set at a
file or directory level so you are not restricted to using only one pattern for
the whole cluster.
There are four variables that combine to determine how OneFS lays out
data.
The variables make the possible outcomes almost unlimited when trying to
understand how the cluster behaves with varying workflow with differing
variables.
You can manually define some aspects of how it determines what is best,
but the process is automated.
You can tune how files are accessed to better suit the workflows. The data
access pattern defines the optimization settings for accessing data.
the given pool, when writing multiple protection stripes for a file. Each file
is written to the same subpool within the node pool. Streaming maximizes
the number of active drives per node as the streaming data is retrieved.
Streaming also influences the prefetch caching algorithm to be highly
aggressive and gather all associated data possible.
3: A random access pattern prefers using a single drive per node for all
protection stripes for a file, like a concurrency access pattern. With
random however, the prefetch caching request is minimal. Most random
data does not benefit from prefetching data into cache.
A 1 MB file is divided into eight data stripe units and three FEC units. The
data is laid out in three stripes. With a streaming access pattern, more
spindles are preferred. 1 MB file split into eight stripe unit and three stripes
- streaming uses spindles.
The graphic is a representation of a Gen6 chassis with four nodes. Each node has five
drive sleds. Each drive sled has three disks. The disk that is used is in the same
neighborhood (orange), do not traverse to disks in the other neighborhoods (gray).
A 1 MB file is divided into eight data stripe units and three FEC units. The
data is laid out in three stripes, one drive wide.
The graphic is a representation of a Gen6 chassis with four nodes. Each node has five
drive sleds. Each drive sled has three disks. The orange disk represents a neighborhood.
Configuring the data access pattern is done on the file pool policy, or
manually at the directory and file level. Set data access patterns using the
WebUI or use isi set for directory and file level or isi file pool
policy for file pool policy level.
For WebUI Administration, go to File systems > Storage pools > File pool policies.
Challenge
Storage Pools
Storage Pools
Scenario
Before configuring the file policies and tiering data, the IT manager wants
to know about the components of storage pools. Give a thorough
explanation that includes all the important details.
Storage pools monitor the health and status at the node pool level. Using
storage pools, multiple node pools can all co-exist within a single file
system, with a single point of management.
Gen6 drive sleds have three, four, or six drives170 whereas the F200 has 4
drive bays F600 has 8 drive bays and F900 has 24 drive bays.
170 Drives are segmented into disk pools, creating a failure domain.
Disk Pool
Disk pools provide separate failure domains. Each drive within the sled is
in a different disk pool, lessening the chance for data unavailability.
171Not spanning disk pools the granularity at which files are striped to the
cluster. Disk pool configuration is automatic and cannot be configured
manually. Removing a sled does not cause data unavailability as only one
disk per disk pool is temporarily lost.
Neighborhood
Gen6 Neighborhood
A Gen6 node pool splits into two neighborhoods when adding the 20th
node172. One node from each node pair moves into a separate
neighborhood.
172After the 20th node up to the 39th node, no two disks in a given drive
sled slot of a node pair share a neighborhood. The neighborhoods split
again when the node pool reaches 40 nodes.
Node Pool
SmartPools
SmartPools Basic173
SmartPools Advanced174
File Pools
File pools are the SmartPools logical layer, at which file pool policies are
applied.
173 The basic version of SmartPools supports virtual hot spares, enabling
space reservation in a node pool for re-protection of data. OneFS
implements SmartPools basic by default. You can create multiple node
pools, but only a single tier and only a single file pool. A single tier has
only one file pool policy that applies the same protection level and I/O
optimization settings to all files and folders in the cluster.
174 More advanced features are available in SmartPools with a license.
With the advanced features you can create multiple tiers and file pool
policies that direct specific files and directories to a specific node pool or a
specific tier.
User created, and defined policies are set on the file pools.
CloudPools
Moving the cold archival data to the cloud, lowers storage cost and
optimizes storage resources.
Serviceability
Listed are the CLI options that can help get information about storage
pools.
• isi status -p
Challenge
Lab Assignment: Go to the lab and verify the storage pool settings.
File Pools
File Pools
Scenario
The media team needs their storage on disks that do not compete with the
other disks. Provide information on segregating data into different node
pools.
File pool policies automate file movement, enabling users to identify and
move logical groups of files.
• User-defined filters175
• File-based, not hardware-based176
• User-defined or default protection and policy settings177
The example shows that each policy has a different optimization and
protection level. A file that meets the policy criteria for tier 3 is stored in the
tier 3 node pool with +3d:1n1d protection. Also, the file is optimized for
streaming access.
The default file pool policy is defined under the default policy.
175 Files and directories are selected using filters and apply actions to files
matching the filter settings. The policies are used to change the storage
pool location, requested protection settings, and I/O optimization settings.
176 Each file is managed independent of the hardware, and is controlled
policies. File pool policies add the capability to modify the settings at any
time, for any file or directory.
1: The individual settings in the default file pool policy apply to files not
defined in another file pool policy that you create. Default file pool policy
cannot be reordered or removed.
2: To modify the default file pool policy, click File system, click Storage
pools, and then click the File pool policies tab. On the File pool
policies page, next to the Default policy, click View/Edit.
3: You can choose to have the data that applies to the Default policy
target a specific node pool or tier or go anywhere. Without a license, you
cannot change the anywhere target. If existing file pool policies direct
data to a specific storage pool, do not configure other file pool policies with
anywhere.
5: You can specify a node pool or tier for snapshots. The snapshots can
follow the data, or go to a different storage location.
6: Assign the default requested protection of the storage pool to the policy,
or set a specified requested protection.
8: In the Data access pattern section, you can choose between Random,
Concurrency, or Streaming.
A tier that is called media tier with a node pool has been created.
The business unit targets their mp4 marketing segments to the media_tier
where the hosting application can access them.
Create the filters in the File matching criteria section when creating or
editing a file pool policy.
Filter elements:
• Filter type178
178File pool policies with path-based policy filters and storage pool
location actions are run during the write of a file matching the path criteria.
Path-based policies are first started when the SmartPools job runs, after
that they are started during the matching file write. File pool policies with
storage pool location actions, and filters that are based on other attributes
besides path, write to the node pool with the highest available capacity.
The initial write ensures that write performance is not sacrificed for initial
data placement.
• Operators179
• Multiple criteria180
SSD Options
With the exception of F-Series nodes, if a node pool has SSDs, by default
the L3 cache is enabled on the node pool. To use the SSDs for other
strategies, first disable L3 cache on the node pool. Manually enabling SSD
strategies on specific files and directories is not recommended.
Pros Cons
179 Operators can vary according to the selected filter. You can configure
the comparison value, which also varies according to the selected filter
and operator. The Ignore case box should be selected for files that are
saved to the cluster by a Windows client.
180 The policy requires at least one criterion, and allows multiple criteria.
You can add AND or OR statements to a list of criteria. Using AND adds a
criterion to the selected criteria block. Files must satisfy each criterion to
match the filter. You can configure up to three criteria blocks per file pool
policy.
Pros Cons
All metadata is on SSDs - Need enough SSD space to hold all the
speeds random lookups and metadata. Typically, two SSDs per
treewalks. node is good, especially with denser
options.
Pros Cons
Avoid SSDs
Using the 'Avoid SSD' option affects performance. This option writes all
file data and all metadata mirrors to HDDs. Typically, use this setting when
implementing L3 cache and global namespace acceleration (GNA) in the
same cluster. You create a path-based file pool policy that targets an L3
cache enabled node pool. The data SSD strategy and snapshot SSD
strategy for this L3 cache enabled node pool should be set to ‘Avoid SSD’.
181 The SetProtectPlus job applies the default file pool policy.
182 When SmartPools is licensed, the SmartPools job processes and
applies all file pool policies. By default, the job runs at 22:00 hours every
day at a low priority.
183 Uses a file system index database on the file system instead of the file
system itself to find files needing policy changes. By default, the job runs
at 22:00 hours every day at a low priority. The FilePolicy job was
introduced in OneFS 8.2.0.
Policy Template
Template settings are preset to the name of the template along with a brief
description. You can change the settings.
Template considerations:
• Opens a partially populated, new file pool policy.
• You must rename the policy.
• You can modify and add criteria and actions.
• Use in web administration interface only.
Plan to add more node capacity when the cluster reaches 80% so that it
does not reach 90%. The cluster needs the extra capacity for moving
around data, and for the VHS space to rewrite data when a drive fails.
Listed are more considerations.
• Avoid overlapping file policies where files may match more than one
rule. If data matches multiple rules, only the first rule is applied.
• File pools should target a tier and not a node pool within a tier.
• You can use the default policy templates as examples.
Serviceability
Listed here are the CLI options that can help get information about file
pools.
• If file pool policy rules are not being applied properly, check the policy
order.
• Test file pool policy before applying.
Challenge
SmartPools
SmartPools
Scenario
Before configuring the file tiering, the IT manager wants to know about the
OneFS SmartPools settings. Describe the SmartPools settings and then
configure SmartPools.
SmartPools Overview
SmartPools enables the grouping of nodes into storage units that include
node pools, CloudPools, and tiers.
With SmartPools, you can segregate data based on its business value,
putting data on the appropriate tier of storage with appropriate levels of
performance and protection.
SmartPools Licensing
185
Node pool membership changes through the addition or removal of
nodes to the cluster. Typically, tiers are formed when adding different
node pools on the cluster.
SmartPool Settings
Cache Statistics
GNA
• Minimum 1.5% of all disk capacity must be SSD and 20% of nodes
must contain SSDs.
• Use SSDs to store metadata mirror in different node pools.
GNA Aspects
Pros Cons
Allows metadata read acceleration for Difficult to manage and size the
non-SSD nodes - need some nodes disk
with SSDs.
Hard rules and limits
Helps Job Engine and random reads Links expansion of one tier to
another tier to adhere to the
limits.
L3 Cache
L3 cache is enabled by default for all new node pools that are added to a
cluster.
Any node pool with L3 cache enabled is excluded from GNA space
calculations and do not participate in GNA enablement.
VHS
Virtual hot spare, or VHS, allocation enables space to rebuild data when a
drive fails.
When selecting the option to reduce the amount of available space, free-
space calculations exclude the VHS reserved space.
OneFS uses the reserved VHS free space for write operations unless you
select the option to deny new data writes.
Spillover
With the licensed SmartPools module, you can direct data to spillover to a
specific node pool or tier group.
Actions
If you clear the box (disable), SmartPools does not modify or manage
settings on the files.
Protection example: If a +2d:1n protection is set and the disk pool suffers
three drive failures, the data that is not lost can still be accessed. Enabling
the option ensures that intact data is still accessible. If the option is
disabled, the intact file data is not accessible.
GNA can be enabled if 20% or more of the nodes in the cluster contain
SSDs and 1.5% or more of the total cluster storage is SSD-based. The
recommendation is that at least 2.0% of the total cluster storage is SSD-
based before enabling GNA. Going below the 1.5% SSD total cluster
space capacity requirement automatically disables GNA metadata. If you
SmartFail a node that has SSDs, the SSD total size percentage or node
percentage containing SSDs could drop below the minimum requirement,
disabling GNA. Any node pool with L3 cache enabled is excluded from
GNA space calculations and do not participate in GNA enablement.
GNA also uses SSDs in one part of the cluster to store metadata for
nodes that have no SSDs. The result is that critical SSD resources are
maximized to improve performance across a wide range of workflows.
VHS example: If specifying two virtual drives or 3%, each node pool
reserves virtual drive space that is equivalent to two drives or 3% of their
total capacity for VHS, whichever is larger. You can reserve space in node
pools across the cluster for this purpose, equivalent to a maximum of four
full drives. If using a combination of virtual drives and total disk space, the
larger number of the two settings determines the space allocation, not the
sum of the numbers.
SmartPools Considerations
• Node pools must have at least four nodes for Gen6 and at least three
nodes for the F200/600/900. The default is one node pool per node
type and configuration.
• The file pool policy default is all files are written anywhere on cluster.
To target more node pools and tiers, activate the SmartPools license.
Activity: SmartPools
Challenge
CloudPools
CloudPools
Scenario
Next, take the file pool policies to the CloudPools level. For some of the
long-term archive data, the group is looking at cloud options. Explain
CloudPools and how file pool policies are used with CloudPools.
Customers who want to run their own internal clouds can use a
PowerScale installation as the core of their cloud.
The video provides a CloudPools overview and use case. See the student
guide for a transcript or download from the player.
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=ysxbj+pXwvWoRyg1bMcWuA=
=&attachments=true&autoplay=true
The next policy moves the archive data off the cluster and into the cloud
when data is not accessed for more than 180 days. Stub files that are also
called SmartLinks are created. Stub files consume approximately 8 KB
space on the Isilon cluster. Files that are accessed or retrieved from the
cloud, or files that are not fully moved to the cloud, have parts that are
cached on the cluster and are part of the stub file. The storing of
CloudPools data and user access to data that is stored in the cloud is
transparent to users.
CloudPools Considerations
187 In OneFS 8.2 and later, CloudPools compress data before sending it
over the wire.
188 CloudPools in OneFS 8.2 prevents enabling compliance mode on
CloudPools Administration
Configure and manage CloudPools from the WebUI File system, Storage
pools page, CloudPools tab. Managing CloudPools using the CLI is done
with the isi cloud command.
CloudPools Tab
Once the SmartPools and CloudPools licenses are applied, the WebUI
shows the cloud storage account options.
The file pool policies enable the definition of a policy to move data out to
the cloud.
The graphic shows the window for creating a cloud storage account.
• You can create and edit one or more cloud storage accounts in
OneFS.
• Before creating a cloud storage account, establish an account with one
of the supported cloud providers.
• OneFS attempts to connect to the cloud provider using the credentials
you provide in the cloud storage account.
CloudPools SmartLink
Run the isi get -D command to see files archived to the cloud using
CloudPools.
SmartPools file pool policies are used to move data from the cluster to the
selected CloudPools storage target.
When configuring a file pool policy, you can apply CloudPools actions to
the selected files.
CloudPools Settings
You may want to modify the settings for the file pool policy based on your
requirements. Modifications are not necessary for most workflows. You
can elect to encrypt and compress data.
The graphic shows various default advanced CloudPool options that are
configured.
1: You can encrypt data prior to archiving it to the cloud. Cloud data is
decrypted when accessed or recalled.
2: You can compress data prior to archiving to the cloud. Cloud data is
decompressed when accessed or recalled.
3: Set how long to retain cloud objects after a recalled file replaces the
SmartLink file. After the retention period, the cloud objects garbage
collector job cleans up the local resources allocated for the SmartLink
files, and removes the associated cloud objects.
4: If a SmartLink file has been backed up and the original SmartLink file is
subsequently deleted, associated cloud objects are deleted only after the
retention time of the backed-up SmartLink file has expired.
5: If a SmartLink file has been backed up and the original SmartLink file is
subsequently deleted, associated cloud objects are deleted only after the
original retention time, or a longer incremental or full backup retention
period, has expired.
6: Specifies how often SmartLink files modified on the cluster are written
to their associated cloud data objects.
9: Specifies how long the system retains recalled cloud data that is in the
cache of associated SmartLink files.
The output of the isi cloud command shows the actions that can be
taken.
1: Use to grant access to CloudPool accounts and file pool policies. You
can add and remove cloud resource, list cluster identifiers, and view
cluster details.
3: Use to archive or recall files from the cloud. Specify files individually, or
use a file matching pattern. Files that are targeted for archive must match
the specified file pool policy, or any file pool policy with a cloud target.
8: Files that are stored in the cloud can be fully recalled using the isi
cloud recall command. Recall can only be done using the CLI. When
recalled, the full file is restored to its original directory. The file may be
subject to the same file pool policy that originally archived it, and rearchive
it to the cloud on the next SmartPools job run. If re-archiving is
unintended, the recalled file should be moved to a different, unaffected,
directory. The recalled file overwrites the stub file. You can start the
command for an individual file or recursively for all files in a directory path.
10: Use to restore the cloud object index (COI) for a cloud storage
account on the cluster. The isi cloud access add command also
restores the COI for a cloud storage account.
• Support196
• Integration197
• No Internet connection198
196 C2S support delivers full CloudPools functionality for a target endpoint,
and supports the use with C2S Access Portal (CAP), and X.509 client
certificate authority. C2S also provides support (from AIMA) to securely
store certificates, validate, and refresh if needed.
197 The CloudPools C2S feature offers an integrated solution with AWS
CloudPools Limitations
In a standard node pool, file pool policies can move data from high-
performance tiers to storage tiers and back as defined by their access
policies. However, data that moves to the cloud remains stored in the
cloud unless an administrator explicitly requests data recall to local
storage. If a file pool policy change is made that rearranges data on a
normal node pool, data is not pulled from the cloud. Public cloud storage
often places the largest fees on data removal, thus file pool policies avoid
removal fees by placing this decision in the hands of the administrator.
The connection between a cluster and a cloud pool has limited statistical
features. The cluster does not track the data storage that is used in the
cloud, therefore file spillover is not supported. Spillover to the cloud would
present the potential for file recall fees. As spillover is designed as a
temporary safety net, once the target pool capacity issues are resolved,
data would be recalled back to the target node pool and incur an
unexpected fee.
Statistic details, such as the number of stub files on a cluster or how much
cache data is stored in stub files and would be written to the cloud on a
flush of that cache, is not easily available. No historical data is tracked on
the network usage between the cluster and cloud either in writing traffic or
in read requests. These network usage details should be viewed from the
cloud service management system.
Activity: Cloudpools
Challenge
File Filtering
File Filtering
Scenario
It appears to be that there are some types of files that need not be stored
in the production directories. The IT manager wants you to explain file
filtering and configure a policy to filter unnecessary files.
Considerations:
• No limit to extension list.
• Per access zone.201
• Configurable for the SMB defaults202.
• No license is required.
199 Explicit deny lists are used to block only the extensions in the list.
OneFS permits all other file types to be written. Administrators can create
custom extension lists based on specific needs and requirements.
200 Explicit allow list permits access to files only with the listed file
enable file filtering in an access zone, OneFS applies file filtering rules
only to files in that access zone.
202 OneFS does not take into consideration which file sharing protocol was
used to connect to the access zone when applying file filtering rules.
However, you can apply additional file filtering at the SMB share level.
203 Users cannot access files with a denied extension. OneFS can deny a
file through the denied extensions list, or because the extension is not
included as part of the allowed extensions list.
204 Administrators have full control to read or delete all files. Administrators
Select each item below to learn more about its use case.
205 How the file filtering rule is applied to the file determines where the file
filtering occurs. If a user or administrator accesses the cluster through an
access zone or SMB share without applying file filtering, files are fully
available.
206 File filters are applied only when accessed over the supported
protocols.
Another use case is to limit the cost of storage. Organizations may not
want typically large files, such as video files, to be stored on the cluster, so
they can deny .mov or .mp4 file extension
Another use case is to limit an access zone for a specific application with
its unique set of file extensions. File filtering with an explicit allow list of
extensions limits the access zone or SMB share for its singular intended
purpose.
When enabling file filtering in an access zone, OneFS applies file filtering
rules only to files in that access zone.
You can configure file filters on the Protocols > Windows sharing (SMB)
> Default share settings page207.
Access zone level: Web UI: Access > File filter > File filter settings.
Modify file filtering settings by changing the filtering method or editing file
extensions.
CLI: isi smb shares create and isi smb shares modify
commands. The administrative user must have the
ISI_PRIV_FILE_FILTER privilege.
Challenge
Your Challenge: Login to the cluster and configure file filtering on an SMB
share.
SmartQuotas
SmartQuotas
Scenario
This video provides an overview for SmartQuotas. See the student guide
for a transcript of the video.
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=tCIE1bGAUz6k3W1ic8tZfw==&
autoplay=true
Quotas are a useful way to ensure that a user or department uses only
their share of the available space. SmartQuotas are also useful for
enforcing an internal chargeback system. SmartQuotas contain flexible
Before OneFS 8.2, SmartQuotas reports the quota free space only on
directory quotas with a hard limit. For user and group quotas,
SmartQuotas reports the size of the entire cluster capacity or parent
directory quota, not the size of the quota. OneFS 8.2.0 includes
enhancements to report the quota size for users and groups. The
enhancements reflect the true available capacity that is seen by the user.
Quota Types
The File System uses the concept of quota types as the fundamental
organizational unit of storage quotas. Storage quotas comprise a set of
resources and an accounting of each resource type for that set. Storage
quotas are also called storage domains.
Enforcement Quotas
SmartQuotas Implementation
2: User and default user quotas: User quotas are applied to individual
users and track all data that is written to a specific directory. User quotas
enable the administrator to control the capacity any individual user
3: Group and default group quotas: Group quotas are applied to groups
and limit the amount of data that the collective users within a group can
write to a directory. Group quotas function in the same way as user
quotas, except for a group of people and instead of individual users.
Default group quotas are applied to all groups, unless a group has an
explicitly defined quota for that directory. Default group quotas operate like
default user quotas, except on a group basis.
Default Quotas
Default Directory
Default user quotas are applied to all users unless a user has an explicitly
defined quota for that directory. Default user quotas enable the
administrator to apply a quota to all users, instead of individual user
quotas. Default group quotas operate like default user quotas, except on a
group basis.
For example, HR team's requirement policy states that every user should
be assigned a quota of 1 GB for their home directories. Creating a user
quota for each user in the organization is cumbersome and inefficient.
Instead, the lead administrator can create a default user quota 1 GB on
the folder hosting all the home directories.
The WebUI can be used to view the created quotas and their links. See
the student guide for information about quota links.
208 The 'isi quota' command is used to create the default directory quota.
The Unlink option makes the quota independent of the parent, meaning
modifications to the default directory quota no longer apply to the sub
directory. This example shows removing the link on the Screen_shots sub
directory and then modifying the default directory quota on the parent,
Quota, directory. Remove the link using the button on the WebUI or isi
quota quotas modify --
path=/ifs/training/Features/Quota/Screen_shots --
type=directory --linked=false. Using the --linked=true
option re-links or links to the default directory quota.
Quota Accounting
The quota accounting options are Include snapshots in the storage quota,
209and enforce the limits for this quota based on:
209 Tracks both the user data and any associated snapshots. A single path
can have two quotas that are applied to it, one without snapshot usage
(default) and one with snapshot usage. If snapshots are in the quota, more
files are in the calculation.
210 Enforces the File system logical size quota limits. The default setting is
Application logical size is typically equal or less than file system logical
Overhead Calculations
213 With thin provisioning, the cluster can be full even while some users or
directories are well under their quota limit. Configuring quotas that exceed
the cluster capacity enables a smaller initial purchase of capacity/nodes.
214 Thin provisioning lets you add more nodes as needed, promoting a
Quota Nesting
The isi quota quotas list command is used to compare the size of
a quota to the amount of data it holds.
At the top of the hierarchy, the /ifs/sales folder has a directory quota of 1
TB. Any user can write data into this directory, up to a combined total of 1
In OneFS 8.2.0 and later, you can view advisory and soft quota limits as a
percent of the hard quota limit.
A hard limit must exist to set the advisory and soft percentage.
Quota Notifications
Overview
Design Options
• When the limit is exceeded, based on the rule, you can create a cluster
event and email the owner or specific addresses.
• The email integrates with the provider email configuration or can be
manually mapped217.
• While SmartQuotas comes with default email message templates, you
can also create custom message templates as .txt files using
predefined variables.
Example
safer side, the administrator wants to monitor usage so that the quota is
not exceeded by more than 60% and wants to be notified of usage daily.
The graphic shows one of the available quota templates that are located in
the /etc/ifs directory.
• PAPI support218.
• OneFS 8.2 enhancements219.
Template Variables
Considerations
Best Practice:
– Do not enforce quotas on file system root (/ifs).
– Do not configure quotas on SyncIQ target directories.
Activity: SmartQuotas
Challenge
SmartDedupe
SmartDedupe
Scenario
The cluster is hosting home directories for the users. Much of the data is
shared and has multiple copies. Deduplication should help address the
inefficient use of space. Describe the benefits of deduplication, explain
how deduplication works, and schedule deduplication on a directory.
SmartDedupe Overview
SmartDedupe Architecture
When SmartDedupe runs for the first time, it scans the dataset and
selectively samples blocks from it, creating the fingerprint index. This
index contains a sorted list of the digital fingerprints, or hashes, and their
associated blocks. Then, if they are determined to be identical, the block’s
pointer is updated to the already existing data block and the new,
duplicate data block is released.
3: Shadow stores are similar to regular files but are hidden from the file
system namespace, so cannot be accessed using a path name. A shadow
store typically grows to a maximum size of 2 GB (or about 256K blocks),
with each block able to be referenced by 32,000 files. If the reference
count limit is reached, a new block is allocated, which may or may not be
in the same shadow store. Also shadow stores do not reference other
shadow stores. And snapshots of shadow stores are not permitted
because the data that is stored in shadow stores cannot be overwritten.
SmartDedupe Considerations
• SmartDedupe License.220
• Best for static files and directories.221
• Post process222 - not immediate - eventual.
The data starts out at the full literal size on the drives, and might get
deduplicated hours or days later.
• Snapshots.229
• One deduplication job runs at a time.230
− Schedule deduplication to run during the cluster’s low usage hours,
by default, the SmartDedupe job runs automatically.
− After the initial dedupe job, schedule incremental dedupe jobs to
run about every two weeks, depending on the size and rate of
change of the dataset.
− Run SmartDedupe with the default "low" impact Job Engine policy.
• Rehydrates files from shadow store.231
SmartDedupe Function
A job in the OneFS Job Engine232 runs through blocks that are saved in
every disk pool, and compares the block hash values.233
replicated or backed up. To avoid running out of space, ensure that target
clusters and tape devices have free space to store deduplicated data.
229 SmartDedupe will not deduplicate the data stored in a snapshot.
2: Compare 8 KB blocks.
5: Free blocks
232 The job first builds an index of blocks, against which comparisons are
done in later phase, and ultimately confirmations and copies take place.
The deduplication job can be a time consuming, but because it happens
as a job the system load throttles, the impact is absolute. Administrators
find that their cluster space usage has dropped once the job completes.
233 If a match is found, and confirmed as a true copy, the block is moved to
the shadow store, and the file block references are updated in the
metadata.
234 A home directory scenario where many users save copies of the same
file can offer excellent opportunities for deduplication.
235 Static, archival data is seldom changing, therefore the storage that is
images can benefit from deduplication. Deduplication does not work well
with compressed data, the compression process tends to rearrange data
to the point that identical files in separate archives are not identified as
such. Environments with many unique files do not duplicate each other, so
the chances of blocks being found which are identical are low.
SmartDedupe Jobs
Because the sharing phase is the slowest deduplication phase, a dry run,
or DedupeAssessment, returns an estimate of capacity savings.
SmartDedupe Administration
Considerations
• Compressed and deduplicated data does not exit the file system as
compressed or deduplicated in any form.
• There is no OneFS WebUI support for inline data reduction.
Configuration and management are through the CLI only.
• Since data reduction extends the capacity of a cluster, it also has the
effect of reducing the per-TB compute resource ratio (CPU, memory,
I/O). Hence, it is less suited for heavily accessed data, or high-
performance workloads.
• Depending on an application’s I/O profile and the effect of data
reduction on the data layout, read and write performance and overall
space savings can vary considerably.
• SmartDedupe does not attempt to deduplicate files smaller than 32 KB
whereas inline deduplication has no limit on file size.
Best Practices
Activity: SmartDedupe
Challenge
SnapshotIQ
SnapshotIQ
Scenario
There is a need to use snapshots to give users the ability to recover files.
Describe snapshot behavior, identify snapshot types, and configure and
manage snapshot functionality.
SnapshotIQ Overview
If you modify a file and determine that the changes are unwanted, you can
copy or restore the file from the earlier file version.
Snapshot Operations
Snapshot growth: The data is modified, and only the changed data
blocks are contained241 in snapshots.
OneFS uses both Copy on Write, or CoW and Redirect on Write, or, RoW.
SnapshotIQ uses both copy on write (CoW) and redirect on write (RoW)
strategies for its differential snapshots and uses the most appropriate
method for a given situation. Both have pros and cons, and OneFS
240 A snapshot is not a copy of the original data, but only an extra set of
pointers to the original data. At the time it is created, a snapshot
consumes a negligible amount of storage space on the cluster. The
original file references the snapshots.
241 If data is modified on the cluster (Block D’ in the graphic), only one
copy of the changed data is made. With CoW the original block (Block D)
is copied to the snapshot. The snapshot maintains a pointer to the data
that existed at the time that the snapshot was created.
242 Snapshots do not consume a set amount of storage space, there is no
Creating Snapshots
3: Optional: In the Snapshot Name field, type a name for the snapshot.
4: In the Path field, specify the directory that you want the snapshot to
contain.
Snapshot location
243 Use shorter expiration periods for snapshots that are generated more
frequently, and longer expiration periods for snapshots that are generated
less frequently.
244 For example, if snapping a directory located at
can only open the .snapshot directories for which they already have
permissions. Without access rights users cannot open or view any
.snapshot file for any directory.
Accessing snapshots
Preserving Permissions
The snapshot preserves248 the file and directory permissions at that point
in time of the snapshot.
246 This is a virtual directory where all the snaps listed for the entire cluster
are stored.
247 To view the snapshots on /ifs/eng/media, user can change directory
Restoring Snapshots
Restore Theory
QUESTION: What happens when the user wants to recover block A data
that was overwritten in Time 3 with A’?
Clients with Windows Shadow Copy Client can restore the data from the
snapshot.
Clients accessing the export over NFS can navigate using the .snapshot
directory.
To recover a deleted file, right-click the folder that previously contained the
file, click Restore Previous Version, and select the required file to recover.
To restore a corrupted or overwritten file, right-click the file itself, instead
of the folder that contains the file, and then click Restore Previous
Version.
Writable Snapshots
Writable Snapshots
The source snapshot and its writable copy must reside in a directory in the
/ifs file system. The writable snapshots feature creates a directory quota
on the root of the writable snapshot that monitors its space usage.
Considerations
• Hard links to files within the domain of the writable snapshot cannot be
created from outside the writable snapshot domain.
• Files that reside in the writable snapshot domain cannot be renamed
from outside that writable snapshot domain.
SnapshotIQ Considerations
Activity: SnapshotIQ
Challenge
SyncIQ
SyncIQ
Scenario
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=OZC9t92nwmWVLWNjfT/+5w=
=&attachments=true&autoplay=true
Shown is a cluster with the source directory using SyncIQ to replicate data
to a remote target directory. OneFS SyncIQ uses asynchronous
replication, enabling you to maintain a consistent backup copy of your
data on another Isilon cluster. Asynchronous replication is similar to an
asynchronous file write.
The target system passively acknowledges receipt of the data and returns
an ACK once the target receives the entire file or update. Then the data is
passively written to the target. SyncIQ enables you to replicate data from
one PowerScale cluster to another. Activate a SyncIQ license on both the
primary and the secondary Isilon clusters before replicating data between
them. You can replicate data at the directory level while optionally
excluding specific files and sub-directories from being replicated.
Under each deployment, the configuration could be for the entire cluster or
a specified source directory. Also, the deployment could have a single
policy that is configured between the clusters or several policies, each with
different options aligning to RPO and RTO requirements.
Select the tabs to know more about each type of deployment Typologies.
One-to-one
One-to-many
Many-to-one
Local Target
Cascaded
Considerations
Capabilities
249 The SyncIQ Job Engine is separate from the cluster maintenance
activity Job Engine in OneFS. SyncIQ runs based on SyncIQ policies that
you can schedule or run as required manually.
250 Semi-automated failovers from source to target, and semi-automated
failback from target to original source. Failover and failback only include
the cluster preparation activities and do not include DNS changes, client
redirection or any required networking changes.
251 The semi-automated failover process preserves the synchronization
Limitations
reservation evenly among all policies. Using the CLI, you can make
bandwidth reservations for individual policies.
257 The target cluster contains a copy of the source data synchronized on
Compatibility
The table shows the versions of OneFS you can synchronize using
SyncIQ. Target cluster running OneFS 7.1.x version of OneFS is no longer
supported. For information about the support and service life-cycle dates
for hardware and software products, see the Isilon Product Availability
Guide.
CloudPools
SyncIQ provides data protection for CloudPools data and provides failover
and failback capabilities.
Failover
Failback
Like failover, failback must be performed on a per policy basis. The user
must make the same network changes to restore access to direct clients
to the source cluster.
The example shows a failback where the client accesses source data.
261 A failback can happen when the primary cluster is available once again
for client activities. The reason could be from any number of
circumstances including the natural disasters are no longer impacting
operations, or site communication or power outages have been restored to
normal. You must failback each SyncIQ policy.
Failback Preparation
The mirror policy is placed under Data Protection > SyncIQ > Local
Targets on the primary cluster. On the secondary cluster, the mirror policy
is placed under Data Protection > SyncIQ > Policies.
Failover Revert
262Failover revert stops the failover job and restores the cluster to a sync
ready state. Failover reverts enables replication to the target cluster to
once again continue without performing a failback.
263Use revert if the primary cluster once again becomes available before
any writes happen to the target. A temporary communications outage or if
doing a failover test scenario are typical use cases for a revert.
264 You create and start replication policies on the primary cluster. A policy
specifies what data is replicated, where the data is replicated to, and how
often the data is replicated.
265 The primary cluster holds the source root directory, and the secondary
The panels describe the files for creating the SyncIQ policy. Refer to the
student guide for more information.
Settings
Creating a SyncIQ policy is done of the Data protection > SyncIQ >
Policies page or using the isi sync policy create command.
266SyncIQ jobs are the operations that do the work of moving the data
from one PowerScale cluster to another. SyncIQ generates these jobs
according to replication policies.
Target Cluster
Advanced
The final segment of the policy creation are the advanced fields.
Source cluster directories: In the Source Cluster criteria, the Source root
directory is the SyncIQ domain. The path has the data that you want to
protect by replicating it to the target directory on the secondary cluster.
Unless otherwise filtered, everything in the directory structure from the
source root directory and below replicates to the target directory on the
secondary cluster.
Includes and excludes: The Included directories field permits adding one
or more directory paths below the root to include in the replication. Once
an include path is listed that means that only paths listed in the include
path replicate to the target. Without include paths all directories below the
root are included. The Excluded directories field lists directories below the
root you want explicitly excluded from the replication process. You cannot
fail back replication policies that specify includes or exclude settings. The
DomainMark job does not work for policies with subdrectories mentioned
in Include or Exclude. Using includes or excludes for directory paths does
not affect performance.
File matching criteria: The File matching criteria enables the creation of
one or more rules to filter which files do and do not get replicated.
Creating multiple rules connect them together with Boolean AND or OR
statements. When adding a new filter rule, click either the Add an “And”
condition or Add an “Or” condition links. File matching criteria says that if
the file matches these rules then replicate it. If the criteria does not match
the rules, do not replicate the file.
Target: Snapshots are used on the target directory to retain one or more
consistent recover points for the replication data. You can specify if and
how these snapshots generate. To retain the snapshots SyncIQ takes,
select Enable capture of snapshots on the target cluster. SyncIQ always
retains one snapshot of the most recently replicated delta set on the
secondary cluster to facilitate failover, regardless of this setting. Enabling
capture snapshots retains snapshots beyond the time period that is
needed for SyncIQ. The snapshots provide more recover points on the
secondary cluster.
Deep copy: The Deep copy for CloudPools setting applies to those
policies that have files in a CloudPools target. Deny is the default. Deny
enables only stub file replication. The source and target clusters must be
at least OneFS 8.0 to support Deny/ Allow the SyncIQ policy to determine
if a deep copy should be performed. Force automatically enforces a deep
copy for all CloudPools data that are contained within the SyncIQ domain.
Allow or Force are required for target clusters that are not CloudPools
aware.
267 If a mirrored copy of the source is the goal, create a sync policy.
268 If the goal is to have all source data that is copied and to retain deleted
Movie:
Link:
https://edutube.dell.com/Player.aspx?vno=6cyyA4XvBqkyHJwXs6ltdg==&
attachments=true&autoplay=true
Let's begin at our host. Here I have the browser open and noticed I am
connected to both the Boston Cluster and the Seattle Cluster. The
directory that I will replicate are the user home directories.
For this demonstration, note that I also have the share from the Seattle
cluster map to the host. The Seattle cluster is configured with identical
access zones and naming structure. Both clusters have sync IQ licensed.
I'll go to the sync IQ page on the source cluster, mouseover data
protection, select SYNC IQ, and then the policies tab.
Next select create sync policy. All the fields are explained in detail in the
eye Salon disaster recovery training course and the eye salon. One FS
Web administration guide. For the demonstration, I am keeping the
configuration very simple. The policy name is Boston to Seattle holders.
The type of replication will be synchronized. The job type is manual. Slash
IF slash home slash DES is to syncIQ domain.
Down in the Target Field, the target host is Seattle DES Lab and the
Target directory is slash F slash home slash DES. Click on create policy.
That's it for a very basic policy before moving on, let's take a look at an
assessment to validate the process. Select assess sync. Now go to the
reports tab and view the assessment. We can also navigate to the Seattle
Cluster Data Protection SYNC IQ page on the local targets tab to view the
policy target.
Now let's start the job. This is done on the source cluster. Here we are on
the sync IQ page policies tab. In the actions column for a job, select start
job. Next, we'll go to the summary tab and verify the job status is running
for demonstration purposes. There are only a few files to replicate and the
initial replication should complete fairly quickly. My host has both the
source and target directories map.
Here we can see the animation directory has replicated to the target
cluster. This directory is read only and if I go into the animation directory
and try to write, I'll get an error. Back in the source directory, I am able to
write data. Here I'm just doing a copy and paste of one of the files.
Next, let's failover to the Seattle cluster. First, I'll stop writes to the cluster
by disabling SMB services. This may not be practical in a real environment
where you may want read and write access to other workflows. Our sync
IQ policy is already set to manual. If the policy is not set to manual, we
would modify the sync IQ policy to manual at this time. A failover is done
on the target cluster will go to the Seattle Cluster Sync IQ page and local
targets tab. On the policy to fail over, I'll choose allow rights. In a real
Now let's failback to put the cluster back in a normal state. On the Boston
Cluster Sync IQ page policies tab, I'll select Resync prep. Next on the
target cluster SYNC IQ page policies tab of verify the mirror policy is
created. Let's go ahead and stop writes to the secondary cluster. Undo
this by disabling the SMB service. Back on the Sync IQ page, I will start
job on the mirror policy. The summary tab will show the active jobs. Now
go to the source cluster sync IQ page, local targets tab and verify the
status is finished. And then select allow rights. This puts the target cluster
back to a readonly state. Next, go to the target cluster Sync IQ page
policies tab and on the mirror policy, select Resync Prep. On both the
source and target clusters, I'll re-enable the SMB service.
Activity: SyncIQ
Challenge
SmartSync
SmartSync
Smartsync Overview
• Data protection
• Data repurposing (copy)
• Data archive
269
SmartLock
SmartLock
Scenario
SmartLock Overview
SmartLock Concepts
• Retention Period
• Compliance
• WORM
Compliance Enterprise
Only use if SEC 17a-4 must be Does not restrict cluster to follow
followed. SEC 17a-4 rules.
273 Compliance directories can be created only if the cluster has been
upgraded to SmartLock compliance mode.
274 SmartLock enterprise mode is the default SmartLock operation mode.
When using SmartLock, there are two types of directories: enterprise and
compliance. A third type of directory is a standard or non-WORM275
directory.
If using the compliance clock, you must copy data into the Compliance
SmartLock directory structure before committing the data to a WORM
state.
SmartLock Configuration
1: Setting to "On" enables the root user to delete files that are currently
committed to a WORM state.
5: After a specified period, a file that has not been modified is committed
to a WORM state.
6: Files committed to a WORM state are not released from a WORM state
until after the specified date, regardless of the retention period.
Use case:
• The administrator requires a WORM directory where files are in a
WORM state for at least 30 days and are removed from the WORM
state after 60 days.
• The default retention is 60 days.
• Set minimum and maximum retention dates.
CLI:
For a file to have a file retention date applied, and set to a read-only state,
you must commit the file to WORM.
Until the files are committed to WORM, files that are in a SmartLock
directory act as standard files that you can move, modify, or delete.
First set the retention date on the file, then Set per SmartLock domain.
Commit files to WORM state using Windows After the time period
controls or UNIX commands. expires, the file is
Example: # chmod ugo-w automatically committed to
/ifs/finance/worm/JulyPayroll. WORM.
xls
SmartLock Considerations
• Do not use rm -rf . The command option r deletes all files and
directories recursively, and option f avoids prompting before deleting.
• In OneFS versions later than OneFS 8.0.1, SyncIQ failback is
supported on SmartLock directories.
Activity: SmartLock
Challenge
Monitoring Administration
PowerScale HealthCheck
PowerScale HealthCheck
Scenario
HealthCheck Overview
The OneFS HealthCheck tool is a service that helps evaluate the cluster
health status and provides alerts to potential issues.
You can use HealthCheck to verify the cluster configuration and operation,
proactively manage risk, reduce support cycles and resolution times, and
improve uptime.
The graphic shows the checklist items for the cluster_capacity check. The
HealthCheck terms and their definition are:
The CLI can be used to view the parameters of a checklist item. The
example shows CLI window viewing the node capacity item parameters.
Running a HealthCheck
The example shows selecting the Run option for the cluster_capacity
checklist. The HealthCheck table shows the status of the checklist.
HealthCheck Schedule
Viewing an Evaluation
HealthCheck Resources
Challenge
InsightIQ
InsightIQ
Scenario
InsightIQ Overview
Qualifying Questions
InsightIQ Dashboard
You can modify the view to represent any time period where InsightIQ has
collected data. Also, breakouts and filters can be applied to the data. In
the Aggregated Cluster Overview section, you can view the status of all
monitored clusters as a whole. There is a list of all the clusters and nodes
that are monitored. Total capacity, data usage, and remaining capacity are
shown. Overall health of the clusters is displayed. There are graphical and
numeral indicators for connected clients, active clients, network
throughput, file system throughput, and average CPU usage. Depending
on the chart type, preset filters enable you to view specific data. For
example, In/Out displays data by inbound traffic compare with outbound
traffic.
You can also view data by file access protocol, individual node, disk,
network interface, and individual file or directory name. If displaying the
data by the client only, the most active clients are represented in the
displayed data. Displaying data by event can include an individual file
system event, such as read, write, or lookup. Filtering by operation class
displays data by the type of operation being performed.
Capacity Analysis
Default Reports
• Performance reports
• File system reports
• Live reporting
You can drill down to file system reporting to get a capacity reporting
interface that displays more detail about usage, overhead and anticipated
capacity.
The administrator can select cluster information and use that as a typical
usage profile to estimate when the cluster reaches 90% full. The
information is useful for planning node/cluster expansion ahead of time to
avoid delays around procurement and order fulfillment.
The Plot data shows the granularity of the reporting available. The
Forecast data shows the breakout of information that is shown in the
forecast chart. Depending on the frequency and amount of variation,
outliers can have a major impact on the accuracy of the forecast usage
data.
InsightIQ collects the FSA data from the cluster for display to the
administrator.
Enable FSA
Before you can view and analyze data usage and properties through
InsightIQ, you must enable the FSA feature.
To enable FSA, Open the Monitored Clusters page by clicking Settings >
Monitored Clusters. In the Actions column for the cluster that you want to
enable or disable FSA, click Configure. The Configuration page displays.
Click the Enable FSA tab. To enable the FSA job, select Generate FSA
reports on the monitored cluster. To enable InsightIQ for FSA report,
select View FSA reports in InsightIQ.
280 The job collects information across the cluster, such as the number of
files per location or path, the file sizes, and the directory activity tracking.
If there are long time periods between the FSAnalyze job runs, the
snapshot can grow very large, possibly consuming much of the cluster's
space. To avoid large snapshot, you can disable the use of snapshots for
FSAnalyze. Disabling snapshot use means that the jobs may take longer
to run.
Considerations
Challenge
isi statistics
isi statistics
Scenario
The three main commands that enable you to view the cluster from the
command line are isi status, isi devices, and isi statistics.
isi statistics
The output shows the operations by protocol. The example shows that
NFS clients are connected to node 6 with 278.5k bytes per second input
rate.
isi devices
281 Other services such as InsightIQ, the WebUI, and SNMP gather
information using the "isi statistics" command.
isi status
The isi status command displays information about the current status
of the cluster, alerts, and jobs. The example of the isi status output gives a
general node status, performance metrics, critical alerts, and Job Engine
status.
The --quiet option omits the alerts and Job Engine status output.
Tip: See the CLI Reference guide for a complete list of the
command options and output definitions.
The isi statistics command dumps all collected stats, and you can
run the "query" subcommand on a specific statistic.
• You can build a custom isi statistics query that is not in the
provided subcommands
• Cluster and node statistics from kernel counters
• isi_stats_d
282 A cron job can run on UNIX-based systems to schedule periodic jobs.
The example output shows the isi statistics drive command for
the SSD drives on node 6.
The head -10 option displays the first 10 most active most accessed files
and directories.
Practical Skills
Combining large sets of collected data with log analysis can help identify
long-term trends and sources of trouble.
2: isi Statistics can fill the gaps. Skillful use of isi statistics
can produce equivalent information to what InsightIQ offers and contains
many performance-related options.
Challenge
Click the Save Progress & Exit button in the course menu or below
to record this content as complete.
Go to the next learning or assessment, if applicable.
The default value to create an export task is "all" and is optional. All
components that are currently supported are http, quota, snapshot, nfs,
smb, s3, and ndmp.
DNS Primer
2: A single period (.) represents the root domain, and is the top level of the
DNS architecture.
3: Below the root domain are the top-level domains. Top-level domains
represent companies, educational facilities, nonprofits, and country codes
such as *.com, *.edu, *.org, *.us, *.uk, *.ca, and so on. A name registration
authority manages the top-level domains.
5: The last record in the tree is the hosts record, which indicates an
individual computer or server.
What is an A record?283
For example, a server that is named centos would have an A record that
mapped the hostname centos to the IP address assigned to it:
centos.delledu.lab A 192.168.3.3 Where centos is the hostname,
delledu.lab is the domain name, and centos.delledu.lab is the FQDN.
284Companies that want to divide their domain into sub domains use NS
records. Sub domains indicate a delegation of a portion of the domain
name to a different group of name servers. You create NS records to point
the name of this delegated sub domain to different name servers.
You must create an address (A) record in DNS for the SmartConnect
service IP. Delegating to an A record means that if you failover the entire
cluster, you can do so by changing one DNS A record. All other name
server delegations can be left alone. In many enterprises, it is easier to
update an A record than a name server record, because of the perceived
complexity of the process.
Delegation recommendation.285
3: All clients are configured to make requests from the resident DNS
server using a single DNS hostname. Because all clients reference a
single hostname, isilon.xattire.com, it simplifies the management for large
numbers of clients.
4: The resident DNS server forwards the delegated zone lookup request
to the delegated zone server of authority, here the SIP address of the
cluster.
7: The client then connects to the appropriate cluster node using their
protocol.
NFS Connectivity
NFS relies upon remote procedure call (RPC) for client authentication and
port mapping. RPC is the NFS method that is used for communication
between a client and server over a network. RPC is on Layer 5 of the OSI
model. Because RPC deals with the authentication functions, it serves as
gatekeeper to the cluster.
NFS connectivity
287 When the server receives the CALL, it performs the service that is
requested and sends back the REPLY to the client. During a CALL and
REPLY, RPC looks for client credentials, that is, identity and permissions.
288 If the server is not running a compatible version of the RPC protocol, it
Video
Movie:
The web version of this content contains a movie.
HDFS Topic
• Data Lakes and Analytics
• HDFS Overview Video
• OneFS with Hadoop
• OneFS vs. Hadoop
• HDFS Administration
• Best Practices Resources
number of the RPC program it wants to call and the IP port number.
When a node boots, it first checks its own vault resources before querying
its paired node. This way if the node can recover its journal from its own
resources, there is no need to query the paired node. But, if the journal is
bad, the node can identify the journal condition from its node state block
data, and recovery should be possible. There is a consequence to the
nodes running in pairs. If a node runs unpaired, it is under-protected.
Concurrency Examples
The process of striping spreads all write operations from a client291 across
the nodes of a cluster. Each tab illustrates a file that is broken down into
chunks, after which it is striped across disks292 in the cluster along with the
FEC.
291 A client is connected to only one node at a time. However when that
client requests a file from the cluster, the client connected node does not
have the entire file locally on its drives. The client-connected node
retrieves and rebuilds the file using the back-end network.
292 Even though a client is connected to only one node, when that client
saves data to the cluster, the write operation occurs in multiple nodes. The
scheme is true for read operations also.
All files 128 KB or less are mirrored. For a protection strategy of N+1 the
128 KB file has 2 instances, the original data and one mirrored copy.
The example shows a file that is not evenly distributed in 128 KB chunks.
Blocks in the chunk that are not used are free for use in the next stripe
unit. Unused blocks in a chunk are not wasted.
Select each i buttons for information about ingest and OneFS storage.
2: Utilizing PowerScale to hold the Hadoop data gives you all of the
protection benefits of the OneFS operating systems. You can select any of
the data protection levels that OneFS offers giving you both disk and node
fault tolerance.
Movie:
URL:
https://edutube.emc.com/Player.aspx?vno=kVTYthcIg4kBHj0ryts2HA==&a
utoplay=true
Hadoop enables the distributed process in a large data set across clusters
of servers. Hadoop clusters can dynamically scale up and down based on
the available resources and the required service levels. Let’s see a
traditional Hadoop cluster.
The components are the Name Nodes, Secondary Name Nodes and Data
Nodes. The Name Node holds the metadata or the location information for
every file in the cluster.There is also a Secondary Name Node that is a
backup for the Name Node. The secondary Name Node is passive.As the
name implies, the Data Node is where the data resides. Data is spread
across the node with a 3X mirror.A logical compute process runs on each
data node handling compute operations such as, MapReduce that runs
analytic jobs. In a traditional Hadoop only environment, the HDFS is a
read-only file system. As you can imagine, it would be difficult to do
analysis on a data set that constantly changes.Typically, Hadoop data
exists in silos. Production data is maintained on production servers and
then copied to a landing zone server which then imports or ingests the
data into HDFS. It is important to note that the data on HDFS is not
production data, it is copied from another source. Where does the
PowerScale fit into this solution?
connection, and a Data Node connection. For Data Node IP allocation, the
Name Node gets all the IPs in the access zone.
If the first node shows any issues, the client will use the second and then
a third IP to finish their Hadoop jobs. This provides an automated retry for
clients. Data Node load balancing and pipeline write recover fixes
issues where a Data Node runs out of threads. Features are available in
OneFS 8.0.1 and later. In closing, there are 2 top known issues with Name
Node to Data Node IP addresses allocation. First is when there are
multiple access zones for HDFS, the Name Node can give out IP
addresses from a different access zone. Second, opening multiple security
context can cause the error “status: too may files open”, all Data Nodes
are bad errors. The pipeline write recovery feature fixes the security
context issue.
Next a Name Node looks at the rack configuration and gets the IP
addresses for the rack. The Name Node also checks if any IP addresses
are blacklisted. Then the Name Node gives out rack IP addresses first
based on client IP otherwise it returns IP addresses across the entire
zone. Data Node load balancing is a PowerScale feature that allocates IP
addresses from a Name Node. The IP addresses given when metadata is
requested is from the nodes that have the lowest connection count. When
a client sends a request to write to the cluster, PowerScale’s pipeline
write recovery feature provides 3 Data Node IP addresses to the client.
The next policy moves the archive data off the cluster and into the cloud
when data is not accessed for more than 180 days. Stub files that are also
called SmartLinks are created. Stub files consume approximately 8 KB
space on the Isilon cluster. Files that are accessed or retrieved from the
cloud, or files that are not fully moved to the cloud, have parts that are
cached on the cluster and are part of the stub file. The storing of
CloudPools data and user access to data that is stored in the cloud is
transparent to users.
requires this Hadoop front end to do the data analysis. PowerScale holds
the data so that Hadoop, applications, or clients can manipulate it.
HDFS Administration
The graphic shows the WebUI Protocols, Hadoop (HDFS), Settings
page, and the corresponding isi hdfs settings command output.
294 Hadoop assumes that all members of the domain are trusted.
PowerScale supports integrating with AD or LDAP, and gives you the
ability to safely segment access.
295 Each physical HDFS cluster can only support one distribution of
1: The Default block size determines how the HDFS service returns data
upon read requests from a Hadoop compute client. The server-side block
size determines how the OneFS HDFS daemon returns data to read
requests. Leave the default block size at 128 MB. If the customer runs an
older version of HDFS, consider a 64 MB block size. If the block size is set
to high, many read/write errors and performance problems occur. Tune on
setup.
• Visit the Using Hadoop with OneFS Info Hub web page for
documentation.
• Use the Isilon Hadoop tools to create users and groups in the local
provider.
A2000
The A2000 is an ideal solution for high-density, deep archive storage that
safeguards data efficiently for long-term retention. The A2000 stores up to
1280 TB per chassis and scales to over 80 PB in a single cluster
A300
An ideal active archive storage solution that combines high performance,
nearline accessibility, value, and ease of use. The A300 provides between
120 TB to 1.2 PB per chassis and scales to 75 PB in a single cluster. The
A300 includes inline compression and deduplication capabilities.
A3000
An ideal solution for high-performance, high-density, deep archive storage
that safeguards data efficiently for long-term retention. The A3000 stores
up to 1.6 PB per chassis and scales to 100 PB in a single cluster. The
A3000 includes inline compression and deduplication capabilities.
Cache - L1
Client-side cache. L1 cache refers to read transaction requests, or when
a client requests data from the cluster. L1 cache is stored in a segmented
area of the node RAM and as a result is fast. Related to L1 cache is the
write cache or the write coalescer that buffers write transactions from the
client. The write cache is flushed after successful write transactions. In
OneFS, the two similar caches are distinguished based on their read or
write functionality. Client-side caching includes both the in and out client
transaction buffers.
Cache - L2
Cache - L3
Stored on SSDs, L3 cache holds file data and metadata released from L2
cache, effectively increasing L2 cache capacity.
Chimer Nodes
Chimers are nodes which can contact the external NTP servers. By
default, if the cluster has more than three nodes, three of the nodes are
selected as ‘chimers’. If the cluster comprises three nodes or less, only
one node will be selected as a chimer. If no external NTP server is set,
they will use the local clock instead. The other non-chimer nodes will use
the chimer nodes as their NTP servers. The chimer nodes are selected by
the lowest node number which is not excluded from chimer duty.
F200
Provides the performance of flash storage in a cost-effective form factor to
address the needs of a wide variety of workloads. Each node can scale
raw storage capacity from 3.84 TB to 30.72 TB per node and up to 7.7 PB
of raw capacity per cluster. The F200 includes in-line compression and
deduplication. The minimum number of F200 nodes per cluster is three
while the maximum cluster size is 252 nodes. The F200 is best suited for
remote offices, small M&E workloads, small hospitals, retail outlets, IoT,
factory floor, and other similar deployment scenarios.
F600
F800
The F800 is suitable for workflows that require extreme performance and
efficiency. It is an all-flash array with ultra-high performance. It delivers up
to 250,000 IOPS and up to 15 GB/s aggregate throughput in a single
chassis configuration. Also, it delivers up to 15.75M IOPS and 945 GB/s of
aggregate throughput in a 252 node cluster. The raw storage capacity
scales from 96 TB to 924 TB in a single 4U chassis and up to 58 PB in a
single cluster.
F810
The F810 is suitable for workflows that require extreme performance and
efficiency. The F810 also provides high-speed inline data deduplication
and in-line data compression. It delivers up to 3:1 efficiency, depending on
your specific dataset and workload. The F810 delivers up to 250,000 IOPS
and up to 15 GB/sec aggregate throughput in a single chassis
configuration. It delivers up to 15.75M IOPS and 945 GB/s of aggregate
throughput in a 252 node cluster. The raw storage capacity scales from
230 TB to 924 TB in a 4U chassis and up to 58 PB in a single cluster.
F900
File Provider
A file provider enables you to supply an authoritative third-party source of
user and group information to a cluster. A third-party source is useful in
UNIX and Linux environments that synchronize /etc/passwd,
/etc/group, and etc/netgroup files across multiple servers.
Groupnet
Groupnet
The groupnet is a top-level networking container that manages hostname
resolution against DNS nameservers and contains subnets and IP
address pools. Groupnets are how the cluster communicates with the
world. The groupnet specifies which networking properties the Active
Directory provider will use when communicating with external servers. If
the cluster communicates to another authentication domain, it must find
that domain. To find another authentication domain, you need a DNS
setting to route to that domain. With OneFS 8.0 and later releases,
groupnets can contain individual DNS settings.
H400
The H400 provides a balance of performance, capacity, and value to
support a wide range of file workloads. It delivers up to 3 GB/s bandwidth
per chassis and provides capacity options ranging from 120 TB to 960 TB
per chassis. The H400 uses a medium compute performance node with
SATA drives.
H500
H5600
The H5600 combines massive scalability – 1.28 PB (raw) per chassis and
up to 8 GB/s bandwidth in an efficient, highly dense, deep 4U chassis. The
H5600 delivers inline data compression and deduplication. It is designed
to support a wide range of demanding, large-scale file applications and
workloads.
H600
The H600 is designed to provide high performance at value, delivers up to
120,000 IOPS and up to 12 GB/s bandwidth per chassis. It is ideal for
high-performance computing (HPC) workloads that don’t require the
extreme performance of all-flash. These are spinning media nodes with
various levels of available computing power - H600 combines our turbo
compute performance nodes with 2.5" SAS drives for high IOPS
workloads.
H700
Provides maximum performance and value to support demanding file
workloads. The H700 provides capacity up to 1.2 PB per chassis and 75
PB per cluster. The H700 includes inline compression and deduplication
capabilities.
H7000
Provides versatile, high-performance, high-capacity hybrid platform with
up to 1.6 PB per chassis and 100.8 PB per cluster. The deep-chassis
based H7000 is ideal to consolidate a range of file workloads on a single
platform. The H7000 includes inline compression and deduplication
capabilities.
Home Directory
Home directory provisioning creates a single home share that redirects
users to their SMB home directories. If one does not exist, a directory is
automatically created.
isi get
Job - Schedule
With the Schedule options, you can start the job manually or set to run on
a regularly scheduled basis.
Layers of Access
Protocol Layer - The first layer is the protocol layer. Protocols may be
Server Message Block, or SMB, Network File System, or NFS, File
Transfer Protocol, or FTP, or some other protocol.
Authentication Layer - The authentication layer identifies a user using a
system such as NIS, local files, or Active Directory.
Identity Assignment Layer - The third layer is identity assignment. This
layer is based on the results of the authentication layer, but there are
some cases that need identity mediation within the cluster, or where roles
are assigned within the clusters that are based on user identity.
Authorization Layer - Finally, based on the established connection and
authenticated user identity, the file and directory permissions are
evaluated. The evaluation determines whether the user is entitled to
perform the requested data activities.
MTTDL
MTTDL is a statistical calculation that estimates the likelihood of a
hardware failure resulting in data loss. MTTDL is a system view of
reliability and asks the question “What happens when hardware does fail,
and will I lose any data when it does?”
NFS
Network File System, or NFS, is an open standard that UNIX clients use.
The NFS protocol enables a client computer to access files over a
network. NFS clients mount the OneFS export that is accessible under a
client mountpoint. The mountpoint is the directory that displays files from
the server. The NFS service enables you to create as many NFS exports
as needed.
NFS Failover
NFS enables clients to transparently fail over to another node when a
network or node fails. The failover ability enables movement from one
node to another and no manual intervention on the client side. Movement
to another node enables a continuous workflow from the client side with
no appearance or disruption to their working time.
OneFS Multi-Tenancy
With OneFS, multi-tenancy enables the PowerScale cluster to
simultaneously handle more than one set of networking configurations.
Multi-Tenant Resolver, or MTDNS is a subset of multi-tenancy that
pertains to hostname resolution against DNS name servers. Each tenant
on the cluster can have its own network settings. Before OneFS 8.0, you
could only define one set of DNS servers on the cluster.
Quotas - Accounting
Quotas - Advisory
Advisory quotas do not deny writes to the disk, but they can trigger alerts
and notifications after the threshold is reached.
Quotas - Enforcement
Enforcement quotas include the functionality of accounting quotas and
enable the sending of notifications and the limiting of disk storage.
Reed-Solomon
OneFS uses the Reed-Solomon algorithm, which is an industry standard
method to create error-correcting codes, or ECC, at the file level.
Retention Period
Scale-out Solution
Not all clustered NAS solutions are the same. Some vendors overlay a
management interface across multiple independent NAS boxes. An
overlay gives a unified management interface but does not unify the file
system. While this approach does ease the management overhead of
traditional NAS, it still does not scale well.
Scale-up Solution
The two controllers can run active/active or active-passive. For more
capacity, add another disk array. Each of these components is added
individually. As more systems are added, NAS sprawl becomes an issue.
Scale-up Storage
Scale-up storage is an architecture type that is common in the enterprise
space. High performance, high availability single systems that have a fixed
capacity ceiling characterize scale-up.
SmartCache
SmartCache is a globally coherent read and write caching infrastructure
that provides low latency access to content. Like other resources in the
cluster, as more nodes are added, the total cluster cache grows, enabling
OneFS to deliver predictable, scalable performance within a single
filesystem. OneFS write caching uses write buffering to aggregate, or
coalesce, multiple write operations to the NVRAM file systems journals so
that they can be written to disk safely and more efficiently. This form of
buffering reduces the disk write penalty which could require multiple reads
and writes for each write operation.
SmartDedupe
OneFS deduplication saves a single instance of data when multiple
identical instances of that data exist, in effect, reducing storage
consumption. Deduplication can be done at various levels: duplicate files,
duplicate blocks in files, or identical extents of data within files. Stored
data on the cluster is inspected, block by block, and one copy of duplicate
blocks is saved, thus reducing storage expenses by reducing storage
consumption. File records point to the shared blocks, but file metadata is
not deduplicated.
SmartLock WORM
SmartLock provides WORM (write-once/read-many) status on files. In a
WORM state, files can be read but not modified. "Committing" a file is
changing a file from a read/write state to a WORM state that has a
retention expiration date. Files are committed to a WORM state when
using SmartLock.
SmartPools
SmartPools is a software module that enables administrators to define and
control file management policies within a cluster.
SnapshotIQ
OneFS snapshots are used to protect data against accidental deletion and
modification. Because snapshots are available locally, users can restore
their data without administrative intervention.
System Zone
The default access zone of the cluster is ”System,” and it uses an internal
authentication provider. You can configure external providers for the
System access zone.