NetBackup10 AdminGuide HBase
NetBackup10 AdminGuide HBase
Administrator's Guide
Release 10.0
NetBackup™ for HBase Administrator's Guide
Last updated: 2022-02-27
Legal Notice
Copyright © 2022 Veritas Technologies LLC. All rights reserved.
Veritas, the Veritas Logo, and NetBackup are trademarks or registered trademarks of Veritas
Technologies LLC or its affiliates in the U.S. and other countries. Other names may be
trademarks of their respective owners.
This product may contain third-party software for which Veritas is required to provide attribution
to the third party (“Third-party Programs”). Some of the Third-party Programs are available
under open source or free software licenses. The License Agreement accompanying the
Software does not alter any rights or obligations you may have under those open source or
free software licenses. Refer to the Third-party Legal Notices document accompanying this
Veritas product or available at:
https://www.veritas.com/about/legal/license-agreements
The product described in this document is distributed under licenses restricting its use, copying,
distribution, and decompilation/reverse engineering. No part of this document may be
reproduced in any form by any means without prior written authorization of Veritas Technologies
LLC and its licensors, if any.
The Licensed Software and Documentation are deemed to be commercial computer software
as defined in FAR 12.212 and subject to restricted rights as defined in FAR Section 52.227-19
"Commercial Computer Software - Restricted Rights" and DFARS 227.7202, et seq.
"Commercial Computer Software and Commercial Computer Software Documentation," as
applicable, and any successor regulations, whether delivered by Veritas as on premises or
hosted services. Any use, modification, reproduction release, performance, display or disclosure
of the Licensed Software and Documentation by the U.S. Government shall be solely in
accordance with the terms of this Agreement.
http://www.veritas.com
Technical Support
Technical Support maintains support centers globally. All support services will be delivered
in accordance with your support agreement and the then-current enterprise technical support
policies. For information about our support offerings and how to contact Technical Support,
visit our website:
https://www.veritas.com/support
You can manage your Veritas account information at the following URL:
https://my.veritas.com
If you have questions regarding an existing support agreement, please email the support
agreement administration team for your region as follows:
Japan [email protected]
Documentation
Make sure that you have the current version of the documentation. Each document displays
the date of the last update on page 2. The latest documentation is available on the Veritas
website:
https://sort.veritas.com/documents
Documentation feedback
Your feedback is important to us. Suggest improvements or report errors or omissions to the
documentation. Include the document title, document version, chapter title, and section title
of the text on which you are reporting. Send feedback to:
You can also see documentation information or ask a question on the Veritas community site:
http://www.veritas.com/community/
https://sort.veritas.com/data/support/SORT_Data_Sheet.pdf
Contents
Index .................................................................................................................... 50
Chapter 1
Introduction
This chapter includes the following topics:
■ Limitations
SSH communication
BigData policy
HMaster
Application_Type=hbase
Backup Host 1
Region Server 1
Master server
Region Server 2
Backup Host 2
Media server
Region Server 3
Storage
Backup Host 3
Region Server n Parallel Streams HBase plug-in is deployed
HBase cluster on all the backup hosts
...
3 Discovery of 4 Workload
Backup job
workload for backup discovery file 1
is triggered.
2
6 Child
Region Server 2 job 2
Backup Host 2
Region Server 3 3
6 Child
job 3
Backup Host 3 Storage
Region Server n 7 Data is backed up in = Workload
parallel streams n distribution files
HBase Cluster
4. A workload discovery file is created on the backup host. The workload discovery
file contains the details of the data that needs to be backed up from the different
Region servers.
5. The backup host uses the workload discovery file and decides how the workload
is distributed amongst the backup hosts. Workload distribution files are created
for each backup host.
6. Individual child jobs are executed for each backup host. As specified in the
workload distribution files, data is backed up.
7. Data blocks are streamed simultaneously from different Region servers to
multiple backup hosts.
The compound backup job is not completed until all the child jobs are completed.
After the child jobs are completed, NetBackup cleans all the snapshots from the
HMaster. Only after the cleanup activity is completed, the compound backup job is
completed.
See “About backing up a HBase cluster” on page 33.
2
Backup host connects
with HMaster
HMaster
1 Master server
Restore job
is triggered
Region Server 1
Backup host
Region Server 2 4
Storage
Objects are restored on
HBase Cluster the associated Region
servers 3
Restore
Starts
Introduction 10
NetBackup for HBase terminologies
Terminology Definition
■ The backup job runs a discovery job for getting information of the
data to be backed up.
■ Child jobs are created for each backup host that performs the
actual data transfer.
■ After the backup is complete, the job cleans up the snapshots on
the HMaster and is then marked complete.
Discovery job When a backup job is executed, first a discovery job is created. The
discovery job communicates with the HMaster and gathers information
of the block that needs to be backed up and the associated Region
servers. At the end of the discovery, the job populates a workload
discovery file that NetBackup then uses to distribute the workload
amongst the backup hosts.
Child job For backup, a separate child job is created for each backup host to
transfer data to the storage media. A child job can transfer data blocks
from multiple Region servers.
Workload discovery During discovery, when the backup host communicates with the
file HMaster, a workload discovery file is created. The file contains
information about the data blocks to be backed up and the associated
Region servers.
Introduction 11
Limitations
Terminology Definition
Parallel streams The NetBackup parallel streaming framework allows data blocks from
multiple Region servers to be backed up using multiple backup hosts
simultaneously.
Backup host The backup host acts as a proxy client. All the backup and restore
operations are executed through the backup host.
Primary HMaster In a high-availability scenario, you need to specify one HMaster with
the BigData policy and with the tpconfig command. This HMaster
is referred as the primary HMaster.
Fail-over HMaster In a high-availability scenario, the HMaster other than the primary
HMaster that are updated in the hbase.conf file are referred as
fail-over HMaster.
Limitations
Review the following limitations before you deploy the HBase plug-in:
■ Only RHEL and SUSE platforms are supported for HBase clusters and backup
hosts.
■ HBase plug-in does not capture Extended Attributes (xattrs) or Access Control
Lists (ACLs) of an object during backup and hence these are not set on the
restored files or folders.
■ For highly available HBase cluster, if fail-over happens during a backup or restore
operation, the job fails.
Introduction 12
Limitations
■ If you cancel a backup and restore job manually while the discovery job for a
backup operation is in progress, the snapshot entry does not get removed from
the HBase web graphical user interface (GUI).
■ Backup of read-only tables is not supported.
■ You need to add tables one-by-one in the backup selection when you create
the backup policy.
■ Backup and restore operations are not supported with Kerberos authentication
if NB_FIPS_MODE is enabled at the bp.conf.
■ If the CRL expires during the backup of an HTTPS-based Hadoop cluster, the
backup runs partially.
■ If you have multiple CRL-based Hadoop clusters, ensure that you add different
backup hosts for every cluster.
Chapter 2
Deploying HBase plug-in
for NetBackup
This chapter includes the following topics:
Task Reference
Pre-requisites and See “Pre-requisites for installing the HBase plug-in” on page 14.
requirements
Task Reference
Best practices See “Best practices for deploying the HBase plug-in” on page 15.
Verifying the See “Verifying the deployment of the HBase plug-in ” on page 16.
deployment
■ export HADOOP_HOME=PATH_OF_HDFS_DIR
■ export HADOOP_MAPRED_HOME=$HADOOP_HOME
■ export HADOOP_COMMON_HOME=$HADOOP_HOME
■ export HADOOP_HDFS_HOME=$HADOOP_HOME
■ export YARN_HOME=$HADOOP_HOME
■ export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
■ export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin
■ export HADOOP_INSTALL=$HADOOP_HOME
■ export HADOOP_OPTS="$HADOOP_OPTS
-Djava.library.path=$HADOOP_HOME/lib/native"
■ PATH=$PATH:$HBASE_HOME/bin:$JAVA_HOME/bin
■ export CLASSPATH=$CLASSPATH:/usr/local/hadoop/hbase/lib/*
■ Add the entries of all the HMaster and region servers to the /etc/hosts file on
all the backup hosts. You must add the hostname in FQDN format.
Or
Add the appropriate DNS entries in the /etc/resolve.conf file.
■ Always specify the HMaster and region servers in FQDN format.
■ Ping all the nodes (use FQDN) from the backup hosts.
Task Reference
Task Reference
Configuring the See “Configuring the HBase plug-in using the HBase configuration file”
HBase plug-in on page 22.
using the HBase
See “Configuring NetBackup for a highly-available HBase cluster”
configuration file
on page 23.
Configuring the See “Configuration for a HBase cluster that uses Kerberos” on page 26.
backup hosts for
HBase clusters
that use Kerberos
Configuring See “Configuring NetBackup policies for HBase plug-in” on page 26.
NetBackup policies
for HBase plug-in
For more information on how to create a policy, see See “Creating a BigData backup
policy” on page 27.
To add a backup host
1 In the Backup Selections tab, click New and add the backup host in the
following format:
Backup_Host=<hostname>
For more information on how to create a policy, See “Creating a BigData backup
policy” on page 27.
Alternatively, you can also add a backup host using the following command:
bpplinclude PolicyName -add "Backup_Host=hostname"
For more information, See “Using NetBackup Command Line Interface (CLI)
to create a BigData policy for HBase clusters ” on page 29.
2 As a best practice, add the entries of all the HMaster and Region servers to
the /etc/hosts file on all the backup hosts. You must add the host name in
FQDN format.
OR
Add the appropriate DNS entries in the /etc/resolve.conf file.
To remove a backup host
1 In the Backup Selections tab, select the backup host that you want to remove.
2 Right click the selected backup host and click Delete.
Alternatively, you can also remove a backup host using the following command:
bpplinclude PolicyName -delete "Backup_Host=hostname"
■ For Windows
The directory path to the command:
<Install_Path>\NetBackup\bin\admincmd\bpsetconfig
bpsetconfig -h masterserver
bpsetconfig> APP_PROXY_SERVER = clientname1.domain.org
bpsetconfig> APP_PROXY_SERVER = clientname2.domain.org
bpsetconfig>
Windows systems: <ctl-Z>
■ For a highly-available HBase cluster, ensure that the user for the primary and
fail-over HMaster is the same.
■ Use the credentials of the application server that you will use when configuring
the BigData policy.
■ For the HBase cluster that uses Kerberos, specify the actual Kerberos user
name as application_server_user_id value.
■ RSA key must be in the SHA-256 format.
■ Ensure that RSA is supported on the backup host and to obtain the RSA key,
run the following command:
ssh_host_rsa_key.pub | awk '{print $2}' | base64 -d| sha256sum
|awk '{print $1}'
This utility is available at /etc/ssh.
To add credentials in NetBackup
1 Run tpconfig command from the following directory paths:
On UNIX systems, /usr/openv/volmgr/bin/
On Windows systems, install_path\Volmgr\bin\
2 Run the tpconfig --help command. A list of options which are required to
add, update, and delete HBase credentials is displayed.
3 Run the tpconfig -add -application_server application_server_name
-application_server_user_id user_ID -application_type hbase
-password password command by providing appropriate values for each
parameter to add HBase credentials.
For example, if you want to add credentials for HBase server which has
application_server_name as HBase1, then run the following command using
the appropriate <user_ID> and <password> details.
tpconfig -add -application_server HBase1 -application_type hbase
-application_server_user_id HBase -password password
For example:
tpconfig -add -application_server testserver.veritas.com
-application_server_user_id hadoop-application_type hbase
-requiredport 60010 -password hadoop -host_user_id hadoop
-host_password hadoop-host_RSA_key
b2352722053ac9f40bc1XXXXXXXXXXXXXXXXXXXXXXXXX419fa241ba9431fd6b9
Note: Ensure that HBase user has admin permissions for SSH and for HBase
folders.
Note: You must not provide a blank value for any of the parameters, or the backup
job fails.
Note: For non-HA environment, the fail-over parameters are not required.
Configuring NetBackup for HBase 23
Configuring the HBase plug-in using the HBase configuration file
{
"application_servers":
{
"hostname_of_the_primary_HMaster":
{
"failover_HMaster":
[
{
"hostname":"hostname_of_failover_HMaster"
}
]
}
}
{
"application_servers":
{
"hostname_of_primary_HMaster1":
{
"failover_HMaster":
[
{
"hostname": "hostname_of_failover_HMaster1"
}
]
}
}
}
Configuring NetBackup for HBase 25
Configuring the HBase plug-in using the HBase configuration file
2 If you have multiple HBase clusters, use the same hbase.conf file to update
the details. For example,
{
"application_servers":
{
"hostname_of_primary_HMaster1":
{
"failover_HMaster":
[
{
"hostname": "hostname_of_failover_HMaster1"
}
],
},
"hostname_of_primary_HMaster2":
{
"failover_HMaster":
[
{
"hostname": "hostname_of_failover_HMaster2",
}
],
}
}
}
3 Copy this file to the following location on all the backup hosts:
/usr/openv/var/global/
Note: The table names that are specified for backup selection while defining
BigData Policy with Application_Type=hbase must not contain space or
comma in their names.
4 View the details about the new policy using the -L option.
bpplinfo policyname -L
For UNIX:
bpplinclude PolicyName -add 'Application_Type=hbase
7 Specify the backup host on which you want the backup operations to be
performed for HBase.
For Windows:
bpplinclude PolicyName -add "Backup_Host=hostname"
For UNIX:
bpplinclude PolicyName -add 'Backup_Host=hostname'
Note: The backup host must be a Linux computer. The backup host can be a
NetBackup client or a media server or a master server.
8 Specify the HBase directory or folder name that you want to backup.
For Windows:
bpplinclude PolicyName -add "/namespace:table_name"
For UNIX:
bpplinclude PolicyName -add '/namespace:table_name'
Note: The table names that are specified for backup selection while defining
BigData Policy with Application_Type=hbase must not contain space or comma
in their names.
9 Modify and update the policy storage type for BigData policy.
bpplinfo PolicyName -residence STUName -modify
10 Specify the IP address or the host name of the HMaster for adding the client
details.
For Windows:
bpplclients PolicyName -M "MasterServerName" -add
"HBaseServerHMaster" "Linux" "RedHat"
For UNIX:
bpplclients PolicyName -M 'MasterServerName' -add
'HBaseServerHMaster' 'Linux' 'RedHat'
Configuring NetBackup for HBase 31
Disaster recovery of a HBase cluster
11 Assign a schedule for the created BigData policy as per your requirements.
bpplsched PolicyName -add Schedule_Name -cal 0 -rl 0 -st
sched_type -window 0 0
Task Description
After the HBase cluster and nodes are up, Perform the following tasks:
prepare the cluster for operations with
Update firewall settings so that the backup
NetBackup.
hosts can communicate with the HBase
cluster.
The backup hosts use the HBase.conf file With this release, the following plug-in
to save the configuration settings of the settings can be configured
HBase plug-in. You need to create separate
■ See “Configuring NetBackup for a
file for each backup host and copy it to
highly-available HBase cluster”
/usr/openv/var/global/. You need to
on page 23.
create the HBase.conf file in JSON format.
Update the BigData policy with the original See “Configuring NetBackup policies for
HMaster name. HBase plug-in” on page 26.
Chapter 4
Performing backups and
restores of HBase
This chapter includes the following topics:
Task Reference
(Optional) See “Pre-requisite for running backup and restore operations for a HBase
Complete the cluster with Kerberos authentication” on page 34.
pre-requisite for
Kerberos
Task Reference
Best practices See “Best practices for backing up a HBase cluster” on page 35.
Troubleshooting For discovery and cleanup related logs, review the following log file on
tips the first backup host that triggered the discovery.
/usr/openv/var/global/logs/nbaapidiscv
For data transfer related logs, search for corresponding backup host
(using the hostname) in the log files on the master server.
Note: During the backup and restore operations, the TGT must be valid. Thus,
specify the TGT validity accordingly or renew it when required during the operation.
For example,
kinit -k -t /usr/openv/var/global/nbusers/hbase_mykeytabfile.keytab
[email protected]
Considerations
■ On the Hmaster, set the PasswordAuthentication field to Yes in the
/etc/ssh/sshd_config file. After you update the file, restart sshd.
Ensure that all the cluster servers support supports same Hash Key algorithm
( RSA)
■ Snapshots are not cleaned up if you cancel a job manually. After cancelling the
job you must manually delete snapshots from the HBase shell.
■ If you take backup of an empty table, you need to clean the snapshot manually
from the HBase shell.
■ See “Best practices for backing up a HBase cluster” on page 35.
■ Ensure that the local time on the HBase nodes and the backup host are
synchronized with the NTP server.
Task Reference
Complete the See “Pre-requisite for running backup and restore operations for a HBase
pre-requisites for cluster with Kerberos authentication” on page 34.
Kerberos
Restoring HBase ■ See “Using the Restore Wizard to restore HBase data on the same
data on the same cluster” on page 37.
HMaster or HBase ■ See “Using the bprestore command to restore HBase data on the
cluster same HBase cluster” on page 38.
Restoring HBase See “Restoring HBase data on an alternate HBase cluster” on page 39.
data to an
alternate HMaster
or HBase cluster
Best practices See “Best practices for restoring a HBase cluster” on page 44.
Troubleshooting See “About NetBackup for HBase debug logging” on page 45.
tips
Performing backups and restores of HBase 37
About restoring a HBase cluster
Considerations
When you restoring disabled table, the table will be enabled after successful restore.
■ On the Specify NetBackup Machines and Policy Type wizard, enter the
policy type details for restore.
From the Policy type for restores list, choose BigData as the policy type
for restore.
Click Ok.
6 Go to the Backup History and select the backup images that you want to
restore.
7 In the Directory Structure pane, expand the Directory.
All the subsequent files and folders under the directory are displayed in the
Contents of Selected Directory pane.
8 In the Contents of Selected Directory pane, select the check box for the
HBase files that you want to restore.
9 Click Restore.
10 In the Restore Marked Files dialog box, select the destination for restore as
per your requirement.
■ Select Restore everything to its original location if you want to restore
your files to the same location where you performed your backup.
■ Select Restore everything to a different location if you want to restore
your files to a location which is not the same as your backup location.
Where,
-S master_server
Specifies a file (listfile) that contains a list of files to be restored and can be
used instead of the file names option. In listfile, list each file path must be on
a separate line.
-L progress_log
Specifies the name of whitelisted file path in which to write progress information.
-t 44
■ NetBackup supports redirected restores only using the Command Line Interface
(CLI).
■ Make sure that you have added the credentials for the alternate HMaster or
HBase cluster in NetBackup master server.
Performing backups and restores of HBase 41
Restoring HBase data on an alternate HBase cluster
Parameter Value
rename_file Change
/<namespace:source_table_name>
to
/<namespace:destination_table_name>
ALT_APPLICATION_SERVER=<alternate
name node>
Specifies a file (listfile) that contains a list of files to be restored and can be
used instead of the file names option. In listfile, list each file path must be on
a separate line.
-L progress_log
Specifies the name of whitelisted file path in which to write progress information.
-t 44
Specifies the name of a file with name changes for alternate-path restores.
Use the following form for entries in the rename file:
change backup_tablename to restore_tablename
ALT_APPLICATION_SERVER=<Application Server Name>
Note: Ensure that you have whitelisted all the file paths such as
<rename_file_path>, <progress_log_path> that are already not included as
a part of NetBackup install path.
Performing backups and restores of HBase 43
Restoring truncated tables
6 Go to the Backup History and select the backup images that you want to
restore.
7 In the Directory Structure pane, expand the Directory.
All the subsequent files and folders under the directory are displayed in the
Contents of Selected Directory pane.
8 In the Contents of Selected Directory pane, select the check box for the
HBase files that you want to restore.
9 Click Restore.
10 In the Restore Marked Files dialog box, select Restore individual directories
and files to different locations.
11 Select the source HBase directory.
Performing backups and restores of HBase 44
Best practices for restoring a HBase cluster
14 Click OK.
15 Click Start Restore.
16 Verify the restored files.
Workaround:
Download and install the HBase plug-in.
Workaround:
Remove non-existing tables from the backup selection.
Troubleshooting 47
Backup fails with error 6623
Workaround:
HMaster or Data nodes are offline. Ensure that HMaster or Data nodes are online.
Workaround:
Ensure that the destination client is a backup host.
Workaround:
Ensure that the backup host is online and connects to the HMaster.
Workaround
Either the HBase credentials are not added in NetBackup master server, or the
credentials added are invalid.
Ensure that the HBase credentials are added in NetBackup master server.
You can use the NetBackup tpconfig command to add the credentials. See “Adding
HBase credentials in NetBackup” on page 20.
Troubleshooting 48
Backup operation fails with error 6654
If the SSL certificates and CRL path files are stored at a location that is not persistent
the appliance restart, the backups and restore operations fail.
Workaround:
If custom configuration files for Hadoop and HBase get deleted after a restart, you
can manually create the files at the following location:
■ Hadoop:/usr/openv/var/global/hadoop.conf
■ HBase:/usr/openv/var/global/hbase.conf
You can store the CA certificate that has signed the Hadoop or HBase SSL certificate
and CRL at the following location:
/usr/openv/var/global/
{
"application_servers":
{
"primary.host.com":
{
"use_ssl":true
"failover_namenodes":
[
{
"hostname":"secondary.host.com",
"use_ssl":true
"port":11111
}
],
"port":11111
}
},
"number_of_threads":5
}
Index
A L
Adding License 14
backup host 18 Limitations 11
B N
Backup 33–34 NetBackup
backup 8 debug logging 45
BigData policy NetBackup Appliance
Command Line Interface 29 backup host 20
NetBackup Administration Console 27
Policies utility 28 O
Policy Configuration Wizard 27
overview
backup 6
C configuration 6
compatibility deployment 6
supported operating system 14 installation 6
Creating restore 6
BigData backup policy 27
P
D parallel streaming framework 6
disaster recovery 31 policies
configuring 26
H Preparing 14
HBase credentials
adding 20 R
Removing
I backup host 18
Restore 36
Installing
bprestore command 38
verifying 16
restore 9
Restoring 37
K alternate HMaster 39
Kerberos
post installation 26
kerberos
T
terms 10
backup 34
restore 34
W
Whitelisting
backuphost 19