Moving CLDB Data
Describes how to move CLDB data to another node.
About this task
In a Community Edition-licensed cluster, CLDB data must be recovered from a failed CLDB node and installed on another node. The cluster can continue normally as soon as the CLDB is started on another node.
For more information, see CLDB Failover.
Use the maprcli dump zkinfo
command to identify the latest epoch of the
CLDB, identify the nodes where replicates of the CLDB are stored, and select one of
those nodes to serve the new CLDB node. Perform the following steps on any cluster
node:
Procedure
-
Log in as
root
or usesudo
for the following commands. -
Issue the
maprcli dump zkinfo
command using the-json
flag.# maprcli dump zkinfo -json
The output displays the ZooKeeper znodes. -
In the
/datacenter/controlnodes/cldb/epoch/1
directory, locate the CLDB with the latest epoch.
The Latest Epoch field identifies the current epoch of the CLDB data. In this example, the latest epoch is 13.{ "/datacenter/controlnodes/cldb/epoch/1/KvStoreContainerInfo":" Container ID:1 VolumeId:1 Master:10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--13-VALID Servers: 10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--13-VALID Inactive Servers: Unused Servers: Latest epoch:13" }
-
Select a CLDB from among the copies at the latest epoch. For example,
10.250.2.41:5660--13-VALID
indicates that the node has a copy at epoch 13 (the latest epoch).