Removing Disks from the File System

Explains how to remove disks using either the Control System or the CLI.

About this task

When you remove a disk from the file system, other disks in the storage pool are also removed automatically from the file system and are no longer in use (they are available but off-line). Their disk storage goes to 0%, and they are eligible to be added again to the file system to build a new storage pool. You can either replace the disk and re-add it along with the other disks that were in the storage pool, or just re-add the other disks if you do not plan to replace the disk you removed. See Adding Disks to file system for more information.

WARNING
Removing a disk in the storage pool that contains Container ID 1 shuts down CLDB, triggering a CLDB failover. Container ID 1 contains CLDB data for the primary CLDB. From the command-line, run the maprcli disk remove command without the -force 1 option first and examine the warning messages to make sure you are not removing the disk with Container ID 1. To safely remove such a disk, perform a CLDB Failover to make one of the other CLDB nodes the primary CLDB, then remove the disk as normal with addition of the -force 1 option.
Before removing or replacing disks, make sure the Replication Alarm (VOLUME_ALARM_DATA_UNDER_REPLICATED), Data Alarm (VOLUME_ALARM_DATA_UNAVAILABLE), Warm-Tier Data Node Down (VOLUME_ALARM_DEGRADED_EC_STRIPES), and EC Degraded Alarm (VOLUME_ALARM_CRITICALLY_DEGRADED_EC_STRIPES) are not raised. These alarms can indicate potential or actual data loss. If either alarm is raised, you might be able to repair the problem using the /opt/mapr/server/fsck utility before removing or replacing disks.
NOTE
Using the /opt/mapr/server/fsck utility with the -r flag to repair a file system risks data loss. Call HPE Ezmeral Data Fabric support before using /opt/mapr/server/fsck -r.

Removing Disks from file system Using the Control System

About this task

Complete the following steps to remove disks using the Control System:

Procedure

  1. Log in to the Control System and go to the Summary tab in the node information page.
  2. Select the disks to remove in the Disks pane and click Remove Disk(s) from File System.
    The Remove Disk(s) from File System confirmation dialog displays.
    WARNING
    One or more disks you selected may have unreplicated data on it and this action will forcefully remove the disks.
  3. Review the list and click Remove Disk.
    Wait several minutes while the removal process completes. After you remove the disks, any other disks in the same storage pools are taken offline and marked as available (not in use by HPE Ezmeral Data Fabric).
  4. Remove the physical disks from the node or nodes according to the correct hardware procedure.
  5. From a command line terminal, remove the failed disk log file from the /opt/mapr/logs directory.
    These log files are typically named like this:
    diskname.failed.info

Removing Disks from file system Using the CLI or REST API

Procedure

  1. On the node, determine which disk to remove/replace by examining Disk entries in the /opt/mapr/logs/faileddisk.log file.
  2. Run the following command, substituting the hostname or IP address for <host> and a list of disks for <disks>
    maprcli disk remove -disks <disk names> -host <host>
    NOTE
    This command does not remove a disk containing unreplicated data unless forced.

    For complete reference information, see disk remove.

  3. Examine the screen output in response to the command you ran in step 2.
    For example:
    maprcli disk remove -host `hostname -f` -disks /dev/sdd
    message   host   disk
    removed.  host1  /dev/sdd
    removed.  host1  /dev/sde
    removed.  host1  /dev/sdf
    Make a note of the additional disks removed when the disk is removed. For example, the disks /dev/sde and /dev/sdf are part of the same storage pool and therefore removed along with the disk (/dev/sdd).
  4. Confirm that the removed disks do not appear in the disktab file.
  5. Remove the disk log file from the /opt/mapr/logs directory.
    For failed disks, these log files are typically named in the pattern diskname.failed.info.

What to do next

When you replace a failed disk, add it back to the file system along with the other disks from the same storage pool that were previously removed. Adding only the replacement disk to the file system, results in a non-optimal storage pool layout, which can lead to degraded performance.

Once you add the disks to the file system, the cluster automatically allocates properly-sized storage pools. For example, if you add ten disks, HPE Ezmeral Data Fabric allocates two storage pools of three disks each and two storage pools of two disks each.