4 Managing Oracle Key Vault Multi-Master Clusters

You can create, configure, manage, and administer an Oracle Key Vault multi-master cluster by using the Oracle Key Vault management console.

4.1 About Managing Oracle Key Vault Multi-Master Clusters

You can add or remove nodes from the cluster, disable or enable cluster nodes, and manage activities such as node conflicts and replication.

4.2 Setting Up a Cluster

After you add an initial node to a cluster, you can add more nodes, thereby creating read-write pairs of nodes or read-only nodes.

4.2.1 About Setting Up a Cluster

You create a multi-master cluster by converting a single Oracle Key Vault server to become the initial node.

This Oracle Key Vault server seeds the cluster data and converts the server into the first cluster node, which is called the initial node.

After the initial cluster is created in the Oracle Key Vault server, you can add the different types of nodes that you need for the cluster. The cluster is expanded when you induct additional Oracle Key Vault servers, and add them as read-write nodes, or as simple read-only nodes.

4.2.2 Creating the First (Initial) Node of a Cluster

To create a cluster, you must convert an existing standalone Oracle Key Vault server to become the first node in the cluster.

This first node is called the initial node. The standalone Oracle Key Vault server can also be a server that has been upgraded to Oracle Key Vault release 18.1 or later from a previous release, or it can also be the server that is unpaired from a primary-standby configuration. Check Oracle Key Vault Release Notes for known issues about unpair operations and upgrades.

You can use this node to add one or more nodes to the cluster. The node operates in read-only restricted mode until it is part of a read-write pair.

  1. Perform a full backup of the Oracle Key Vault server to a remote destination.
  2. Log into the Oracle Key Vault management console as a user who has the System Administrator role.
  3. If the Oracle Key Vault server was upgraded from a release earlier than Oracle Key Vault release 12.2 (bundle patch 8), then generate and activate (rotate) a new certificate for the node.
    If this step does not apply to your deployment, then you can bypass it.
  4. Select the Cluster tab, and then select Configure from the left navigation bar.
  5. In the Configure as Candidate Node page, enter the following settings:
    • First Node of Cluster: Select the Yes button.

    • Node Name: Enter a unique name for this node. This field is auto-populated with the Oracle Key Vault server name as the node name. You may specify another name. You cannot change this name later.

    • Cluster Name: Enter a name for this cluster of nodes. You cannot change this name later.

    • Cluster Subgroup: Enter a name for this sub-group of nodes, such as a data center name or a logical group name.

  6. Click Convert to Candidate Node.
The conversion can take several minutes. After the conversion is complete, the Cluster Management page is displayed and the node is now operating in read-only restricted mode. At this stage, you can add another node to this cluster. In that case, the initial node will no longer be a candidate node: it will become the controller node and the new node will be the candidate node. The data on the controller node will overwrite the data on the candidate node.

4.2.3 Adding Nodes to a Cluster

After you have created a cluster to have the initial node, you can add more nodes to the cluster.

4.2.3.1 Adding a Node to Create a Read-Write Pair

After you create the initial node, you must add an additional read-write peer to the cluster.

You can configure any two nodes as a read-write pair. However, any single node can be read-write paired with only one other node.

To create a read-write pair using two nodes in a cluster, you pair a node (referred to as the controller node) with a newly configured server (referred to as the candidate node). Note that this will take some time: an hour or more, depending on the speed of your server, network, and volume of data in the cluster. Be aware that the controller node will be unable to service endpoints during certain parts of this operation.

  1. Perform a full backup of the controller node to a remote destination before continuing.
  2. Ensure that the following network requirements are in place:
    • There is good network connectivity between the servers that host the controller node and the candidate node.
    • The ports that are required for Oracle Key Vault are open in the network firewall. These ports are described in Oracle Key Vault Installation and Upgrade Guide .
  3. As a user who has the System Administrator role, log in to the Oracle Key Vault management console of the node that you want to use as a controller node .
    You can use any existing node, including the first node, that does not have a read-write peer to be the controller node for this operation. If necessary, add a read-only node first.
  4. Select the Cluster tab, and then select Management from the left navigation bar.
  5. Under Cluster Details, select Add.
  6. In the Add Candidate Node to Cluster page, under Add Cluster Details, enter the recovery passphrase in the Recovery Passphrase of the Cluster field.
    This value will be used later when you pair with the candidate node. The recovery passphrase is that of the first node of the cluster and will be used later across all the cluster nodes.
  7. Enter the following details under Add Candidate Node Details.

    While you enter these details, do not save any of this information or click the Add Node button until you reach Step 14.

    • Add Node as Read-Write Peer: Select Yes.
    • Node ID: Select a unique ID for the candidate node. Remember that after you create this ID, you cannot change it. Node ID is auto populated but you may change it. Ensure that the candidate node ID is unique in the cluster.

    • Node Name: Enter a unique name of the candidate node. After you create this name, you cannot change it.

    • Cluster Subgroup: Enter the sub-group name for the candidate node. You can provide an existing subgroup name. If you provide a subgroup name that does not exist, it will be created. (This field is auto-populated with the cluster subgroup name of the controller node.)

    • Cluster Name is auto-populated and cannot be changed.

    • IP Address: Enter the IP address of the candidate node.

    • Certificate of Candidate Node: The next steps explain how you can find the certificate of the candidate node.

    Do not exit this page.

  8. In a new browser window, log into the Oracle Key Vault management console of the candidate node as a user who has the System Administrator role.
  9. Select the Cluster tab, and then select Configure from the left navigation bar.
  10. In the Configure as Candidate Node page, enter the following details:
    • First Node of the Cluster: Select No. Selecting No shows additional fields to enter.
    • Recovery Passphrase of the Cluster: enter the recovery passphrase of the cluster that you entered earlier for the controller node.
    • IP Address: enter the IP address of the controller node.
    • Certificate of Controller Node: Use these steps to enter certificate of controller node.
      • In the browser window for the controller node, scroll to the bottom of the screen. Select and copy the entire certificate value shown for Certificate of Controller Node.
      • In the browser window for the candidate node, paste the copied certificate from the controller node into the Certificate of the Controller Node field.
    • Check the recovery passphrase, the IP address, and the pasted in certificate very carefully to ensure that you copied it correctly. If there is an error, then after you click Convert to Candidate Node, you will need to terminate the pairing process or potentially reinstall Oracle Key Vault on this node.
  11. Click Convert to Candidate Node.
    The conversion can take several minutes. After the conversion is complete, the screen will refresh and the Adding Candidate Node to Cluster page is displayed. The certificate for the candidate node appears on this page.
  12. Select and copy the entire candidate node certificate.
  13. In the browser window of the controller node, paste the copied certificate from the candidate node into the Certificate of Candidate Node box.
  14. Click Add Node.

    This process will take an hour or more, depending on the speed of your server, network, and volume of data in the cluster. During this time, the network management interface of the Oracle Key Vault will be restarted and you might momentarily get a Server Error 500 on the controller node. On the candidate node, errors may also appear, such as Bad Gateway. The candidate node will restart as part of the induction process. This is normal. During the pairing process, the status of the candidate node will display as PAIRING on all cluster nodes.

    To view the status of any server, view the output on the management console.

    After the candidate node restarts and completes the pairing process, you can log in to either cluster node to view the cluster status by selecting the Cluster tab. Both nodes will now show as ACTIVE. The candidate node may briefly display that it is in read-only restricted mode after it automatically restarts. This node is now a synchronously paired cluster node and no longer a candidate node. After a node is part of a cluster, the console will display the node name, subgroup name, and cluster name in the top right area of the console header.

4.2.3.2 Adding a Node as a Read-Only Node

To add a new read-only cluster node, you add a newly configured server from any existing cluster node.

The existing cluster node is referred to as the controller node, and the newly configured server is referred to as the candidate node. This process will take an hour or more, depending on the speed of your server, network, and volume of data in the cluster.

  1. Perform a server backup before continuing.
  2. Ensure that the following network requirements are in place:
    • There is good network connectivity between the servers that host the controller node and the candidate node.
    • The ports that are required for Oracle Key Vault are open in the network firewall. These ports are described in Oracle Key Vault Installation and Upgrade Guide.
  3. Log into the controller node Oracle Key Vault management console as a user who has the System Administrator role.
    You can use any existing node as a controller for this operation.
  4. Select the Cluster tab, and then select Management from the left navigation bar.
  5. In the Cluster Details section click Add.
  6. In the Add Cluster Details section, enter the cluster recovery passphrase in the Recovery Passphrase of the Cluster field.
    This value will be used later when pairing with the candidate node.
  7. Under Add Candidate Node Details, enter the following information:
    • Add Node as Read-Write Peer: Select No.
    • Node ID: Select a unique ID for the candidate node. Remember that after you create this ID, you cannot change it. Node ID is auto populated but you may change it. Ensure that the candidate node ID is unique in the cluster.
    • Node Name: Enter a unique name of the candidate node. After you create this name, you cannot change it.
    • Cluster Subgroup: Enter the subgroup name for the candidate node. You can provide an existing subgroup name. If you provide a subgroup name that does not exist, then it will be created.
    • Cluster Name: This name is populated automatically.
    • IP Address: Enter the IP address of the candidate node.

    Do not exit this page.

  8. In a new browser window, log into the Oracle Key Vault management console of the candidate node as a user who has the System Administrator role.
  9. Select the Cluster tab, and then select Configure from the left navigation bar.
    The Configure as Cluster Candidate page appears.
  10. In the Configure as Candidate Node page, enter the following details:
    • First Node of the Cluster: Select No. Selecting No shows additional fields to enter.
    • Recovery Passphrase of the Cluster: enter the recovery passphrase of the cluster that you entered earlier for the controller node.
    • IP Address: enter the IP address of the controller node.
    • Certificate of Controller Node: Use these steps to enter certificate of controller node.
      • In the browser window for the controller node, scroll to the bottom of the screen. Select and copy the entire certificate value shown for Certificate of Controller Node.
      • In the browser window for the candidate node, paste the copied certificate from the controller node into the Certificate of the Controller Node field.
    • Check the recovery passphrase, the IP address, and the pasted in certificate very carefully to ensure that you copied it correctly. If there is an error, after you click Convert to Candidate Node, you will need to terminate the pairing process or potentially reinstall Oracle Key Vault on this node.
  11. Click Convert to Candidate.
    The conversion can take several minutes. After the conversion is complete, the screen will refresh and the Adding Candidate Node to Cluster page is displayed. The certificate for the candidate node appears on this page.
  12. Select and copy the entire candidate node certificate.
  13. In the browser window of the controller node, paste the copied certificate from the candidate node into the Certificate of Candidate Node box.
  14. Click Add Node.

    This process will take an hour or more, depending on the speed of your server, network, and volume of data in the cluster. During this time, the Oracle Key Vault console of the controller node will become unresponsive and can display an error such as Server Error 500. On the candidate node, errors may also appear, such as Bad Gateway. The candidate node will restart as part of the synchronization process. This is normal. During the pairing process, the status of the candidate node will display as PAIRING on all other cluster nodes not involved in this pairing process.

    To view the status of any server, view the output on the server console.

    After the candidate node restarts and completes the pairing process, you can log into either cluster node to view the cluster status by selecting the Cluster tab. Both nodes will now show as ACTIVE. The candidate node may briefly display that it is in read-only restricted mode after it automatically restarts.  This node is now a read-only paired cluster node and no longer a candidate node. After a node is part of a cluster, the console will display the node name, sub-group name, and cluster name in the top right area of the console header.

4.2.3.3 Creating an Additional Read-Write Pair in a Cluster

Any node can be read-write paired with only one other node, and there can be multiple read-write pairs in a cluster.

  1. Select a read-only cluster node as the controller node to pair with a new candidate node.
  2. Follow the steps to create a read-write pair of nodes in a cluster.

4.3 Terminating the Pairing of a Node

On the controller node, you can terminate the pairing process for a new node.

Be aware that when the controller node performs this operation, then it will experience network changes that will temporarily prevent it from serving endpoints.
  1. On the controller node, in the Status section of the Adding Candidate Node to Cluster page, click the Abort button.
  2. A dialog with the message Are you sure you want to ABORT the addition of the new node? will appear. Select OK.
    After the pairing process terminates, you will be returned to the Add Node to Cluster page on the controller node.
If you terminate the pairing of a candidate node from the controller node, then depending on how far the candidate node is into the induction process, you may or may not be able to return the candidate node to a standalone state. You can attempt to terminate the pairing process from the candidate node. If you are unable to terminate the pairing process from the candidate node, then the candidate node is no longer usable in its current state, and you must re-image the candidate node.

4.4 Disabling a Cluster Node

You can temporarily disable a cluster node, which is required for upgrades and maintenance.

However, be aware that a node can only be disabled for a set period of time. When it exceeds that time, it cannot be enabled again. The default maximum disable node duration time is 24 hours, but you can set it for as high as 240 hours. Note that as this value is increased, the average amount of disk space consumed by cluster-related data also increases.

Do not attempt disabling a node unless replication between all nodes is healthy. Failure to do so will result in a node stuck in the DISABLING state. To return such a node back to the ACTIVE state, you can cancel the disable operation by clicking the Cancel Disable button.
  1. Log into any cluster Oracle Key Vault management console as a user who has the System Administrator role.
  2. Select the Cluster tab, and then select Management from the left navigation bar.
  3. Under Cluster Details, in the Select Node column, select the check box of the node to disable.
  4. Click Disable.
    On the node that is being disabled, the node status will display as DISABLING during the disabling process. The other nodes will display the status for this node as DISABLED. When the disabling process is complete, the node that you disabled displays the DISABLED status.

4.5 Enabling a Disabled Cluster Node

You can enable any cluster node that was previously disabled. You must perform this operation from the disabled node.

  1. As a user who has the System Administrator role, log into the Oracle Key Vault management console of any active node in the cluster.
  2. Select the Cluster tab, and then select Management from the left navigation bar.
  3. In the Cluster Details section, note the dates in which the nodes have been disabled.
    Oracle recommends that you enable the nodes in the reverse order in which they were disabled. Otherwise, the enabling action may not be able to complete.
  4. In the Cluster Details section, under Name, click the name of the node that was disabled most recently.
    Clicking the node name enables you to log in to this node. You can only enable disabled nodes from the disabled node itself.
  5. Select Enable.
    You do not need to check the check box of the disabled node in the Cluster Details table.
  6. Repeat this step for each disabled node, from the most recent to the node that was disabled first.

4.6 Deleting a Cluster Node

You can permanently delete a node from the cluster.

Deleted nodes cannot be added back to any cluster, not just to the current cluster from which they were deleted. However, you can reinstall the Oracle Key Vault appliance software on this server and add the deleted node to a cluster. All data will be synchronized with the cluster before the node is deleted. A node cannot delete itself. Be aware that if a deleted node has a read-write peer, then this read-write peer node will experience network changes that will temporarily prevent it from serving endpoints.

  1. As a user who has the System Administrator role, on a different node, log into the Oracle Key Vault management console.
    A node cannot delete itself.
  2. Select the Cluster tab, and then select Management from the left navigation bar.
  3. Under Cluster Details, in the Select Node column, select the check box of the node to delete.
  4. Click Delete.
    The node status will display as DELETING. After it is deleted, it will show as DELETED, or later be removed from the cluster management page.

    This action is immediate. The node status will display as DELETING. Do not shut down the deleted server until it no longer shows in the Cluster Details table on the Cluster Management page. However, Oracle recommends that you wait an hour after deleting a cluster node before reusing the node ID of the node that was deleted.

4.7 Force Deleting a Cluster Node

You can permanently force delete a node from a cluster that is dead, unresponsive, or has exceeded the maximum disabled node time limit.

Forcefully deleting a node that is still a part of a cluster may cause inconsistency in the cluster. Be aware that if the read-write peer of the node that was forcefully deleted is also removed from the cluster before confirming that all critical data from the forcefully deleted node has reached other nodes, then data loss can result. When you forcefully delete a node, ensure that the node to be deleted has first been shut down. A node cannot be deleted from its own management console. When you must forcefully delete a node, ensure that the node to be deleted has first been shut down. Deleted nodes cannot be added back to the cluster. However, you can reinstall the Oracle Key Vault appliance software on a server and then add the deleted node to a cluster. Be aware that if a deleted node has a read-write peer, then this read-write peer node will experience network changes that will temporarily prevent it from serving endpoints.

  1. On a different node, log into the Oracle Key Vault management console as a user who has the System Administrator role.
    A node cannot delete itself. Oracle recommends that if the node to be deleted has a read-write peer, to force delete the node from its read-write peer.
  2. Select the Cluster tab, and then select Management from the left navigation bar.
  3. Under Cluster Details, in the Select Node column, select the check box of the node to disable.
  4. Click Force Delete.
    The node status will display as DELETING. After it is deleted, it will show as DELETED, or later be removed from the cluster management page. Oracle recommends that you wait an hour after force deleting a cluster node before reusing the node ID of the node that was deleted.

4.8 Managing Replication Between Nodes

You can enable and disable node replication from the Oracle Key Vault management console.

4.8.1 Restarting Cluster Services

While managing replication between nodes, you can restart a node's cluster services when the cluster service status for the node is down.

  1. Log into Oracle Key Vault management console of any cluster node as a user who has the System Administrator role.
  2. Select the Cluster tab, and then Monitoring from the left navigation bar.
  3. In the Node State pane, click the Restart Cluster Services button.

4.8.2 Disabling Node Replication

You can disable the replication link between the current node and any other node in a cluster.

  1. Log into Oracle Key Vault management console of any cluster node as a user who has the System Administrator role.
  2. Select the Cluster tab, and then Monitoring from the left navigation bar.
  3. Under Cluster Link State, select the check boxes for the nodes for which you want to disable replication.
  4. Click Disable.
  5. Click OK to confirm in the dialog box.

4.8.3 Enabling Node Replication

You can enable the replication link between the current node and any other node in a cluster.

  1. As a user who has the System Administrator role, log in to the Oracle Key Vault management console of the node for which replication should be managed.
  2. Select the Cluster tab, and then Monitoring from the left navigation bar.
  3. Under Cluster Link State, select the check boxes for the nodes for which you want to enable replication.
  4. Click Enable.
  5. Click OK to confirm in the dialog box.

4.9 Cluster Management Information

The Cluster Management page provides a concise overview of the cluster and the status of each node. 

You can also manage the cluster from the cluster details section. When a node is performing a cluster operation it becomes the controller node.

Be aware that because the replication across the cluster takes time, there may be a delay before the Cluster Management page refreshes with the new cluster status. The replication lag in the monitoring page will help estimate the delay.

To view the Cluster Management page, click the Cluster tab, and then Management  from the left navigation bar.

Description of 21_cluster_management.png follows
Description of the illustration 21_cluster_management.png

Cluster Information

  • Cluster Name: The name of the cluster.

  • Cluster Subgroups: All subgroups within the cluster.

  • Maximum Disable Node Duration: The maximum time, in hours, that a node can be disabled before it can no longer be enabled.

  • Cluster Version: The version of Oracle Key Vault in which the cluster is operating.

Current Node Information

  • Node Name: The name of this node.

  • Node Type: The type of node, such as whether it is read-only or read-write.

  • Cluster Subgroup: The subgroup to which this node belongs.

Cluster Details

  • Select Node: Used to select a node for a specific operation, such as delete, force delete, or disable.

  • Node ID: The ID of the node.

  • Node Name: The name of the node. Clicking the node name takes you to the Cluster Management page of that node.

  • IP Address: The IP address of the node.

  • Mode: The mode in which the node is operating, such as read-write or read-only restricted.

  • Status: The status of the node, such as active, pairing, disabling, disabled, enabling, deleting, or deleted.

  • Read-Write Peer: The read-write peer of the node. If blank, it has no read-write peer.

  • Cluster Subgroup: The subgroup to which the node belongs. You can change this by 1) checking the check box next to a node, 2) clicking the Edit button, which displays a window, 3) entering a new cluster subgroup in the field, and 4) clicking Save.

  • Join Date: The date and time that the node was added to the cluster or most recently enabled

  • Disable Date: The date and time that the node was disabled.

  • Node Version: The current version of the Oracle Key Vault node.

4.10 Cluster Monitoring Information

The Cluster Monitoring page provides the replication health of the cluster and the current node.

This page also provides a concise overview of the settings enabled in the cluster. You cannot update the settings on this page. Because the replication across the cluster takes time, there may be a delay before the Cluster Monitoring page refreshes with the new cluster state. Replication lag will help estimate the delay.

To view the cluster monitoring page, click the Cluster tab, and then Monitoring from the left navigation bar.

You can hover the mouse over the checkmarks or X's in the Cluster Settings State pane. It will display one of the following explanations of the state:
  • Enabled in Cluster
  • Enabled in Node
  • Disabled in Cluster
  • Disabled in Node

Description of 21_cluster_monitoring_information.png follows
Description of the illustration 21_cluster_monitoring_information.png

Cluster Link State

  • Select Node: Used to select nodes for a specific operation, such as enabling or disabling replication. You can click the checkbox on the label row to select all nodes.

  • Node ID: The ID of the node.

  • Node Name: The name of the node.

  • State: The current state of the node. The server is either up or down.

  • Heartbeat Lag: The amount of time since a heartbeat was last received from this node. This setting should be around 120 seconds or lower. If this value is consistently greater than 120 seconds, then it means that one or more replication links may be broken. After the replication problem is fixed, it will then trend back down towards 120 or lower.

  • Replication Lag: The average time it takes for data to replicate from this node to the current node.

  • Enable: Enables the replication between the current node and the node selected.

  • Disable: Disables the replication between the current node and the node selected.

Cluster Settings State

  • Node ID: The ID of the node.

  • Node Name: The name of the node.

  • Audit: Indicates if auditing is enabled or disabled.

  • FIPS: Indicates if FIPS mode is enabled or disabled.

  • HSM: Indicates if HSM integration is enabled or disabled.

  • SNMP: Indicates if SNMP is enabled or disabled.

  • SYSLOG: Indicates if syslog is enabled or disabled.

  • DNS: Indicates if DNS is enabled or disabled.

4.11 Naming Conflicts and Resolution

Oracle Key Vault can resolve naming conflicts that can arise as users create objects such as endpoints, endpoint groups, and user groups.

4.11.1 About Naming Conflicts and Resolution

If you create an object that has the same name as another object on another node, Oracle Key Vault resolves this conflict.

You can create a new object with a name that conflicts with an object of the same type created on another node. If a conflict happens, then Oracle Key Vault will make the name of the conflicting object unique by adding _OKVxx, where xx is the node ID of the node on which the object was created. You can choose to accept this new name or change the object name.

Users who have the System Administrator role can resolve the following naming conflicts:

  • User names
  • Endpoint names

Users who have the Key Administrator role can resolve the following naming conflicts:

  • Endpoint groups
  • Security objects
  • User groups
  • Wallets

If an object is stuck in the PENDING state and will not transition to ACTIVE, then check for any broken replication links in the cluster. You can find cluster links in the Oracle Key Vault management console by selecting the Cluster tab and then selecting Monitoring.

4.11.2 Naming Conflict Resolution Information

The Cluster Conflict Resolution page provides a list of objects with names that conflict with objects created on different nodes. 

On this page, you can accept the suggested unique name or edit the object name. To view the Cluster Conflict Resolution page, click the Cluster tab, and then Conflict Resolution from the left navigation bar. Alternatively, you can click on the Click here for details button on a Naming Conflict alert from the Alerts table on the Home page.

Wallet Name Conflicts

  • Unique Name: The unique name assigned to the object by the system.

  • Supplied Name: The original name of the object that conflicts with another object of this type.

  • Name Status: The status of the object. The status can be PENDING or ACTIVE.

  • Created By: The user that created the conflicting object name.

  • Creator Node: The node on which the conflicting object was created.

  • Description: The description of the object as entered by the user.

  • Rename: The button that links to the object page where it can be renamed. When you click the edit icon, the Wallet Overview page appears. Click Make Unique to give the wallet a unique name, and then click Accept Rename.

  • Accept: Allows you to accept the suggested name for the selected objects.

4.11.3 Changing the Suggested Conflict Resolution Name

You can change the suggested name for an object that conflicts with another object of the same type.

  1. As a user who has the appropriate administrator role, log in to the Oracle Key Vault management console.
  2. Select the Cluster tab, and then Conflict Resolution from the left navigation bar.
  3. Locate the object that requires a name change.
  4. Click the edit icon to the right of the object, under Rename.
  5. On the object overview page, enter the new name for the object.
  6. Click Save.

4.11.4 Accepting the Suggested Conflict Resolution Name

You can accept the suggested name for an object name that conflicts with another object of the same type.

  1. As a user who has the appropriate administrator role, log in to the Oracle Key Vault management console.
  2. Select the Cluster tab, and then Conflict Resolution from the left navigation bar.
  3. Select the objects for which you want to accept the suggested name.
  4. Click Accept.

4.12 Multi-Master Cluster Deployment Recommendations

Oracle provides deployment recommendations for deployments that have two or more nodes.

Two-Node Deployment Recommendations

Use a two-node deployments for the following situations:

  • Non-critical environments, such as test and development
  • Simple deployment of read-write pairs with both nodes active, replacing classic primary-standby
  • Single data center environments

Considerations for a two-node deployment:

  • Availability is provided by multiple nodes.
  • Maintenance will require downtime.
  • Good network connectivity between data centers is mandatory.
  • Take regular backups to remote destinations for disaster recovery.

Three-Node Deployment Recommendations

Use a three-node deployment for the following situations:

  • Single data center environments with minimal downtime requirement
  • Single read-write pair with additional read-only node to handle load
  • One read-only node is available for zero downtime during maintenance

Considerations for a three-node deployment:

  • Take regular backups to remote destinations for disaster recovery.

Four or More Node Deployment Recommendations

Use a deployment of four or more nodes for the the following situations:

  • Large data centers distributed across geographical locations
  • Deployment of read-write pairs with pair members spanning geography

Considerations for a large deployment:

  • Availability is provided by multiple nodes.
  • Additional read-only nodes can be used to handle load.
  • Good network connectivity between data centers is mandatory.