2 Monitoring and Managing Oracle Private Cloud Appliance

Monitoring and management of the Private Cloud Appliance is achieved using the Oracle Private Cloud Appliance Dashboard. This browser-based graphical user interface is also used to perform the initial configuration of the appliance beyond the instructions provided in the Quick Start poster included in the packaging of the appliance.

Attention:

Before starting the system and applying the initial configuration, read and understand the Oracle Private Cloud Appliance Release Notes. The section Known Limitations and Workarounds provides information that is critical for correctly executing the procedures in this document. Ignoring the release notes may cause you to configure the system incorrectly. Bringing the system back to normal operation may require a complete factory reset.

The Oracle Private Cloud Appliance Dashboard allows you to perform the following tasks:

  • Initial software configuration (and reconfiguration) for the appliance using the Network Environment window, as described in Network Settings.

  • Hardware provisioning monitoring and identification of each hardware component used in the appliance, accessed via the Hardware View window described in Hardware View.

  • Resetting the passwords used for different components within the appliance, via the Password Management window, as described in Authentication.

The Oracle Private Cloud Appliance Controller Software includes functionality that is currently not available through the Dashboard user interface:
  • Backup

    The configuration of all components within Private Cloud Appliance is automatically backed up based on a crontab entry. This functionality is not configurable. Restoring a backup requires the intervention of an Oracle-qualified service person. For details, see Oracle Private Cloud Appliance Backup.

  • Update

    The update process is controlled from the command line of the active management node, using the Oracle Private Cloud Appliance Upgrader. For details, see Oracle Private Cloud Appliance Upgrader. For step-by-step instructions, see Upgrading Oracle Private Cloud Appliance.

  • Custom Networks

    In situations where the default network configuration is not sufficient, the command line interface allows you to create additional networks at the appliance level. For details and step-by-step instructions, see Network Customization.

  • Tenant Groups

    The command line interface provides commands to optionally subdivide a Private Cloud Appliance environment into a number of isolated groups of compute nodes. These groups of servers are called tenant groups, which are reflected in Oracle VM as different server pools. For details and step-by-step instructions, see Tenant Groups.

Guidelines and Limitations

The Oracle Private Cloud Appliance is provided as an appliance with carefully preconfigured hardware and software stacks. Making any changes to the hardware or software of the appliance, unless explicitly told to do so by the Oracle Private Cloud Appliance product documentation or Oracle support, will result in an unsupported configuration. This section lists some of the operations that are not permitted, presented as guidelines and limitations that should be followed when working within the Oracle Private Cloud Appliance Dashboard, CLI, or Oracle VM Manager interfaces. If you have a question about changing anything not explicitly permitted or described in the documentation, contact Oracle to open an SR.

The following actions must not be performed, except when Oracle gives specific instructions to do so.

Do Not:

  • attempt to discover, remove, rename or otherwise modify servers or their configuration;

  • add, remove, or update RPMs on Private Cloud Appliance server components other than those distributed with Private Cloud Appliance software except when specifically approved by Oracle;

  • attempt to modify the NTP configuration of a server;

  • attempt to add, remove, rename or otherwise modify server pools or their configuration;

  • attempt to change the configuration of server pools corresponding with tenant groups configured through the appliance controller software (except for DRS policy setting);

  • attempt to move servers out of the existing server pools, except when using the pca-admin commands for administering tenant groups;

  • attempt to add or modify or remove server processor compatibility groups;

  • attempt to modify or remove the existing local disk repositories or the repository named Rack1-repository;

  • attempt to delete or modify any of the preconfigured default networks, or custom networks configured through the appliance controller software;

  • attempt to connect virtual machines to the appliance management network;

  • attempt to modify or delete any existing Storage elements that are already configured within Oracle VM, or use the reserved names of the default storage elements – for example OVCA_ZFSSA_Rack1 – for any other configuration;

  • attempt to configure global settings, such as YUM Update, in the Reports and Resources tab (except for tags, which are safe to edit);

  • attempt to select a non-English character set or language for the operating system, because this is not supported by Oracle VM Manager – see support document [PCA 2.3.4] OVMM upgrade fails with Invalid FQN (Doc ID 2519818.1);

  • attempt to connect any Private Cloud Appliance component to a customer's LDAP or Active Directory for authentication, including management nodes, compute nodes, ZFS Storage Appliances, NIS, NIS+, and other naming services;

  • attempt to add users – for example – adding users to management nodes or to WebLogic;

  • attempt to change DNS settings on compute nodes or ZFS Storage Appliances. The Oracle Private Cloud Appliance Dashboard contains the only permitted DNS settings.

  • change settings on Ethernet switches integrated into the Oracle Private Cloud Appliance or Oracle Private Cloud at Customer.

  • attempt to add a server running Oracle VM 3.4.6.x to a tenant group that already contains a compute node running Oracle VM 3.4.7.

  • migrate a virtual machine from a compute node running Oracle VM 3.4.7 to a compute node running Oracle VM 3.4.6.x.

If you ignore this advice, the Private Cloud Appliance automation, which uses specific naming conventions to label and manage assets, may fail. Out-of-band configuration changes would not be known to the orchestration software of the Private Cloud Appliance. If a conflict between the Private Cloud Appliance configuration and Oracle VM configuration occurs, it may not be possible to recover without data loss or system downtime.

Note:

An exception to these guidelines applies to the creation of a Service VM. This is a VM created specifically to perform administrative operations, for which it needs to be connected to both the public network and internal appliance networks. For detailed information and instructions, see support document How to Create Service Virtual Machines on the Private Cloud Appliance by using Internal Networks (Doc ID 2017593.1).

There is a known issue with the Oracle Private Cloud Appliance Upgrader, which stops the upgrade process if Service VMs are present. For the appropriate workaround, see support document [PCA 2.3.4] pca_upgrader Check Fails with Exception - Network configuration error: 'inet' (Doc ID 2510822.1).

Connecting and Logging in to the Oracle Private Cloud Appliance Dashboard

To open the Login page of the Oracle Private Cloud Appliance Dashboard, enter the following address in a Web browser:

https://manager-vip:7002/dashboard

Where, manager-vip refers to the shared Virtual IP address that you configured for your management nodes during installation. By using the shared Virtual IP address, you ensure that you always access the Oracle Private Cloud Appliance Dashboard on the active management node.

Figure 2-1 Dashboard Login


Screenshot showing the login page of the Oracle Private Cloud Appliance Dashboard.

Note:

If you are following the installation process and this is your first time accessing the Oracle Private Cloud Appliance Dashboard, the Virtual IP address in use by the active management node is set to the factory default 192.168.4.216 . This is an IP address in the internal appliance management network, which can only be reached if you use a workstation patched directly into the available Ethernet port 48 in the Cisco Nexus 9348GC-FXP Switch.

Important:

You must ensure that if you are accessing the Oracle Private Cloud Appliance Dashboard through a firewalled connection, the firewall is configured to allow TCP traffic on the port that the Oracle Private Cloud Appliance Dashboard is using to listen for connections.

Enter your Oracle Private Cloud Appliance Dashboard administration user name in the User Name field. This is the administration user name you configured during installation. Enter the password for the Oracle Private Cloud Appliance Dashboard administration user name in the Password field.

Important:

The Oracle Private Cloud Appliance Dashboard makes use of cookies in order to store session data. Therefore, to successfully log in and use the Oracle Private Cloud Appliance Dashboard, your web browser must accept cookies from the Oracle Private Cloud Appliance Dashboard host.

When you have logged in to the Dashboard successfully, the home page is displayed. The central part of the page contains Quick Launch buttons that provide direct access to the key functional areas.

Figure 2-2 Dashboard Home Page


Screenshot showing the home page of the Oracle Private Cloud Appliance Dashboard.

From every Dashboard window you can always go to any other window by clicking the Menu in the top-left corner and selecting a different window from the list. A button in the header area allows you to open Oracle VM Manager.

Hardware View

The Hardware View window within the Oracle Private Cloud Appliance Dashboard provides a graphical representation of the hardware components as they are installed within the rack. The view of the status of these components is automatically refreshed every 30 seconds by default. You can set the refresh interval or disable it through the Auto Refresh Interval list. Alternatively, a Refresh button at the top of the page allows you to refresh the view at any time.

During particular maintenance tasks, such as upgrading management nodes, you may need to disable compute node provisioning temporarily. This Disable CN Provisioning button at the top of the page allows you to suspend provisioning activity. When compute node provisioning is suspended, the button text changes to Enable CN Provisioning and its purpose changes to allow you to resume compute node provisioning as required.

Rolling over each item in the graphic with the mouse raises a pop-up window providing the name of the component, its type, and a summary of configuration and status information. For compute nodes, the pop-up window includes a Reprovision button, which allows you to restart the provisioning process if the node becomes stuck in an intermittent state or goes into error status before it is added to the Oracle VM server pool. Instructions to reprovision a compute node are provided in Reprovisioning a Compute Node when Provisioning Fails.

Caution:

The Reprovision button is to be used only for compute nodes that fail to complete provisioning. For compute nodes that have been provisioned properly and/or host running virtual machines, the Reprovision button is made unavailable to prevent incorrect use, thus protecting healthy compute nodes from loss of functionality, data corruption, or being locked out of the environment permanently.

Caution:

Reprovisioning restores a compute node to a clean state. If a compute node was previously added to the Oracle VM environment and has active connections to storage repositories other than those on the internal ZFS storage, the external storage connections need to be configured again after reprovisioning.

Alongside each installed component within the appliance rack, a status icon provides an indication of the provisioning status of the component. A status summary is displayed just above the rack image, indicating with icons and numbers how many nodes have been provisioned, are being provisioned, or are in error status. The Hardware View does not provide real-time health and status information about active components. Its monitoring functionality is restricted to the provisioning process. When a component has been provisioned completely and correctly, the Hardware View continues to indicate correct operation even if the component should fail or be powered off. See Table 2-1 for an overview of the different status icons and their meaning.

Table 2-1 Table of Hardware Provisioning Status Icons

Icon Status Description


Hardware status OK icon

OK

The component is running correctly and has passed all health check operations. Provisioning is complete.


Hardware status Warning icon

Provisioning

The component is running, and provisioning is in progress. The progress bar fills up as the component goes through the various stages of provisioning.

Key stages for compute nodes include: HMP initialization actions, Oracle VM Server installation, network configuration, storage setup, and server pool membership.


Hardware status Error icon

Error

The component is not running and has failed health check operations. Component troubleshooting is required and the component may need to be replaced. Compute nodes also have this status when provisioning has failed.

Note:

For real-time health and status information of your active Private Cloud Appliance hardware, after provisioning, consult the Oracle VM Manager or Oracle Enterprise Manager UI.

The Hardware View provides an accessible tool for troubleshooting hardware components within the Private Cloud Appliance and identifying where these components are actually located within the rack. Where components might need replacing, the new component must take the position of the old component within the rack to maintain configuration.

Figure 2-3 The Hardware View


Screenshot showing the Hardware View window of the Oracle Private Cloud Appliance Dashboard.

Network Settings

The Network Environment window is used to configure networking and service information for the management nodes. For this purpose, you should reserve three IP addresses in the public (data center) network: one for each management node, and one to be used as virtual IP address by both management nodes. The virtual IP address provides access to the Dashboard once the software initialization is complete.

To avoid network interference and conflicts, you must ensure that the data center network does not overlap with any of the infrastructure subnets of the Oracle Private Cloud Appliance default configuration. These are the subnets and VLANs you should keep clear:

Subnets:

  • 192.168.4.0/24 – internal machine administration network: connects ILOMs and physical hosts

  • 192.168.32.0/21 – internal management network: traffic between management and compute nodes

  • 192.168.64.0/21 – underlay network for east/west traffic within the appliance environment

  • 192.168.72.0/21 – underlay network for north/south traffic, enabling external connectivity

  • 192.168.40.0/21 – storage network: traffic between the servers and the ZFS Storage Appliance

Note:

Each /21 subnet comprises the IP ranges of eight /24 subnets or over 2000 IP addresses. For example: 192.168.32.0/21 corresponds with all IP addresses from 192.168.32.1 to 192.168.39.255.

VLANs:

  • 1 – the Cisco default VLAN

  • 3040 – the default service VLAN

  • 3041-3072 – a range of 31 VLANs reserved for customer VM and host networks

  • 3073-3099 – a range reserved for system-level connectivity

    Note:

    VLANs 3073-3088 are in use for VM access to the internal ZFS feature

    VLANs 3090-3093 are already in use for tagged traffic over the /21 subnets listed above.

  • 3968-4095 – a range reserved for Cisco internal device allocation

The Network Environment window is divided into three tabs: Management Nodes, Data Center Network, and DNS. Each tab is shown in this section, along with a description of the available configuration fields.

You can undo the changes you made in any of the tabs by clicking the Reset button. To confirm the configuration changes you made, enter the Dashboard Admin user password in the applicable field at the bottom of the window, and click Apply Changes.

Note:

When you click Apply Changes, the configuration settings in all three tabs are applied. Make sure that all required fields in all tabs contain valid information before you proceed.

Figure 2-4 shows the Management Nodes tab. The following fields are available for configuration:

  • Management Node 1:

    • IP Address: Specify an IP address within your datacenter network that can be used to directly access this management node.

    • Host Name: Specify the host name for the first management node system.

  • Management Node 2:

    • IP Address: Specify an IP address within your datacenter network that can be used to directly access this management node.

    • Host Name: Specify the host name for the second management node system.

  • Management Virtual IP Address: Specify the shared Virtual IP address that is used to always access the active management node. This IP address must be in the same subnet as the IP addresses that you have specified for each management node.

Figure 2-4 Management Nodes Tab


Screenshot showing the Management Nodes tab in the Network Environment window of the Oracle Private Cloud Appliance Dashboard.

Figure 2-5 shows the Data Center Network tab. The following fields are available for configuration:

  • Management Network VLAN: The default configuration does not assume that your management network exists on a VLAN. If you have configured a VLAN on your switch for the management network, you should toggle the slider to the active setting and then specify the VLAN ID in the provided field.

    Caution:

    When a VLAN is used for the management network, and VM traffic must be enabled over the same network, you must manually configure a VLAN interface on the vx13040 interfaces of the necessary compute nodes to connect them to the VLAN with the ID in question. For instructions to create a VLAN interface on a compute node, see Create VLAN Interfaces in the Oracle VM Manager User's Guide.

  • Domain Name: Specify the data center domain that the management nodes belong to.

  • Netmask: Specify the netmask for the network that the Virtual IP address and management node IP addresses belong to.

  • Default Gateway: Specify the default gateway for the network that the Virtual IP address and management node IP addresses belong to.

  • NTP: Specify the NTP server that the management nodes and other appliance components must use to synchronize their clocks to.

Figure 2-5 Data Center Network Tab


Screenshot showing the Data Center Network tab in the Network Environment window of the Oracle Private Cloud Appliance Dashboard.

Figure 2-6 shows the Data Center Network tab. The following fields are available for configuration:

  • DNS Server 1: Specify at least one DNS server that the management nodes can use for domain name resolution.

  • DNS Server 2: Optionally, specify a second DNS server.

  • DNS Server 3: Optionally, specify a third DNS server.

Figure 2-6 DNS Tab


Screenshot showing the DNS tab in the Network Environment window of the Oracle Private Cloud Appliance Dashboard.

You must enter the current Private Cloud Appliance Admin account password to make changes to any of these settings. Clicking the Apply Changes button at the bottom of the page saves the settings that are currently filled out in all three Network Environment tabs, and updates the configuration on each of the management nodes. The ovca services are restarted in the process, so you are required to log back in to the Dashboard afterward.

Functional Networking Limitations

There are different levels and areas of network configuration in an Oracle Private Cloud Appliance environment. For the correct operation of both the host infrastructure and the virtualized environment, it is critical that the administrator can make a functional distinction between the different categories of networking, and knows how and where to configure all of them. This section is intended as guidance to select the suitable interface to perform the main network administration operations.

In terms of functionality, practically all networks operate either at the appliance level or the virtualization level. Each has its own administrative interface: Oracle Private Cloud Appliance Dashboard and CLI on the one hand, and Oracle VM Manager on the other. However, the network configuration is not as clearly separated, because networking in Oracle VM depends heavily on existing configuration at the infrastructure level. For example, configuring a new public virtual machine network in Oracle VM Manager requires that the hosts or compute nodes have network ports already connected to an underlying network with a gateway to the data center network or internet.

A significant amount of configuration – networking and other – is pushed from the appliance level to Oracle VM during compute node provisioning. This implies that a hierarchy exists; that appliance-level configuration operations must be explored before you consider making changes in Oracle VM Manager beyond the standard virtual machine management.

Network Configuration

This section describes the Oracle Private Cloud Appliance and Oracle VM network configuration.

  • Virtual Machine Network

    By default, a fully provisioned Private Cloud Appliance is ready for virtual machine deployment. In Oracle VM Manager you can connect virtual machines to these networks directly:

    • default_external, created on the vx13040 VxLAN interfaces of all compute nodes during provisioning

    • default_internal, created on the vx2 VxLAN interfaces of all compute nodes during provisioning

    Also, you can create additional VLAN interfaces and VLANs with the Virtual Machine role. For virtual machines requiring public connectivity, use the compute nodes' vx13040 VxLAN interfaces. For internal-only VM traffic, use the vx2 VxLAN interfaces. For details, see Configuring Network Resources for Virtual Machines.

    Note:

    Do not create virtual machine networks using the ethx ports. These are detected in Oracle VM Manager as physical compute node network interfaces, but they are not cabled. Also, the bondx ports and default VLAN interfaces (tun-ext, tun-int, mgmt-int and storage-int) that appear in Oracle VM Manager are part of the appliance infrastructure networking, and are not intended to be used in VM network configurations.

    Virtual machine networking can be further diversified and segregated by means of custom networks, which are described below. Custom networks must be created in the Private Cloud Appliance CLI. This generates additional VxLAN interfaces equivalent to the default vx13040 and vx2. The custom networks and associated network interfaces are automatically set up in Oracle VM Manager, where you can expand the virtual machine network configuration with those newly discovered network resources.

  • Custom Network

    Custom networks are infrastructure networks you create in addition to the default configuration. These are constructed in the same way as the default private and public networks, but using different compute node network interfaces and terminating on different spine switch ports. Whenever public connectivity is required, additional cabling between the spine switches and the next-level data center switches is required.

    Because they are part of the infrastructure underlying Oracle VM, all custom networks must be configured through the Private Cloud Appliance CLI. The administrator chooses between three types: private, public or host network. For detailed information about the purpose and configuration of each type, see Network Customization.

    If your environment has additional tenant groups, which are separate Oracle VM server pools, then a custom network can be associated with one or more tenant groups. This allows you to securely separate traffic belonging to different tenant groups and the virtual machines deployed as part of them. For details, see Tenant Groups.

    Once custom networks have been fully configured through the Private Cloud Appliance CLI, the networks and associated ports automatically appear in Oracle VM Manager. There, additional VLAN interfaces can be configured on top of the new VxLAN interfaces, and then used to create more VLANs for virtual machine connectivity. The host network is a special type of custom public network, which can assume the Storage network role and can be used to connect external storage directly to compute nodes.

  • Network Properties

    The network role is a property used within Oracle VM. Most of the networks you configure, have the Virtual Machine role, although you could decide to use a separate network for storage connectivity or virtual machine migration. Network roles – and other properties such as name and description, which interfaces are connected, properties of the interfaces and so on – can be configured in Oracle VM Manager, as long as they do not conflict with properties defined at the appliance level.

    Modifying network properties of the VM networks you configured in Oracle VM Manager involves little risk. However, you must not change the configuration – such as network roles, ports and so on – of the default networks: eth_management, mgmt_internal, storage_internal, underlay_external, underlay_internal, default_external, and default_internal. For networks connecting compute nodes, including custom networks, you must use the Private Cloud Appliance CLI. Furthermore, you cannot modify the functional properties of a custom network: you have to delete it and create a new one with the required properties.

    The maximum transfer unit (MTU) of a network interface, standard port or bond, cannot be modified. It is determined by the hardware properties or the SDN configuration, which cannot be controlled from within Oracle VM Manager.

  • VLAN Management

    With the exception of the underlay VLAN networks configured through SDN, and the appliance management VLAN you configure in the Network Settings tab of the Oracle Private Cloud Appliance Dashboard, all VLAN configuration and management operations are performed in Oracle VM Manager. These VLANs are part of the VM networking.

    Tip:

    When a large number of VLANs are required, it is good practice not to generate them all at once, because the process is time-consuming. Instead, add (or remove) VLANs in groups of 10.

Network Customization

The Oracle Private Cloud Appliance controller software allows you to add custom networks at the appliance level. This means that certain hardware components require configuration changes to enable the additional connectivity. The new networks are then configured automatically in your Oracle VM environment, where they can be used for isolating and optimizing network traffic beyond the capabilities of the default network configuration. All custom networks, both internal and public, are VLAN-capable.

The virtual machines hosted on the Private Cloud Appliance have access to external compute resources and storage, through the default external facing networks, as soon as the Oracle VM Manager is accessible.

If you need additional network functionality, custom networks can be configured for virtual machines and compute nodes. For example, a custom network can provide virtual machines with additional bandwidth or additional access to external compute resources or storage. Or you can use a custom network if compute nodes need to access storage repositories and data disks contained on external storage. The sections below describe how to configure and cable your Private Cloud Appliance for these custom networks.

Attention:

Do not modify the network configuration while upgrade operations are running. No management operations are supported during upgrade, as these may lead to configuration inconsistencies and significant repair downtime.

Attention:

Custom networks must never be deleted in Oracle VM Manager. Doing so would leave the environment in an error state that is extremely difficult to repair. To avoid downtime and data loss, always perform custom network operations in the Private Cloud Appliance CLI.

Caution:

The following network limitations apply:

  • The maximum number of custom external networks is 7 per tenant group or per compute node.

  • The maximum number of custom internal networks is 3 per tenant group or per compute node.

  • The maximum number of VLANs is 256 per tenant group or per compute node.

  • Only one host network can be assigned per tenant group or per compute node.

Caution:

When configuring custom networks, make sure that no provisioning operations or virtual machine environment modifications take place. This might lock Oracle VM resources and cause your Private Cloud Appliance CLI commands to fail.

Creating custom networks requires use of the CLI. The administrator chooses between three types: a network internal to the appliance, a network with external connectivity, or a host network. Custom networks appear automatically in Oracle VM Manager. The internal and external networks take the virtual machine network role, while a host network may have the virtual machine and storage network roles.

The host network is a particular type of external network: its configuration contains additional parameters for subnet and routing. The servers connected to it also receive an IP address in that subnet, and consequently can connect to an external network device. The host network is particularly useful for direct access to storage devices.

Configuring Custom Networks

For all networks with external connectivity, the spine Cisco Nexus 9336C-FX2 Switch ports must be specified so that these are reconfigured to route the external traffic. These ports must be cabled to create the physical uplink to the next-level switches in the data center. For detailed information, refer to "Appliance Uplink Configuration" in Network Requirements in the Oracle Private Cloud Appliance Installation Guide.

Creating a Custom Network

  1. Using SSH and an account with superuser privileges, log into the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Launch the Private Cloud Appliance command line interface.

    # pca-admin
    Welcome to PCA! Release: 2.4.4
    PCA>
  3. If your custom network requires public connectivity, you need to use one or more spine switch ports. Verify the number of ports available and carefully plan your network customizations accordingly. The following example shows how to retrieve that information from your system:

    PCA> list network-port
    
    Port      Switch          Type                   State           Networks
    ----      ------          ----                   -----           --------
    1:1       ovcasw22r1      10G                    down            None
    1:2       ovcasw22r1      10G                    down            None
    1:3       ovcasw22r1      10G                    down            None
    1:4       ovcasw22r1      10G                    down            None
    2         ovcasw22r1      40G                    up              None
    3         ovcasw22r1      auto-speed             down            None
    4         ovcasw22r1      auto-speed             down            None
    5:1       ovcasw22r1      10G                    up              default_external
    5:2       ovcasw22r1      10G                    down            default_external
    5:3       ovcasw22r1      10G                    down            None
    5:4       ovcasw22r1      10G                    down            None
    1:1       ovcasw23r1      10G                    down            None
    1:2       ovcasw23r1      10G                    down            None
    1:3       ovcasw23r1      10G                    down            None
    1:4       ovcasw23r1      10G                    down            None
    2         ovcasw23r1      40G                    up              None
    3         ovcasw23r1      auto-speed             down            None
    4         ovcasw23r1      auto-speed             down            None
    5:1       ovcasw23r1      10G                    up              default_external
    5:2       ovcasw23r1      10G                    down            default_external
    5:3       ovcasw23r1      10G                    down            None
    5:4       ovcasw23r1      10G                    down            None
    -----------------
    22 rows displayed
    
    Status: Success
  4. For a custom network with external connectivity, configure an uplink port group with the uplink ports you wish to use for this traffic. Select the appropriate breakout mode.

    PCA> create uplink-port-group MyUplinkPortGroup '1:1 1:2' 10g-4x
    Status: Success

    Note:

    The port arguments are specified as 'x:y' where x is the switch port number and y is the number of the breakout port, in case a splitter cable is attached to the switch port. The example above shows how to retrieve that information.

    You must set the breakout mode of the uplink port group. When a 4-way breakout cable is used, all four ports must be set to either 10Gbit or 25Gbit. When no breakout cable is used, the port speed for the uplink port group should be either 100Gbit or 40Gbit, depending on connectivity requirements. See create uplink-port-group for command details.

    Network ports cannot be part of more than one network configuration.

  5. Create a new network and select one of these types:

    • rack_internal_network: an Oracle VM virtual machine network with no access to external networking; no IP addresses are assigned to compute nodes. Use this option to allow virtual machines additional bandwidth beyond the default internal network.

    • external_network: an Oracle VM virtual machine network with access to external networking; no IP addresses are assigned to compute nodes. Use this option to allow virtual machines additional bandwidth when accessing external storage on a physical network separate from the default external facing network.

    • host_network: an Oracle VM compute node network with access to external networking; IP addresses are added to compute nodes. Use this option to allow compute nodes to access storage and compute resources external to the Private Cloud Appliance. This can also be used by virtual machines to access external compute resources just like external_network.

    Use the following syntax:

    • For an internal-only network, specify a network name.

      PCA> create network MyInternalNetwork rack_internal_network
      Status: Success
    • For an external network, specify a network name and the spine switch port group to be configured for external traffic.

      PCA> create network MyPublicNetwork external_network MyUplinkPortGroup
      Status: Success
    • For a host network, specify a network name, the spine switch ports to be configured for external traffic, the subnet, and optionally the routing configuration.

      PCA> create network MyHostNetwork host_network MyUplinkPortGroup \
      10.10.10 255.255.255.0 10.1.20.0/24 10.10.10.250
      Status: Success

      Note:

      In this example the additional network and routing arguments for the host network are specified as follows, separated by spaces:

      • 10.10.10 = subnet prefix

      • 255.255.255.0 = netmask

      • 10.1.20.0/24 = route destination (as subnet or IPv4 address)

      • 10.10.10.250 = route gateway

      The subnet prefix and netmask are used to assign IP addresses to servers joining the network. The optional route gateway and destination parameters are used to configure a static route in the server's routing table. The route destination is a single IP address by default, so you must specify a netmask if traffic could be intended for different IP addresses in a subnet.

      When you define a host network, it is possible to enter invalid or contradictory values for the Prefix, Netmask and Route_Destination parameters. For example, when you enter a prefix with "0" as the first octet, the system attempts to configure IP addresses on compute node Ethernet interfaces starting with 0. Also, when the netmask part of the route destination you enter is invalid, the network is still created, even though an exception occurs. When such a poorly configured network is in an invalid state, it cannot be reconfigured or deleted with standard commands. If an invalid network configuration is applied, use the --force option to delete the network.

      Details of the create network command arguments are provided in create network in the CLI reference chapter.

      Caution:

      Network and routing parameters of a host network cannot be modified. To change these settings, delete the custom network and re-create it with updated settings.

  6. Connect the required servers to the new custom network. You must provide the network name and the names of the servers to connect.

    PCA> add network MyPublicNetwork ovcacn07r1
    Status: Success
    PCA> add network MyPublicNetwork ovcacn08r1
    Status: Success
    PCA> add network MyPublicNetwork ovcacn09r1
    Status: Success
  7. Verify the configuration of the new custom network.

    PCA> show network MyPublicNetwork
    
    ----------------------------------------
    Network_Name         MyPublicNetwork
    Trunkmode            None
    Description          None
    Ports                ['1:1', '1:2']
    vNICs                None
    Status               ready
    Network_Type         external_network
    Compute_Nodes        ovcacn07r1, ovcacn08r1, ovcacn09r1
    Prefix               None
    Netmask              None
    Route Destination    None
    Route Gateway        None
    ----------------------------------------
    
    Status: Success

    As a result of these commands, a VxLAN interface is configured on each of the servers to connect them to the new custom network. These configuration changes are reflected in the Networking tab and the Servers and VMs tab in Oracle VM Manager.

    Note:

    If the custom network is a host network, the server is assigned an IP address based on the prefix and netmask parameters of the network configuration, and the final octet of the server's internal management IP address.

    For example, if the compute node with internal IP address 192.168.4.9 were connected to the host network used for illustration purposes in this procedure, it would receive the address 10.10.10.9 in the host network.

    Figure 2-7 shows a custom network named MyPublicNetwork, which is VLAN-capable and uses the compute node's vx13041 interface.

    Figure 2-7 Oracle VM Manager View of Custom Network Configuration


    Screenshot showing the Servers and VMs tab of the Oracle Private Cloud Appliance Dashboard. Details are shown of the network configuration on one of the compute nodes that was added to the new custom network.
  8. To disconnect servers from the custom network use the remove network command.

    Attention:

    Before removing the network connection of a server, make sure that no virtual machines rely on this network.

    When a server is no longer connected to a custom network, make sure that its port configuration is cleaned up in Oracle VM.

    PCA> remove network MyPublicNetwork ovcacn09r1
    ************************************************************
     WARNING !!! THIS IS A DESTRUCTIVE OPERATION.
    ************************************************************
    Are you sure [y/N]:y
    
    Status: Success

Deleting Custom Networks

This section describes how to delete custom networks.

Deleting a Custom Network

Caution:

Before deleting a custom network, make sure that all servers have been disconnected from it.

  1. Using SSH and an account with superuser privileges, log into the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Launch the Oracle Private Cloud Appliance command line interface.

    # pca-admin
    Welcome to PCA! Release: 2.4.4
    PCA>
  3. Verify that all servers have been disconnected from the custom network. No vNICs or nodes should appear in the network configuration.

    Caution:

    Related configuration changes in Oracle VM must be cleaned up as well.

    PCA> show network MyPublicNetwork
    
    ----------------------------------------
    Network_Name         MyPublicNetwork
    Trunkmode            None
    Description          None
    Ports                ['1:1', '1:2']
    vNICs                None
    Status               ready
    Network_Type         external_network
    Compute_Nodes        None
    Prefix               None
    Netmask              None
    Route_Destination    None
    Route_Gateway        None
    ----------------------------------------
  4. Delete the custom network.

    PCA> delete network MyPublicNetwork
    ************************************************************
     WARNING !!! THIS IS A DESTRUCTIVE OPERATION.
    ************************************************************
    Are you sure [y/N]:y
    
    Status: Success

    Caution:

    If a custom network is left in an invalid or error state, and the delete command fails, you may use the --force option and retry.

VM Storage Networks

Starting with Oracle Private Cloud Appliance Controller Software release 2.4.3, you can configure private storage networks that grant users access to the internal ZFS Storage Appliance from their Oracle VM environment. Private Cloud Appliance administrators with root access to the management nodes can create and manage the required networks and ZFS shares (iSCSI/NFS) using the pca-admin command line interface. To ensure you can use this functionality, upgrade the storage network as described in Upgrading the Storage Network.

Private Cloud Appliance administrators can create up to sixteen VM storage networks which can be accessed by any virtual machine in any tenant group. End users of virtual machines configure their guest operating system to access one or more of these internal storage networks through NFS or iSCSI once the Private Cloud Appliance administrator has completed the setup.

The VM storage networks are designed to isolate different business systems or groups of end users from each other. For example, the HR department can use two VM storage networks for their virtual machines, while the payroll department can have three or four VM storage networks of their own. Each VM storage network is assigned a single, private non-routed VxLAN to ensure the network is isolated from other virtual machines owned by different end users. End users cannot gain root access to mange the internal ZFS Storage Appliance through the VM storage networks.

The ability to define internal storage networks directly for VMs was introduced in Oracle Private Cloud Appliance Controller Software release 2.4.3. Refer to Oracle Support Document 2722899.1 for important details before using this feature. Should you have any questions, contact Oracle support.

Tenant Groups

A standard Oracle Private Cloud Appliance environment built on a full rack configuration contains 25 compute nodes. A tenant group is a logical subset of a single Private Cloud Appliance environment. Tenant groups provide an optional mechanism for a Private Cloud Appliance administrator to subdivide the environment in arbitrary ways for manageability and isolation. The tenant group offers a means to isolate compute, network and storage resources per end customer. It also offers isolation from cluster faults.

Design Assumptions and Restrictions

Oracle Private Cloud Appliance supports a maximum of 8 tenant groups. This number includes the default tenant group, which cannot be deleted from the environment, and must always contain at least one compute node. Therefore, a single custom tenant group can contain up to 24 compute nodes, while the default Rack1_ServerPool can contain all 25.

Regardless of tenant group membership, all compute nodes are connected to all of the default Private Cloud Appliance networks. Custom networks can be assigned to multiple tenant groups. When a compute node joins a tenant group, it is also connected to the custom networks associated with the tenant group. When you remove a compute node from a tenant group, it is disconnected from those custom networks. A synchronization mechanism, built into the tenant group functionality, keeps compute node network connections up to date when tenant group configurations change.

When you reprovision compute nodes, they are automatically removed from their tenant groups, and treated as new servers. Consequently, when a compute node is reprovisioned, or when a new compute node is added to the environment, it is added automatically to Rack1_ServerPool. After successful provisioning, you can add the compute node to the appropriate tenant group.

When you create a new tenant group, the system does not create a storage repository for the new tenant group. An administrator must configure the necessary storage resources for virtual machines in Oracle VM Manager. See Viewing and Managing Storage Resources.

Configuring Tenant Groups

The tenant group functionality can be accessed through the Oracle Private Cloud Appliance CLI. With a specific set of commands, you manage the tenant groups, their member compute nodes, and the associated custom networks. The CLI initiates a number of Oracle VMoperations to set up the server pool, and a synchronization service maintains settings across the members of the tenant group.

Attention:

Do not modify the tenant group configuration while upgrade operations are running. No management operations are supported during upgrade, as these may lead to configuration inconsistencies and significant repair downtime.

Caution:

You must not modify the server pool in Oracle VM Manager because this causes inconsistencies in the tenant group configuration and disrupts the operation of the synchronization service and the Private Cloud Appliance CLI. Only server pool policies may be edited in Oracle VM Manager.

If you inadvertently used Oracle VM Manager to modify a tenant group, see Recovering from Tenant Group Configuration Mismatches.

Note:

For detailed information about the Private Cloud Appliance CLI tenant group commands, see Oracle Private Cloud Appliance Command Line Interface (CLI).

Creating and Populating a Tenant Group

  1. Using SSH and an account with superuser privileges, log into the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Launch the Private Cloud Appliance command line interface.

    # pca-admin
    Welcome to PCA! Release: 2.4.4
    PCA>
  3. Create the new tenant group.

    PCA> create tenant-group myTenantGroup
    Status: Success
    
    PCA> show tenant-group myTenantGroup
    
    ----------------------------------------
    Name                 myTenantGroup
    Default              False
    Tenant_Group_ID      0004fb00000200008154bf592c8ac33b
    Servers              None
    State                ready
    Tenant_Group_VIP     None
    Tenant_Networks      ['storage_internal', 'mgmt_internal', 'underlay_internal', 'underlay_external', 
                         'default_external', 'default_internal']
    Pool_Filesystem_ID   3600144f0d04414f400005cf529410003
    ----------------------------------------
    
    Status: Success

    The new tenant group appears in Oracle VM Manager as a new server pool. It has a 12GB server pool file system located on the internal ZFS Storage Appliance.

  4. Add compute nodes to the tenant group.

    If a compute node is currently part of another tenant group, it is first removed from that tenant group.

    Caution:

    If the compute node is hosting virtual machines, or if storage repositories are presented to the compute node or its current tenant group, removing a compute node from an existing tenant group will fail. If so, you have to migrate the virtual machines and unpresent the repositories before adding the compute node to a new tenant group.

    PCA> add compute-node ovcacn07r1 myTenantGroup
    Status: Success
    
    PCA> add compute-node ovcacn09r1 myTenantGroup
    Status: Success
  5. Add a custom network to the tenant group.

    PCA> add network-to-tenant-group myPublicNetwork myTenantGroup
    Status: Success

    Custom networks can be added to the tenant group as a whole. This command creates synchronization tasks to configure custom networks on each server in the tenant group.

    Caution:

    While synchronization tasks are running, make sure that no reboot or provisioning operations are started on any of the compute nodes involved in the configuration changes.

  6. Verify the configuration of the new tenant group.

    PCA> show tenant-group myTenantGroup
    
    ----------------------------------------
    Name                 myTenantGroup
    Default              False
    Tenant_Group_ID      0004fb00000200008154bf592c8ac33b
    Servers              ['ovcacn07r1', 'ovcacn09r1']
    State                ready
    Tenant_Group_VIP     None
    Tenant_Networks      ['storage_internal', 'mgmt_internal', 'underlay_internal', 'underlay_external', 
                         'default_external', 'default_internal', 'myPublicNetwork']
    Pool_Filesystem_ID   3600144f0d04414f400005cf529410003
    ----------------------------------------
    
    Status: Success

    The new tenant group corresponds with an Oracle VM server pool with the same name and has a pool file system. The command output also shows that the servers and custom network were added successfully.

These configuration changes are reflected in the Servers and VMs tab in Oracle VM Manager. Figure 2-8 shows a second server pool named MyTenantGroup, which contains the two compute nodes that were added as examples in the course of this procedure.

Note:

The system does not create a storage repository for a new tenant group. An administrator must configure the necessary storage resources for virtual machines in Oracle VM Manager. See Viewing and Managing Storage Resources.

Figure 2-8 Oracle VM Manager View of New Tenant Group


Screenshot showing the Servers and VMs tab of the Oracle Private Cloud Appliance Dashboard. The newly created tenant group appears in the Server Pools list and contains the two servers that were added as part of the procedure.

Reconfiguring and Deleting a Tenant Group

  1. Identify the tenant group you intend to modify.

    PCA> list tenant-group
    
    Name                 Default      State
    ----                 -------      -----
    Rack1_ServerPool     True         ready
    myTenantGroup        False        ready
    ----------------
    2 rows displayed
    
    Status: Success
    
    PCA> show tenant-group myTenantGroup
    
    ----------------------------------------
    Name                 myTenantGroup
    Default              False
    Tenant_Group_ID      0004fb00000200008154bf592c8ac33b
    Servers              ['ovcacn07r1', 'ovcacn09r1']
    State                ready
    Tenant_Group_VIP     None
    Tenant_Networks      ['storage_internal', 'mgmt_internal', 'underlay_internal', 'underlay_external', 
                         'default_external', 'default_internal', 'myPublicNetwork']
    Pool_Filesystem_ID   3600144f0d04414f400005cf529410003
    ----------------------------------------
    
    Status: Success
  2. Remove a network from the tenant group.

    A custom network that has been associated with a tenant group can be removed again. The command results in serial operations, not using the synchronization service, to unconfigure the custom network on each compute node in the tenant group.

    PCA> remove network-from-tenant-group myPublicNetwork myTenantGroup
    ************************************************************
     WARNING !!! THIS IS A DESTRUCTIVE OPERATION.
    ************************************************************
    Are you sure [y/N]:y
    
    Status: Success
  3. Remove a compute node from the tenant group.

    Use Oracle VM Manager to prepare the compute node for removal from the tenant group. Make sure that virtual machines have been migrated away from the compute node, and that no storage repositories are presented.

    PCA> remove compute-node ovcacn09r1 myTenantGroup
    ************************************************************
     WARNING !!! THIS IS A DESTRUCTIVE OPERATION.
    ************************************************************
    Are you sure [y/N]:y
    
    Status: Success

    When you remove a compute node from a tenant group, any custom network associated with the tenant group is automatically removed from the compute node network configuration. Custom networks that are not associated with the tenant group are not removed.

  4. Delete the tenant group.

    Before attempting to delete a tenant group, make sure that all compute nodes have been removed.

    Before removing the last remaining compute node from the tenant group, use Oracle VM Manager to unpresent any shared repository from the compute node, and then release ownership of it. For details, see support document Remove Last Compute Node from Tenant Group Fails with "There are still OCFS2 file systems" (Doc ID 2653515.1).

    PCA> delete tenant-group myTenantGroup
    ************************************************************
     WARNING !!! THIS IS A DESTRUCTIVE OPERATION.
    ************************************************************
    Are you sure [y/N]:y
    
    Status: Success

    When the tenant group is deleted, operations are launched to remove the server pool file system LUN from the internal ZFS storage appliance. The tenant group's associated custom networks are not destroyed.

Authentication

The Password Management window is used to reset the global Oracle Private Cloud Appliance password and to set unique passwords for individual components within the appliance. All actions performed via this tab require that you enter the current password for the Private Cloud Appliance admin user in the field labeled Current PCA Admin Password. Fields are available to specify the new password value and to confirm the value:

  • Current PCA Admin Password: You must provide the current password for the Private Cloud Appliance admin user before any password changes can be applied.

  • New Password: Provide the value for the new password that you are setting.

  • Verify Password: Confirm the new password and check that you have not mistyped what you intended.

The window provides a series of check boxes that make it easy to select the level of granularity that you wish to apply to a password change. At this time, do not use the Select All button to apply a global password to all components that are used in the appliance. For more information see Changing Multiple Component Passwords Causes Authentication Failure in Oracle VM Manager. This action resets any individual passwords that you may have set for particular components. For stricter controls, you may set the password for individual components by simply selecting the check box associated with each component that you wish to apply a password to.

Caution:

Password changes are not instantaneous across the appliance, but are propagated through a task queue. When applying a password change, allow at least 30 minutes for the change to take effect. Do not attempt any further password changes during this delay. Verify that the password change has been applied correctly.

  • Select All: Apply the new password to all components. All components in the list are selected.

  • Oracle VM Manager/PCA admin password: Set the new password for the Oracle VM Manager and Oracle Private Cloud Appliance Dashboard admin user.

  • Oracle MySQL password: Set the new password for the ovs user in MySQL used by Oracle VM Manager.

  • Oracle WebLogic Server password: Set the new password for the weblogic user in WebLogic Server.

  • Oracle Data Network Leaf Switch admin password: Set the new password for the admin user for the leaf Cisco Nexus 9336C-FX2 Switches.

  • Oracle Management Network Switch admin password: Set the new password for the admin user for the Cisco Nexus 9348GC-FXP Switch.

  • Oracle Data Network Spine Switch admin password: Set the new password for the admin user for the spine Cisco Nexus 9336C-FX2 Switches.

  • Oracle ZFS Storage root password: Set the new password for the root user for the ZFS Storage Appliance.

  • PCA Management Node root password: Set the new password for the root user for both management nodes.

  • PCA Compute Node root password: Set the new password for the root user for all compute nodes.

  • PCA Management Node SP/ILOM root password: Set the new password for the root user for the ILOM on both management nodes.

  • PCA Compute Node SP/ILOM root password: Set the new password for the root user for the ILOM on all compute nodes.

Figure 2-9 Password Management


Screenshot showing the Password Management window of the Oracle Private Cloud Appliance Dashboard.

The functionality that is available in the Oracle Private Cloud Appliance Dashboard is equally available via the Oracle Private Cloud Appliance CLI as described in update password.

Caution:

Passwords of components must not be changed manually as this will cause mismatches with the authentication details stored in the Oracle Private Cloud Appliance Wallet.

Health Monitoring

The Oracle Private Cloud Appliance Controller Software contains a monitoring service, which is started and stopped with the ovca service on the active management node. When the system runs for the first time, it creates an inventory database and monitor database. Once these are set up and the monitoring service is active, health information about the hardware components is updated continuously.

The inventory database is populated with information about the various components installed in the rack, including the IP addresses to be used for monitoring. With this information, the ping manager pings all known components every 3 minutes and updates the inventory database to indicate whether a component is pingable and when it was last seen online. When errors occur, they are logged in the monitor database. Error information is retrieved from the component ILOMs.

For troubleshooting purposes, historic health status details can be retrieved through the CLI support mode by an authorized Oracle Field Engineer. When the CLI is used in support mode, a number of additional commands are available, two of which are used to display the contents of the health monitoring databases.

  • Use show db inventory to display component health status information from the inventory database.

  • Use show db monitor to display errors logged in the monitoring database.

The appliance administrator can retrieve current component health status information from the Oracle Linux command line on the active management node by using the Oracle Private Cloud Appliance Health Check utility. The Health Check utility is built on the framework of the Oracle Private Cloud Appliance Upgrader, and is included in the Upgrader package. It detects the appliance network architecture and runs the sets of health checks defined for the system in question.

Checking the Current Health Status of an Oracle Private Cloud Appliance Installation

  1. Using SSH and an account with superuser privileges, log in to the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Launch the Health Check utility.

    # pca_healthcheck
    PCA Rack Type: PCA X8_BASE.
    Please refer to log file
    /nfs/shared_storage/pca_upgrader/log/pca_healthcheck_2019_10_04-12.09.45.log
    for more details.

    After detecting the rack type, the utility executes the applicable health checks.

    Beginning PCA Health Checks...
    
    Check Management Nodes Are Running                                     1/24
    Check Support Packages                                                 2/24
    Check PCA DBs Exist                                                    3/24
    PCA Config File                                                        4/24
    Check Shares Mounted on Management Nodes                               5/24
    Check PCA Version                                                      6/24
    Check Installed Packages                                               7/24
    Check for OpenSSL CVE-2014-0160 - Security Update                      8/24
    Management Nodes Have IPv6 Disabled                                    9/24
    Check Oracle VM Manager Version                                       10/24
    Oracle VM Manager Default Networks                                    11/24
    Repositories Defined in Oracle VM Manager                             12/24
    PCA Services                                                          13/24
    Oracle VM Server Model                                                14/24
    Network Interfaces on Compute Nodes                                   15/24
    Oracle VM Manager Settings                                            16/24
    Check Network Leaf Switch                                             17/24
    Check Network Spine Switch                                            18/24
    All Compute Nodes Running                                             19/24
    Test for ovs-agent Service on Compute Nodes                           20/24
    Test for Shares Mounted on Compute Nodes                              21/24
    Check for bash ELSA-2014-1306 - Security Update                       22/24
    Check Compute Node's Active Network Interfaces                        23/24
    Checking for xen OVMSA-2014-0026 - Security Update                    24/24
    
    PCA Health Checks completed after 2 minutes
  3. When the health checks have been completed, check the report for failures.

    Check Management Nodes Are Running                                   Passed
    Check Support Packages                                               Passed
    Check PCA DBs Exist                                                  Passed
    PCA Config File                                                      Passed
    Check Shares Mounted on Management Nodes                             Passed
    Check PCA Version                                                    Passed
    Check Installed Packages                                             Passed
    Check for OpenSSL CVE-2014-0160 - Security Update                    Passed
    Management Nodes Have IPv6 Disabled                                  Passed
    Check Oracle VM Manager Version                                      Passed
    Oracle VM Manager Default Networks                                   Passed
    Repositories Defined in Oracle VM Manager                            Passed
    PCA Services                                                         Passed
    Oracle VM Server Model                                               Passed
    Network Interfaces on Compute Nodes                                  Passed
    Oracle VM Manager Settings                                           Passed
    Check Network Leaf Switch                                            Passed
    Check Network Spine Switch                                           Failed
    All Compute Nodes Running                                            Passed
    Test for ovs-agent Service on Compute Nodes                          Passed
    Test for Shares Mounted on Compute Nodes                             Passed
    Check for bash ELSA-2014-1306 - Security Update                      Passed
    Check Compute Node's Active Network Interfaces                       Passed
    Checking for xen OVMSA-2014-0026 - Security Update                   Passed
    
    ---------------------------------------------------------------------------
    Overall Status                                                       Failed
    ---------------------------------------------------------------------------
    
    Please refer to log file
    /nfs/shared_storage/pca_upgrader/log/pca_healthcheck_2019_10_04-12.09.45.log
    for more details.
  4. If certain checks have resulted in failures, review the log file for additional diagnostic information. Search for text strings such as "error" and "failed".

    # grep -inr "failed" /nfs/shared_storage/pca_upgrader/log/pca_healthcheck_2019_10_04-12.09.45.log
    
    726:[2019-10-04 12:10:51 264234] INFO (healthcheck:254) Check Network Spine Switch Failed -
    731:  Spine Switch ovcasw22r1 North-South Management Network Port-channel check                 [FAILED]
    733:  Spine Switch ovcasw22r1 Multicast Route Check                                             [FAILED]
    742:  Spine Switch ovcasw23r1 North-South Management Network Port-channel check                 [FAILED]
    750:[2019-10-04 12:10:51 264234] ERROR (precheck:148) [Check Network Spine Switch ()] Failed
    955:[2019-10-04 12:12:26 264234] INFO (precheck:116) [Check Network Spine Switch ()] Failed
    
    # less /nfs/shared_storage/pca_upgrader/log/pca_healthcheck_2019_10_04-12.09.45.log
    
    [...]
      Spine Switch ovcasw22r1 North-South Management Network Port-channel check                 [FAILED]
      Spine Switch ovcasw22r1 OSPF Neighbor Check                                               [OK]
      Spine Switch ovcasw22r1 Multicast Route Check                                             [FAILED]
      Spine Switch ovcasw22r1 PIM RP Check                                                      [OK]
      Spine Switch ovcasw22r1 NVE Peer Check                                                    [OK]
      Spine Switch ovcasw22r1 Spine Filesystem Check                                            [OK]
      Spine Switch ovcasw22r1 Hardware Diagnostic Check                                         [OK]
    [...]
  5. Investigate and fix any detected problems. Repeat the health check until the system passes all checks.

Fault Monitoring

Beginning with Oracle Private Cloud Appliance Controller software release 2.4.3, the health checker is a service started by the ovca-daemon on the active management node. Checks can be run manually from the command line, or by using definitions in the scheduler. Depending on the check definition, the Private Cloud Appliance health checker, the Oracle VM health check, and the Private Cloud Appliance pre-upgrade checks can be invoked.

  • pca_healthcheck monitors the health of system hardware components. For details, see Health Monitoring.

  • ovm_monitor monitors the Oracle VM Manager objects and other environment factors.

  • pca_upgrader monitors the system during an upgrade.

Health checking can be integrated with the ZFS Phone Home service to send reports to Oracle on a weekly basis. The Phone Home function must be activated by the customer and requires that the appliance is registered with Oracle Auto Service Request (ASR). No separate installation is required: All functions come with the controller software in Oracle Private Cloud Appliance Controller software release 2.4.3 and later. For configuration information, see Phone Home Service. For more information about ASR, see Oracle Auto Service Request (ASR).

Using Fault Monitoring Checks

The appliance administrator can access current component health status information from the Oracle Linux command line on the active management node by using the Oracle Private Cloud Appliance Fault Monitoring utility. The Fault Monitoring utility is included in the ovca services. In addition, you can schedule checks to run automatically. The Fault Monitoring utility detects the appliance network architecture and runs the sets of health checks that are defined for that system.

Running Fault Monitor Tests Manually

The Fault Monitoring utility enables you to run an individual check, all the checks for a particular monitoring service, or all of the checks that are available.

  1. Using SSH and an account with superuser privileges, log in to the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. List the available checks.

    [root@ovcamn05r1 ~]# pca-faultmonitor --help
    usage: pca-faultmonitor [-h] [--list_all_monitors][--list_ovm_monitors]
                            [--list_pca_healthcheck_monitors]
                            [--list_pca_upgrader_monitors]
                            [--run_all_monitors]
                            [--run_ovm_monitors]
                            [--run_pca_healthcheck_monitors]
                            [--run_pca_upgrader_monitors][-m MONITOR_LIST]
                            [--print_report]
    
    optional arguments:  
      -h, --help show this help message and exit
      --list_all_monitors List all Fault Monitors(Oracle VM, pca_healthcheck and pca_upgrader)
      --list_ovm_monitors List Oracle VM Fault Monitors
      --list_pca_healthcheck_monitors List pca_healthcheck Fault Monitors
      --list_pca_upgrader_monitors List pca_upgrader Fault Monitors
      --run_all_monitors Run all Fault Monitors
      --run_ovm_monitors Run Oracle VM Fault Monitors
      --run_pca_healthcheck_monitors Run pca_healthcheck Fault Monitors
      --run_pca_upgrader_monitors Run pca_upgrader Fault Monitors
      -m MONITOR_LIST Runs a list of Fault Monitors. Each Fault Monitor must
         be specified with -m option
      --print_report Prints the report on console
    None
    PCA Rack type:      hardware_orange
    Please refer the log file in /var/log/ovca-faultmonitor.log
    Please look at fault report in /nfs/shared_storage/faultmonitor/20200512/
    Note: Reports will not be created for success status

    List all monitors of a specified check.

    [root@ovcamn05r1 faultmonitor]# pca-faultmonitor --list_pca_upgrader_monitors
    PCA Rack type:      hardware_orange
    Please refer the log file in /var/log/faultmonitor/ovca-faultmonitor.log
    Please look at fault report in /nfs/shared_storage/faultmonitor/20200221/
    Note: Reports will not be created for success status
    
    Listing all PCA upgrader faultmonitors
    
    check_ib_symbol_errors                    verify_inventory_cns
    validate_image                            check_available_space
    check_max_paths_iscsi                     check_serverUpdateConfiguration
    check_onf_error                           verify_password
    check_rpm_db                              verify_network_config
    check_yum_proxy                           check_motd
    check_yum_repo                            connect_mysql
    check_osa_disabled                        check_xsigo_configs
    check_pca_services                        check_mysql_desync_passwords
    check_storage_space                       verify_xms_cards
    ...
  3. Run the desired checks.

    • Run all checks.

      [root@ovcamn05r1 ~]# pca-faultmonitor --run_all_monitors
    • To run a specific check or list of checks, list one or more checks, each as an argument of a separate -m option.

      [root@ovcamn05r1 ~]# pca-faultmonitor -m event_monitor -m check_storage_space 
    • Run checks for a specific monitor.

      [root@ovcamn05r1 ~]# pca-faultmonitor --run_pca_upgrader_monitors
      [root@ovcamn05r1 faultmonitor]# pca-faultmonitor --run_ovm_monitors
      PCA Rack type:      hardware_orange
      Please refer the log file in /var/log/faultmonitor/ovca-faultmonitor.log
      Please look at fault report in /nfs/shared_storage/faultmonitor/20200220/
      Note: Reports will not be created for success status
      
      Beginning OVM Fault monitor checks ...
      
      event_monitor                        1/13
      repository_utilization_monitor       2/13
      storage_utilization_monitor          3/13
      db_size_monitor                      4/13
      onf_monitor                          5/13
      db_backup_monitor                    6/13
      firewall_monitor                     7/13
      server_connectivity_monitor          8/13
      network_monitor                      9/13
      port_flapping_monitor                10/13
      storage_path_flapping_monitor        11/13
      repository_mount_monitor             12/13
      server_pool_monitor                  13/13
      --------------------------------------------------
      Fault Monitor Report Summary
      --------------------------------------------------
      OVM_Event_Monitor                    Success
      OVM_Repository_Utilization_Monitor   Success
      OVM_Storage_Utilization_Monitor      Success
      DB_Size_Monitor                      Success
      ONF_Monitor                          Success
      DB_Backup_Monitor                    Success
      Firewall_Monitor                     Success
      Server_Connectivity_Monitor          Success
      Network_Monitor                      Warning
      Port_Flapping_Monitor                Success
      Storage_Path_Flapping_Monitor        Success
      Repository_Mount_Monitor             Warning
      Server_Pool_Monitor                  Success
      --------------------------------------------------
      Overall                                    Failure
      --------------------------------------------------
      
      PCA Rack type:      hardware_orange
      Please refer the log file in /var/log/faultmonitor/ovca-faultmonitor.log
      Please look at fault report in /nfs/shared_storage/faultmonitor/20200220/
      Note: Reports will not be created for success status
      Monitor execution completed after 5 minutes
  4. If a check result is failed, review the console or log file for additional diagnostic information.

  5. Investigate and fix any detected problems. Repeat the check until the system passes all checks.

Scheduling Fault Monitor Tests

By default, the run_ovm_monitors, run_pca_healthcheck_monitors, and run_pca_upgrader_monitors checks are scheduled to run weekly. You can change the frequency of these checks or add other checks to the scheduler. You must restart the ovca service to implement any schedule changes.

  1. Using SSH and an account with superuser privileges, log in to the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Change the schedule properties in the ovca-system.conf file.

    Use the following scheduling format:

    * * * * *  command
    - - - - -
    | | | | |
    | | | | –---  day of week (0-7, Sunday= 0 or 7)
    | | | –-----  month (1-12)
    | | –-------  day of month (1-31)
    | –---------  hour (0-23)
    –-----------  minute (0-59)    
    [root@ovcamn05r1 ~]# cat /var/lib/ovca/ovca-system.conf
    
    [faultmonitor]
    report_path: /nfs/shared_storage/faultmonitor/
    report_format: json
    report_dir_cleanup_days: 10
    disabled_check_list: validate_image
    enable_phonehome: 0
    collect_report: 1
    
    [faultmonitor_scheduler]
    run_ovm_monitors: 0 2 * * *
    run_pca_healthcheck_monitors: 0 1 * * *
    run_pca_upgrader_monitors: 0 0 * * *
    repository_utilization_monitor: 0 */2 * * *
    check_ovmm_version: */30 * * * *

Changing Fault Monitoring Options

  1. Using SSH and an account with superuser privileges, log in to the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Change the appropriate property in the ovca-system.conf file.

    The report_format options are json, text, or html.

    [root@ovcamn05r1 ~]# cat /var/lib/ovca/ovca-system.conf
    
    [faultmonitor]
    report_path: /nfs/shared_storage/faultmonitor/
    report_format: json
    report_dir_cleanup_days: 10
    disabled_check_list: validate_image
    enable_phonehome: 1
    collect_report: 1

Phone Home Service

The Fault Monitoring utility is designed so that the management nodes collect fault data reports and copy those reports to the ZFS Storage Appliance.

If you want Oracle Service to monitor these fault reports, you can configure the Phone Home service to push these reports to Oracle on a weekly basis. Oracle Private Cloud Appliance uses the ZFS Storage Appliance Phone Home service.

Activating the Phone Home Service for Oracle Private Cloud Appliance

Use the following procedure to configure your system to send fault reports to Oracle for automated service response.

  1. Make sure Oracle Auto Service Request (ASR) is installed on the Private Cloud Appliance, the appliance is registered with ASR, and ASR is enabled on the appliance. See Oracle Auto Service Request (ASR).

  2. Using SSH and an account with superuser privileges, log in to the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  3. Enable Phone Home in the Fault Monitoring service.

    By default, Phone Home is disabled in Private Cloud Appliance.

    Use one of the following methods to enable the service:

    • Set the enable_phonehome property to 1 in the ovca-system.conf file on both management nodes.

      [root@ovcamn05r1 ~]# edit /var/lib/ovca/ovca-system.conf
      [faultmonitor]
      report_path: /nfs/shared_storage/faultmonitor/
      report_format: json
      report_dir_cleanup_days: 10
      disabled_check_list: validate_image
      enable_phonehome: 1
      collect_report: 1
    • Starting with release 2.4.4.1, you can also set the Phone Home property by using the Private Cloud Appliance CLI set system-property command. Execute the following command on both the active and passive management nodes:
      [root@ovcamn05r1 ~]# pca-admin
      Welcome to PCA! Release: 2.4.4.1
      PCA> set system-property phonehome enable
      Status: Success

      For more information, see set system-property.

  4. Enable Phone Home in the ZFS Storage Appliance browser interface.

    1. Log in to the ZFS Storage Appliance browser interface.

    2. Go to Configuration > Services > Phone Home.

    3. Click the power icon to bring the service online.

Data Collection for Service and Support

If Oracle Auto Service Request (ASR) is enabled on the Private Cloud Appliance, a service request will be created and sent to Oracle Support automatically for some failures. See Oracle Auto Service Request (ASR).

If an issue is not automatically reported by ASR, open a service request at My Oracle Support to request assistance from Oracle Support. You need an Oracle Premier Support Agreement and your Oracle Customer Support Identifier (CSI). If you do not know your CSI, find the correct Service Center for your country (https://www.oracle.com/support/contact.html), then contact Oracle Services to open a non-technical service request (SR) to get your CSI.

When you open a service request, provide the following information where applicable:

  • Description of the problem, including the situation where the problem occurs, and its impact on your operation.

  • Machine type, operating system release, browser type and version, locale and product release, patches that you have applied, and other software that might be affecting the problem.

  • Details of steps that you have taken to reproduce the problem.

  • Applicable logs and other support data.

This section describes how to collect an archive file with relevant log files and other health information, and how to upload the information to Oracle Support. Oracle Support uses this information to analyze and diagnose the issues. Oracle Support might request that you collect particular data and attach it to an existing service request.

Collecting Support Data

Collecting support data files involves logging in to the command line on components in your Private Cloud Appliance and copying files to a storage location external to the appliance environment, in the data center network. This can only be achieved from a system with access to both the internal appliance management network and the data center network. You can set up a physical or virtual system with those connections, or use the active management node.

The most convenient way to collect the necessary files, is to mount the target storage location on the system using nfs, and copy the files using scp with the appropriate login credentials and file path. The command syntax should be similar to the following example:

# mkdir /mnt/mynfsshare
# mount -t nfs storage-host-ip:/path-to-share /mnt/mynfsshare
# scp root@component-ip:/path-to-file /mnt/mynfsshare/pca-support-data/

Using kdump

Oracle Support Services often require a system memory dump. For this purpose, kdump must be installed and configured on the component under investigation. By default, kdump is installed on all Private Cloud Appliance compute nodes and configured to write the system memory dump to the ZFS Storage Appliance at this location: 192.168.4.100:/export/nfs_repository1/.

For more information, see How to Configure 'kdump' for Oracle VM 3.4.x (Doc ID 2142488.1).

Using OSWatcher

Oracle Support Services often recommend that the OSWatcher tool be run for an extended period of time to diagnose issues. OSWatcher is installed by default on all Private Cloud Appliance compute nodes.

For more information, see Oracle Linux: How To Start OSWatcher Black Box (OSWBB) Every System Boot Using RPM oswbb-service (Doc ID 580513.1).

Using pca-diag

Oracle Support Services use the pca-diag tool. This tool is part of the Private Cloud Appliance controller software installed on both management nodes and on all compute nodes.

The pca-diag tool collects troubleshooting information from the Private Cloud Appliance environment. For more information, see Oracle Private Cloud Appliance Diagnostics Tool.

Oracle Support might request specific output from pca-diag to help diagnose and resolve hardware or software issues.

Use the following procedure to use pca-diag to collect support data from your system.

  1. Log in to the active management node as root.

  2. Run pca-diag with the appropriate command-line arguments.

    For the most complete set of diagnostic data, run the command with both the ilom and vmpinfo arguments.

    • pca-diag ilom

      Use the pca-diag ilom command to detect and diagnose potential component hardware and software problems.

      [root@ovcamn05r1 ~]# pca-diag ilom
      
      Oracle Private Cloud Appliance diagnostics tool
      
      Gathering Linux information...
      Gathering system messages...
      Gathering PCA related files...
      Gathering OS version information...
      Gathering host specific information...
      Gathering PCI information...
      Gathering SCSI and partition data...
      Gathering OS process data...
      Gathering network setup information...
      Gathering installed packages data...
      Gathering disk information...
      Gathering ILOM Service Processor data... this may take a while
      Generating diagnostics tarball and removing temp directory
      
      ==============================================================================
      Diagnostics completed. The collected data is available in:
      /tmp/pcadiag_ovcamn05r1_<ID>_<date>_<time>.tar.bz2
      ==============================================================================
    • pca-diag vmpinfo

      Use the pca-diag vmpinfo command to detect and diagnose potential problems in the Oracle VM environment.

      Note:

      To collect diagnostic information for a subset of Oracle VM servers in the environment, run the command with an additional servers parameter, as shown in the following example:

      [root@ovcamn05r1 ~]# pca-diag vmpinfo servers='ovcacn07r1,ovcacn08r1'
      
      Oracle Private Cloud Appliance diagnostics tool
      Gathering Linux information...
      Gathering system messages...
      Gathering PCA related files...
      Gathering OS version information...
      Gathering host specific information...
      Gathering PCI information...
      Gathering SCSI and partition data...
      Gathering OS process data...
      Gathering network setup information...
      Gathering installed packages data...
      Gathering disk information...
      Gathering FRU data and console history. Use ilom option for complete ILOM data.

      When the vmpinfo3 script is called as a sub-process from pca-diag, the console output continues as follows:

      Running vmpinfo tool...
      Starting data collection
      Gathering files from servers: ovcacn07r1,ovcacn08r1  This process may take some time.
      Gathering OVM Model Dump files
      Gathering sosreports from servers
      The following server(s) will get info collected: [ovcacn07r1,ovcacn08r1]
      Gathering sosreport from ovcacn07r1
      Gathering sosreport from ovcacn08r1
      Data collection complete
      Gathering OVM Manager Logs
      Clean up metrics
      Copying model files
      Copying DB backup log files
      Invoking manager sosreport

      When all files have been collected, the data is compressed into two tarballs. One is from the pca-diag tool, while vmpinfo3 writes a separate tarball with its own specific data.

      Compressing VMPinfo3 date-time.
      =======================================================================================
      Please send /tmp/vmpinfo3-version-date-time.tar.gz to Oracle OVM support
      =======================================================================================
      
      Generating diagnostics tarball and removing temp directory
      ==============================================================================
      Diagnostics completed. The collected data is available in:
      /tmp/pcadiag_ovcamn05r1_ID_date_time.tar.bz2
      ==============================================================================
  3. If necessary, run pca-diag, with or without the ilom argument, on some or all compute nodes as well.

  4. To allow better analysis of physical server issues (for example hanging, crashing, or rebooting), also include the system memory dump file: vmcore. See the beginning of this section for a convenient way to collect the files.

    The location of the file is: kdump-partition-mount-point/kdump/compute-node-ip-date-time/vmcore. The partition and mount point are defined during kdump configuration. By default, kdump writes to 192.168.4.100:/export/nfs_repository1/.

    For more information, see How to Configure 'kdump' for Oracle VM 3.4.x (Doc ID 2142488.1).

  5. When required, collect the OSWatcher logs from the compute nodes. The default location is /var/log/oswatcher/archive/.

  6. Copy all diagnostic files to a location external to the appliance environment, as described at the beginning of this section.

Uploading Support Data Files

For support data files up to 2 GB, upload the file as part of the Service Request (SR) process in My Oracle Support.

  • If you are still in the process of creating the SR, upload the support data in the Upload Files/Attachments step.

  • To upload files after you have created the SR, use the following procedure:

    1. Log into My Oracle Support and open the Dashboard or the Service Request tab.

    2. In the Service Request region, click the SR you want to update.

    3. In the Update section, select Add Attachment.

    4. In the pop-up window, select the file for upload, include any notes, and click Attach File.

If uploading the support data with the SR is not an option, or for support data files larger than 2 GB, use the FTPS file upload service from Oracle Support at transport.oracle.com, as described in the following procedure. Oracle Support might request that you upload using a different mechanism.

  1. Using an FTPS client, for example FileZilla or WinSCP, access the My Oracle Support File Upload Service transport.oracle.com in passive mode.

  2. Log in with your Oracle Single Sign-On user name and password.

  3. Select the support data file to upload.

  4. Select a destination for the file.

    Use the directory path provided by Oracle Support.

    Typically, the directory path is constructed as follows: /upload/issue/sr_number/.

    The use of an SR number ensures that the file is correctly associated with the service request. Record the full path to the file and the SR number for future reference in communications with Oracle Support.

  5. Upload the file.

    When the upload is complete, a confirmation message is displayed.

If you prefer to use a command-line client such as cURL, you typically enter a single command to connect, authenticate, and complete the upload. A cURL command will look similar to the following example:

curl -T path_to_file -u "user" ftps://transport.oracle.com/upload/issue/sr_number

For security reasons, you should omit the password from the command and instead enter the password at the prompt.

For detailed information about uploading files to Oracle Support, see How to Upload Files to Oracle Support (Doc ID 1547088.2).

Cloud Backup

The Oracle Private Cloud Appliance Cloud Backup service automates the backup of critical components and configuration data to your customer tenancy in Oracle Cloud Infrastructure (OCI). This feature is designed to recover a Private Cloud Appliance to a running state after a catastrophic event. This feature is not designed to backup virtual machines, guest operating systems, or applications and data hosted on virtual machines. Backups of customer virtual machines and applications can be managed using Oracle Site Guard. See Oracle VM 3: Getting Started with Disaster Recovery using Oracle Site Guard (Doc ID 1959182.1).

The Cloud Backup service requires an Oracle Cloud Infrastructure cloud tenancy. The service is designed to create a snapshot of backup data from the system, store that snapshot on the internal ZFS Storage Appliance, then push that snapshot to your Oracle Cloud Infrastructure cloud tenancy for remote storage. Once configured, the service automatically runs a backup weekly. For resource management reasons, the 10 latest backups are stored locally on the Oracle Cloud Infrastructure and on your Oracle Cloud Infrastructure tenancy. At this time, contact Oracle Service to restore your Private Cloud Appliance from an Oracle Cloud Infrastructure cloud backup.

The Cloud Backup service uses the object storage feature of Oracle Cloud Infrastructure to store your Private Cloud Appliance configuration backup data. With Object Storage, you can safely and securely store or retrieve data directly from the internet or from within the cloud platform. Object Storage is a regional service and is not tied to any specific compute instance. You can access data from anywhere inside or outside the context of the Oracle Cloud Infrastructure, as long you have internet connectivity and can access one of the Object Storage endpoints. For more information about Object Storage, see the Object Storage Overview.

To use the Cloud Backup service with Private Cloud Appliance releases earlier than 2.4.3, or systems that have been upgraded to release 2.4.3, contact Oracle Service.

For the best experience using the Cloud Backup service, consider these items.

  • Use an Oracle Private Cloud Appliance region that is in the same region as your Private Cloud Appliance.

  • Very slow network speeds in the customer premise network (less than 100 Mbps) may result in timeouts, especially when crossing regions.

  • If you experience timeouts, contact Oracle Service.

  • If the connection to the ZFS Storage Appliance is severed, for example when a management node is rebooted, this could corrupt the Cloud Backup service. See "Cloud Backup Task Hangs When a ZFSSA Takeover is Performed During Backup" in Known Limitations and Workarounds in the Oracle Private Cloud Appliance Release Notes.

Configuring the Cloud Backup Service

This section describes how to initially configure the Cloud Backup service, including how to prepare your Oracle Cloud Infrastructure tenancy to receive backups from the Oracle Private Cloud Appliance.

Configuring the Cloud Backup service does three things: creates a location to store your backups on your Oracle Cloud Infrastructure tenancy, activates the script which gathers backup data from the Private Cloud Appliance, and finally pushes those backups from your Private Cloud Appliance to your Oracle Cloud Infrastructure tenancy on a weekly basis.

Configuring the Cloud Backup Service for Oracle Private Cloud Appliance

  1. Create an object storage bucket on your Oracle Cloud Infrastructure tenancy. See "Creating a Bucket" in Putting Data into Object Storage.

    Note:

    To locate the OCID for a bucket, see Managing Buckets.

    Each target must be associated with its own bucket. Perform this operation to set up each target location for your Private Cloud Appliance backups.

  2. Set up the Oracle Private Cloud Appliance Cloud Backup configuration.

    1. Using SSH and an account with superuser privileges, log into the active management node.

      # ssh root@10.100.1.101
      root@10.100.1.101's password:
      root@ovcamn05r1 ~]#
    2. Launch the Private Cloud Appliance command line interface.

      # pca-admin
      Welcome to PCA! Release: 2.4.4
      PCA>
    3. Create an Oracle Cloud Infrastructure target on your Private Cloud Appliance that corresponds with the Oracle Cloud Infrastructure object store bucket created in step 1.

      This step creates a target on your Private Cloud Appliance ZFS Storage Appliance that sends scheduled backups to an object storage bucket on Oracle Cloud Infrastructure. For more information see create oci-target.

      PCA> create oci-target target_name target_location target_user 
           target_bucket target_tenancy keyfile

      For example:

      PCA> create oci-target cloud_target_1 https://objectstorage.us-oci.com ocid1.user.oc1..oos-test 
           mybucket ocid1.tenancy.oc1..nobody /root/oci_api_key.pem
      
      Status: Success

      The cloud backup is now configured to run weekly.

Configuring a Manual Cloud Backup

This section describes how to trigger a manual cloud backup, which can be useful in preparation for a system upgrade.

Creating a Manual Cloud Backup

  1. Using SSH and an account with superuser privileges, log into the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Launch the Private Cloud Appliancecommand line interface.

    # pca-admin
    Welcome to PCA! Release: 2.4.4
    PCA>
  3. Create the Cloud Backup.

    PCA> create oci-backup oci-target-name1 oci-target-name2
    
    The create oci-backup job has been submitted. Use "show task < task id>" to monitor the progress.
    
    Task_ID         Status  Progress Start_Time           Task_Name       
    
    -------         ------  -------- ----------           ---------       
    
    386c911399b38e  RUNNING None     05-29-2020 21:48:24  oci_backup      
    
    ---------------
    
    1 row displayed
    
    Status: Success

    Only one backup can run at a time. If there is a conflict, you see this error:

    Status: Failure
    
    Error Message: Error (SYSTEM_002): Cannot invoke API function oci_backup while lock oci_backup is in place.

    To resolve this issue, run your manual backup again, once the other backup task is complete.

Deleting Cloud Backups

This section describes how to delete a Cloud Backup, which removes the backup from both the Private Cloud Appliance and your Oracle Cloud Infrastructure tenancy.

Deleting a Cloud Backup

  1. Using SSH and an account with superuser privileges, log into the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Launch the Private Cloud Appliance command line interface.

    # pca-admin
    Welcome to PCA! Release: 2.4.4
    PCA>
  3. Delete the backup.

    PCA> delete oci-backup <OVCA/OCI_backups@AK00000000_OCI_snap_2020_06_29-04.56.28
    >
    
    WARNING !!! THIS IS A DESTRUCTIVE OPERATION.
    
    Are you sure [y/N]:y

Deleting Oracle Cloud Infrastructure Targets

This section describes how to remove an Oracle Cloud Infrastructure target from your Private Cloud Appliance. The related object storage buckets in your Oracle Cloud Infrastructure tenancy are not removed. This operation removes only the selected target on your Private Cloud Appliance, thus breaking the link to that target in your Oracle Cloud Infrastructure tenancy.

Deleting a Target

  1. Using SSH and an account with superuser privileges, log into the active management node.

    # ssh root@10.100.1.101
    root@10.100.1.101's password:
    root@ovcamn05r1 ~]#
  2. Launch the Private Cloud Appliance command line interface.

    # pca-admin
    Welcome to PCA! Release: 2.4.4
    PCA>
  3. Delete the Oracle Cloud Infrastructure target on your Private Cloud Appliance.

    PCA> delete oci-target <target>