Skip Headers
Oracle® Application Server Installation Guide
10g Release 3 (10.1.3.2.0) for HP-UX PA-RISC (64-Bit)

Part Number B32406-01
Go to Documentation Home
Home
Go to Book List
Book List
Go to Table of Contents
Contents
Go to Index
Index
Go to Feedback page
Contact Us

Go to previous page
Previous
Go to next page
Next
View PDF

6 Installing in High Availability Environments

This chapter provides an overview of the high availability configurations supported by Oracle Application Server and instructions for installation.

Contents of this chapter:

6.1 Overview of High Availability Configurations

This chapter provides only a brief overview of the high availability configurations in Oracle Application Server. For a complete description of the configurations, see the Oracle Application Server High Availability Guide.

Oracle Application Server supports the following types of high availability configurations at installation time. Note that there are multiple variants of each type.

For a quick summary of the high availability configurations, see Section 6.1.4, "Summary of Differences".

6.1.1 Active-Active Topologies: OracleAS Clusters

Oracle Application Server provides an active-active redundant model for all its components. In an active-active topology, two or more Oracle Application Server instances are configured to serve the same workload. These instances can run on the same machine or on different machines.

The instances are front-ended by an external load balancer, which directs requests to any of the active instances. Instead of an external load balancer, you can also run a software load balancer to distribute the requests. In production environment, however, a hardware load balancer is recommended.

Common properties of an active-active topology include:

  • Similar instance configuration

    The instances need to serve the same workload or applications. Some configuration properties should have similar values across instances so that the instances can deliver the same reply to the same request. Other configuration properties may be instance-specific, such as local host name information.

    If you make a configuration change to one instance, you should also make the same change to the other instances in the active-active topology. The "Configuring and Managing Clusters" chapter in the Oracle Containers for J2EE Configuration and Administration Guide lists the files that contain properties that should be replicated.

  • Independent operation

    If one Oracle Application Server instance in an active-active topology fails, the other instances in the cluster continue to serve requests. The load balancer directs requests only to instances that are alive.

Advantages of an active-active topology include:

  • Increased availability

    An active-active topology is a redundant configuration. Loss of one instance can be tolerated because other instance can continue to serve the same requests.

  • Increased scalability and performance

    Multiple identically-configured instances provide the capability to share a workload among different machines and processes. You can scale the topology by adding new instances as the number of requests increase.

For instructions on creating the OracleAS Clusters configuration, see Section 6.3, "Creating the Active-Active Topology".

6.1.2 Active-Passive Topologies: OracleAS Cold Failover Clusters

Oracle Application Server provides an active-passive model for all its components in OracleAS Cold Failover Clusters. In an OracleAS Cold Failover Cluster topology, two Oracle Application Server instances are configured to serve the same application workload but only one is active at any particular time. The passive instance runs (that is, becomes active) only when the active instance fails. These instances run on nodes that are in a hardware cluster.

Common properties of an OracleAS Cold Failover Cluster topology include:

  • Hardware cluster

    In an OracleAS Cold Failover Cluster topology, you run Oracle Application Server on machines that are in a hardware cluster, with vendor clusterware running on the machines.

  • Shared storage

    You install the Oracle home for the Oracle Application Server instance on storage shared by the machines in the hardware cluster.

    The active node in the OracleAS Cold Failover Cluster topology mounts the shared storage so that it has access to the Oracle home. If it fails, the passive instance mounts the shared storage and accesses the same Oracle home.

  • Virtual hostname

    The virtual hostname gives clients a single system view of the Oracle Application Server middle tier. Clients use the virtual hostname to access the Oracle Application Server middle tier.

    The virtual hostname is associated with a virtual IP. This name-IP entry must be added to the DNS that the site uses. For example, if the two physical hostnames of the hardware cluster are node1.mycompany.com and node2.mycompany.com, the single view of this cluster can be provided by the virtual hostname apps.mycompany.com. In the DNS, apps maps to a virtual IP address that floats between node1 and node2 via a hardware cluster. Clients access Oracle Application Server using apps.mycompany.com; they do not know which physical node is active and actually servicing a particular request.

    You can specify the virtual hostname during installation. See the Oracle Application Server Installation Guide.

  • Failover procedure

    An active-passive configuration also includes a set of scripts and procedures to detect failure of the active instance and fail over to the passive instance while minimizing downtime.

Advantages of an OracleAS Cold Failover Cluster topology include:

  • Increased availability

    If the active instance fails for any reason or must be taken offline, an identically configured passive instance is prepared to take over at any time.

  • Reduced operating costs

    In an active-passive topology only one set of processes is up and serving requests. Managing the active instance is generally easier than managing an array of active instances.

  • Application independence

    Some applications may not be suited to an active-active topology. This may include applications that rely heavily on application state or on information stored locally. An active-passive topology has only one instance serving requests at any particular time.

For instructions on creating the OracleAS Cold Failover Cluster configuration, see Section 6.4, "Creating the Active-Passive Topology".

6.1.3 OracleAS Disaster Recovery

OracleAS Disaster Recovery configurations have the following characteristics:

  • A production site and a standby site that mirrors the production site. Typically, these sites are located some distance from each other to guard against site failures such as floods, fires, or earthquakes. During normal operation, the production site handles all the requests. If the production site goes down, the standby site takes over and handles all the requests.

  • Each site has all the hardware and software to run. It contains nodes for running Oracle Application Server instances, load balancers, and DNS servers.

For installation details, see Section 6.5, "Creating an OracleAS Disaster Recovery Configuration".

6.1.4 Summary of Differences

Table 6-1 summarizes the differences among the high availability configurations.

Table 6-1 Differences Among the High Availability Configurations


OracleAS Cold Failover Cluster
OracleAS Clusters
OracleAS Disaster Recovery

Node configuration

Active-Passive

Active-Active

Active-Passive

Hardware cluster

Yes

No

Optional (hardware cluster required only if you installed the OracleAS Infrastructure in an OracleAS Cold Failover Cluster configuration)

Virtual hostname

Yes

No

Yes

Load balancer

No

Yes

No

Shared storage

Yes

No

No


6.2 Requirements for High Availability Configurations

This section describes the requirements common to all high availability configurations. In addition to these common requirements, each configuration has its own specific requirements. See the individual chapters for details.

Note:

You still need to meet the requirements listed in Chapter 2, "Requirements", plus requirements specific to the high availability configuration that you plan to use.

The common requirements are:

6.2.1 Check Minimum Number of Nodes

You need at least two nodes in a high availability configuration. If a node fails for any reason, the second node takes over.

6.2.2 Check That Groups Are Defined Identically on All Nodes

Check that the /etc/group file on all nodes in the cluster contains the operating system groups that you plan to use. You should have one group for the oraInventory directory, and one or two groups for database administration. The group names and the group IDs must be the same for all nodes.

See Section 2.7, "Operating System Groups" for details.

6.2.3 Check the Properties of the oracle User

Check that the oracle operating system user, which you log in as to install Oracle Application Server, has the following properties:

  • Belongs to the oinstall group and to the osdba group. The oinstall group is for the oraInventory directory, and the osdba group is a database administration group. See Section 2.7, "Operating System Groups" for details.

  • Has write privileges on remote directories.

6.2.4 Check for Previous Oracle Installations on All Nodes

Check that all the nodes where you want to install in a high availability configuration do not have existing oraInventory directories.

Details of all Oracle software installations are recorded in the Oracle Installer Inventory directory. Typically, this directory is unique to a node and named oraInventory. The directory path of the Oracle Installer Inventory directory is stored in the oraInst.loc file.

The existence of this file on a node confirms that the node contains some Oracle software installation. Since the high availability configurations require installations on multiple nodes with Oracle Installer Inventory directories on a file system that may not be accessible on other nodes, the installation instructions in this chapter and subsequent chapters for high availability configurations assume that there have not been any previous installations of any Oracle software on any of the nodes that are used for this high availability configuration. The oraInst.loc file and the Oracle Installer Inventory directory should not exist on any of these nodes prior to these high availability installations.

To check if a node contains an oraInventory directory that could be detected by the installer:

  1. On each node, check for the existence of the oraInst.loc file. This file is stored in the /var/opt directory.

    If a node does not contain this file, then it does not have an oraInventory directory that will be used by the installer. You can check the next node.

  2. For nodes that contain the oraInst.loc file, rename the file and the oraInventory directory. The installer then prompts you to enter a location for a new oraInventory directory.

    For example, enter the following commands as root:

    # cat /etc/oraInst.loc
    inventory_loc=/localfs/app/oracle/oraInventory 
    inst_group=dba 
    # mv /etc/oraInst.loc /etc/oraInst.loc.orig 
    # mv /localfs/app/oracle/oraInventory /localfs/app/oracle/oraInventory.orig
    
    

Since the oraInst.loc file and the Oracle Installer Inventory directory are required only during the installation of Oracle software, and not at runtime, renaming them and restoring them later does not affect the behavior of any installed Oracle software on any node. Make sure that the appropriate oraInst.loc file and Oracle Installer Inventory directory are in place before starting the Oracle Universal Installer.

Note:

For an OracleAS Disaster Recovery configuration, the correct oraInst.loc file and associated oraInventory directory are required during normal operation, not just during installation.

6.3 Creating the Active-Active Topology

This section describes how to install Oracle Application Server in an active-active topology with OracleAS Clusters. OracleAS Clusters is one of the high availability environments supported by Oracle Application Server.

Contents of this section:

6.3.1 Active-Active Topologies: Introduction

An active-active topology consists of redundant Oracle Application Server instances that deliver greater scalability and availability than a single instance. Active-active topologies remove the single point of failure that a single instance poses. While a single Oracle Application Server instance leverages the resources of a single host, a cluster of Oracle Application Server instances spans multiple hosts, distributing application execution over a greater number of CPUs. A single Oracle Application Server instance is vulnerable to the failure of its host and operating system, but an active-active topology continues to function despite the loss of an operating system or a host, hiding any such failure from clients.

In active-active topologies, all the instances are active at the same time. This is different from active-passive topologies, where only one instance is active at any time.

The nodes in the active-active topologies are not in a hardware cluster.

Load Balancer Requirements

Active-active topologies use a load balancer to direct requests to one of the Oracle Application Server instances in the topology. In other words, the Oracle Application Server instances are fronted by the load balancer.

You configure the load balancer with virtual server names for HTTP and HTTPS traffic. Clients use the virtual server names in their requests. The load balancer directs requests to an available Oracle Application Server instance.

See the Oracle Application Server High Availability Guide for a list of features that your load balancer should have.

Shared Storage for Oracle Metadata Services (MDS)

In addition to the local drives on each node, you need a shared storage that can be accessed by all the nodes in an active-active topology. On the shared storage, you need to create a directory for Oracle Metadata Services (MDS). The MDS directory stores data about your deployed applications.

By having the MDS directory on the shared storage, you ensure that all Oracle Application Server instances access the same data and you do not have to worry about replicating the data.

The mount point for the MDS directory must be the same for all nodes in the active-active topology (for example, /oracle/mds).

You do not need the shared storage during installation, but you need it when you deploy your applications. You specify the path to the MDS directory during application deployment.

Figures of Active-Active Topologies

The following figures show two active-active topologies. The difference in the topologies is whether you install Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB in the same Oracle home or in separate Oracle homes.

Figure 6-1 shows an active-active topology with Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB in the same Oracle home. Figure 6-2 shows an active-active topology with Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB in separate Oracle homes.

Figure 6-1 Active-Active Topology with Oracle HTTP Server and WebCenter Suite in the Same Oracle Home

Collocated Active-Active topology
Description of "Figure 6-1 Active-Active Topology with Oracle HTTP Server and WebCenter Suite in the Same Oracle Home"

Figure 6-2 Active-Active Topology with Oracle HTTP Server, WebCenter Framework, and Oracle Content DB in Separate Oracle Homes

Distributed Active-Active Topology
Description of "Figure 6-2 Active-Active Topology with Oracle HTTP Server, WebCenter Framework, and Oracle Content DB in Separate Oracle Homes"

6.3.2 OracleAS Clusters in Active-Active Topologies

All the Oracle Application Server instances in an active-active topology belong to the same cluster. Oracle HTTP Server can forward application requests to any OC4J instance in the same cluster. This is useful because the WebCenter Framework and Oracle Content DB components run in OC4J containers:

Table 6-2 WebCenter Framework and Oracle Content DB and their OC4J Instances

Component Runs in this OC4J Instance

WebCenter Framework


OC4J_WebCenter

Oracle Content DB


OC4J_Content


You can group instances in a cluster using one of the following ways:

  • All the instances use the same multicast IP address and port.

  • All the instances are chained to the same discovery server.

  • Each instance specifies all other instances in the opmn.xml configuration file.

  • If the instances run on nodes that are on different subnets, you have to designate a node to be the gateway server, which bridges the instances on the different subnets.

OracleAS Clusters also enable you to use the @cluster parameter in some opmnctl commands. Commands that use the @cluster parameter apply to all instances in the cluster. For example, you can use the @cluster parameter to start all components in all instances in the cluster.

OC4J instances in a cluster have the following features:

  • OC4J instances have cluster-wide properties as well as instance-specific properties. Cluster-wide properties are properties whose values should be similar for all OC4J instances in the cluster. Instance-specific properties are properties that have different values for each OC4J instance. For a list of cluster-wide properties, see the "Configuring and Managing Clusters" chapter in the Oracle Containers for J2EE Configuration and Administration Guide.

  • If you modify a cluster-wide property in one OC4J instance, you should propagate the change to all other OC4J instances in the cluster.

  • When you deploy an application to an OC4J instance, you also need to deploy it on all other OC4J instances in the cluster.

  • The number of OC4J processes is an instance-specific property: it can be different for each OC4J instance. This must be configured for each Oracle Application Server instance in the cluster. The OC4J process configuration provides flexibility to tune according to the specific hardware capabilities of the host. By default, each OC4J instance is instantiated with a single OC4J process.

For details, see the "Configuring and Managing Clusters" chapter in the Oracle Containers for J2EE Configuration and Administration Guide.

6.3.3 Properties of Oracle Application Server Instances in Active-Active Topologies

Because the load balancer can send a request to any Oracle Application Server instance in the topology, you need to ensure that the instances are configured in the same manner so that clients get the same response regardless of which instance handles the request. This includes the following:

  • Deploy the same applications on each OC4J instance in the topology.

  • Ensure that you replicate state and stateful session bean information across OC4J instances so that in the event that an OC4J instance fails, another OC4J instance contains the state information and can continue the session.

  • Ensure that configuration properties for all the OC4J instances in the topology are identical. These configuration properties are listed in chapter 8, "Configuring and Managing Clusters", in section "Replicating Changes Across a Cluster", in the Oracle Containers for J2EE Configuration and Administration Guide.

6.3.4 Installation Steps for Active-Active Topologies

To create the topology shown in Figure 6-1 or Figure 6-2, you perform the following steps:

Step 1: Install a High Availability Oracle Database

Step 2: Install Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB and Cluster the Instances using OPMN

Step 3: Disable Oracle HTTP Server on Nodes with Oracle Content DB (Distributed Installations Only)

Step 4: Configure the Load Balancer with Virtual Server Names

Step 5: (optional) Configure SSL on Both Nodes

Step 6: Update Domain Properties for Oracle Content DB

Step 7: Reload OPMN

The following sections describe the steps in detail.


Step 1   Install a High Availability Oracle Database

You need a database for Oracle Content DB. In high availability topologies, you should use high availability databases, such as cold failover cluster database or Real Application Clusters database.

See Section 2.4, "Requirements for Oracle Content Database" for more information on database requirements for Oracle Content DB.

Step 2   Install Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB and Cluster the Instances using OPMN

You can install Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB in the same Oracle home (see Figure 6-1), or in different Oracle homes (see Figure 6-2).

For Oracle Application Server instances that you want to group in the same active-active topology, you need to place them in the same cluster. This enables communication between the Oracle HTTP Server and OC4J instances, and simplifies the management of Oracle Application Server instances. OracleAS Clusters enable you to use the @cluster parameter for the opmnctl command to manage all the instances in the cluster.

You can create clusters using one of the following methods:

  • Dynamic Discovery Method

    In this method, each ONS node within the same subnet announces its presence with a multicast message. The cluster topology map for each node is automatically updated as nodes are added or removed, enabling the cluster to be self-managing.

    If you use this method and you are performing a basic installation, you can define a cluster postinstallation by following the steps in Section 6.3.5.1, "Setting up Clusters with the Dynamic Discovery Method". If you use this method and you are performing an advanced installation, you should specify the multicast address and port on the Cluster Topology Configuration screen in the installer.

  • Discovery Server Method

    In this method, specific nodes within a cluster are configured to serve as "discovery servers", which maintain the topology map for the cluster; the remaining nodes then connect with one another via this server.

    If you use this method, you can define a cluster for OPMN by specifying the names of the Oracle Application Server instances explicitly in the opmn.xml file of each instance by following the steps in Section 6.3.5.2, "Setting up Clusters with the Discovery Server Method" after installation.

  • Gateway Method

    This configuration is used to connect topologies separated by firewalls or on different subnets using specified "gateway" nodes.

    If you use this method, see the section "Configuring Cross-Topology Gateways" in the Oracle Containers for J2EE Configuration and Administration Guide for configuration details.

You can perform either an integrated installation or a distributed installation.

  • For Integrated Installations (Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB in the Same Oracle Home)

    You install Oracle Application Server on the local storage of each node in the active-active topology.

    Perform a basic installation as described in Section 4, "Basic Installation" so that Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB will run from the same Oracle home.

    Note the following:

    • Set the Oracle home to be on the local storage of each node.

    • For deployment to work across all instances in a cluster, the oc4jadmin password must be the same for all instances in the cluster.

    • To simplify administering the instances, use the same Oracle home path and the same instance name for each node.

    • If you are using the dynamic discovery method to cluster the Oracle Application Server instances for OPMN, be sure to perform the steps in Section 6.3.5.1, "Setting up Clusters with the Dynamic Discovery Method" after installation.

    • If you are using the discovery server method to cluster the Oracle Application Server instances for OPMN, be sure to perform the steps in Section 6.3.5.2, "Setting up Clusters with the Discovery Server Method" after installation.

    • If you are using the gateway method to cluster the Oracle Application Server instances for OPMN, see the section "Configuring Cross-Topology Gateways" in the Oracle Containers for J2EE Configuration and Administration Guide for configuration details.

  • For Distributed Installations (Oracle HTTP Server, Oracle WebCenter Framework, and Oracle Content DB in Different Oracle Homes)

    You install Oracle Application Server on the local storage of each node in the active-active topology.

    For the nodes where you want to run Oracle HTTP Server, follow the steps in Section 5.2.5, "Installing Oracle HTTP Server".

    For the nodes where you want to run Oracle WebCenter Framework and Oracle Content DB, you must perform the following two installations on each node:

    During installation, select the following options:

    • On the Administration Settings screen:

      • If you want this node to administer the cluster using Application Server Control, select Start Oracle Enterprise Manager 10g ASControl in this instance. In a cluster topology, only one instance should be configured as an Administration OC4J instance. Note that the Administration OC4J instance for the cluster does not have to be the first installed node.

      • If you do not want this node to administer the cluster, deselect Start Oracle Enterprise Manager 10g ASControl in this instance.

    • If you are using the dynamic discovery method to cluster the Oracle Application Server instances for OPMN, perform the following:

      • If you are installing Oracle HTTP Server, select Configure this Oracle HTTP Server instance to be part of an Oracle Application Server cluster on the "Cluster Topology Configuration" screen. Specify the IP Address and Port for the multicast address shared by all the nodes in the cluster.

      • If you are installing Oracle WebCenter Framework, select Configure this instance to be part of an Oracle Application Server cluster topology on the "Cluster Topology Configuration" screen. Specify the IP Address and Port for the multicast address shared by all the nodes in the cluster and select Access this OC4J Instance from a separate Oracle HTTP Server.

      • If you are installing Oracle Content DB, select Configure this instance to be part of an Oracle Application Server cluster topology on the "Cluster Topology Configuration" screen. Specify the IP Address and Port for the multicast address shared by all the nodes in the cluster.

      Note that the multicast address must be between 224.0.1.0 and 239.255.255.255. If you are installing on the first node in the cluster, you may choose any IP address and port, as long as it falls in the multicast address range.

    Note the following:

    • Set the Oracle home to be on the local storage of each node.

    • For deployment to work across all instances in a cluster, the oc4jadmin password must be the same for all instances in the cluster.

    • Ensure that the same component uses the same port number in each Oracle Application Server instance in the cluster. For example, ensure that Oracle HTTP Server is listening at the same port number for all instances in the cluster.

    • To simplify administering the instances, use the same Oracle home path and the same instance name for each node.

    • If you are using the discovery server method to cluster the Oracle Application Server instances for OPMN, be sure to perform the steps in Section 6.3.5.2, "Setting up Clusters with the Discovery Server Method" after installation.

    • If you are using the gateway method to cluster the Oracle Application Server instances for OPMN, see the section "Configuring Cross-Topology Gateways" in the Oracle Containers for J2EE Configuration and Administration Guide for configuration details.

Step 3   Disable Oracle HTTP Server on Nodes with Oracle Content DB (Distributed Installations Only)

This step is required only if you performed a distributed installation in step 2.

Disable Oracle HTTP Server on the Oracle homes where you installed Oracle Content DB.

  1. Set the status of Oracle HTTP Server to disabled in the ORACLE_HOME/opmn/conf/opmn.xml file:

    <ias-component id="HTTP_Server" status="disabled">
    
    
  2. Stop and restart all components in the Oracle home.

    > ORACLE_HOME/bin/opmnctl stopall
    > ORACLE_HOME/bin/opmnctl startall
    
    

Step 4   Configure the Load Balancer with Virtual Server Names

Refer to your load balancer documentation for configuration steps. On your load balancer, you need to configure a virtual server name and port for HTTP traffic, and another virtual server name and port for HTTPS traffic. The port numbers for the virtual server names should match the port numbers at which Oracle HTTP Server is listening. Clients will use the virtual server names and ports to access Oracle Application Server instances.

Step 5   (optional) Configure SSL on Both Nodes

This step is required only if you are using SSL.

Edit the ORACLE_HOME/Apache/Apache/conf/httpd.conf file on each node as follows:

  1. Add the following line to the httpd.conf to load the certheaders_module.

    LoadModule certheaders_module libexec/mod_certheaders.so
    
    

    Note:

    If you installed Oracle HTTP Server 2.0 using the Oracle Application Server Companion CD, add the following line to the httpd.conf file:
    LoadModule certheaders_module modules/mod_certheaders.so
    
    
  2. Add a virtual host directive.

    <VirtualHost *:7777>
    ServerName myworkplace.com
    Port 443
    ServerAdmin you@youraddress.com
    RewriteEngine On
    RewriteOptions inherit
    SimulateHTTPS On
    </VirtualHost>
    
    
  3. Restart Oracle HTTP Server on all the nodes.

Step 6   Update Domain Properties for Oracle Content DB

Update these domain properties using Application Server Control on each node:

  • Set IFS.DOMAIN.APPLICATION.ApplicationHost to the virtual hostname configured on the load balancer.

  • Set IFS.DOMAIN.APPLICATION.ApplicationPort to the port number configured on the load balancer.

To edit these properties using Application Server Control:

  1. On the Cluster Topology page, expand OC4J_Content.

  2. Click content located under OC4J_Content. This displays the "Application: content" page.

  3. On the "Application: content" page, click Content DB Extension. This displays the "Content DB: content" page.

  4. On the "Content DB: content" page, click the Administration tab.

  5. Click the icon in the Go To Task column for Domain Properties.

  6. On the Domain Properties page, click the IFS.DOMAIN.APPLICATION.ApplicationHost property, and set its value to the virtual hostname configured on the load balancer. Click OK.

  7. On the Domain Properties page, click the IFS.DOMAIN.APPLICATION.ApplicationPort property, and set its value to the port number configured on the load balancer. Click OK.

Step 7   Reload OPMN

Reload OPMN on each node.

> ORACLE_HOME/opmn/bin/opmnctl reload

6.3.5 Supporting Procedures for Creating the Active-Active Topology

This section describes some common procedures that you may need to perform to maintain the active-active topology:

6.3.5.1 Setting up Clusters with the Dynamic Discovery Method

In this method, you define the same multicast address and port for each Oracle Application Server instance in the cluster. An advantage in using this method is that you do not have to specify the name of each Oracle Application Server instance in the cluster. You can add or remove instances from the cluster by editing the multicast address and port.

  1. For each Oracle Application Server instance that you want to group in the same cluster, run the following command:

    opmnctl config topology update discover="*<multicastAddress>:<multicastPort>"
    
    

    multicastAddress specifies the multicast address that you want to use for this cluster. The multicast address must be within the valid address range, which is 224.0.1.0 to 239.255.255.255. Note that the multicast address is preceded by a * character in the command.

    multicastPort can be any unused port number.

    Example:

    > ORACLE_HOME/opmn/bin/opmnctl config topology update
                             discover="*225.0.0.20:8001"
    
    

    In distributed installations (Oracle HTTP Server and OC4J on different Oracle homes), you need to cluster all the Oracle Application Server instances into the same cluster. You need to use the same multicast IP and port for all the instances.

  2. For each Oracle Application Server instance where you ran the opmnctl config topology update command, run the opmnctl reload command to force OPMN to read the updated opmn.xml file.

    > ORACLE_HOME/opmn/bin/opmnctl reload
    
    

6.3.5.2 Setting up Clusters with the Discovery Server Method

If you do not want to use the multicast method, you can define a cluster by specifying the names of the nodes running the Oracle Application Server instances in the opmn.xml file of each instance.

Example: if you want to cluster four instances (inst1.node1.mycompany.com, inst2.node2.mycompany.com, inst3.node3.mycompany.com, inst4.node4.mycompany.com), you would perform these steps:

  1. Designate at least one of the instances to serve as the discovery server. The discovery server maintains the topology for the cluster.

    This example assumes that inst1.node1.mycompany.com and inst2.node2.mycompany.com will be the discovery servers for the cluster.

    In distributed installations (Oracle HTTP Server and OC4J on different Oracle homes), any instance, whether running Oracle HTTP Server or OC4J, can serve as the discovery server.

  2. In the opmn.xml file for all instances in the cluster, specify the nodes that are running the discovery servers (node1.mycompany.com and node2.mycompany.com in the example).

    In the example, the opmn.xml file is changed to include the following lines:

    <notification-server>
       <topology>
          <discover
              list="node1.mycompany.com:6201,node2.mycompany.com:6201"/>
       </topology>
    ...
    </notification-server>
    
    

    The 6201 specifies the port number at which the notification server is listening. You can find this value in the opmn.xml file of that instance.

    If you have more than one discovery server, you separate them with the comma character.

  3. On all the instances, run opmnctl reload to force OPMN to read the updated opmn.xml file.

    > ORACLE_HOME/opmn/bin/opmnctl reload
    
    

6.4 Creating the Active-Passive Topology

This section describes how to install Oracle Application Server in an active-passive topology with OracleAS Cold Failover Cluster. OracleAS Cold Failover Cluster is one of the high availability environments supported by Oracle Application Server.

Contents of this section:

6.4.1 Active-Passive Topologies: Introduction

An active-passive topology consists of the following:

  • Two nodes in a hardware cluster

  • A virtual hostname and IP address

  • A shared storage, to be shared between the two nodes

You install the Oracle home on the shared storage. During runtime in an active-passive topology, only one node is active. The other node is passive. The active node mounts the shared storage so that it can access the files and runs all the processes and handles all the requests. Clients access the active node through the virtual hostname. Clients do not need to know the physical hostnames of the nodes in the topology.

If the active node fails for any reason, a failover event occurs and the passive node takes over and becomes the active node. It mounts the shared storage and runs all the processes and handles all the requests. The virtual hostname and IP now point to the passive node. Clients, because they access the nodes using the virtual hostname, do not know that it is the passive node that is servicing their requests.

The nodes need to be in hardware cluster to enable failover.

Note:

Installing the Oracle home on the local storage of each node in the OracleAS Cold Failover Cluster topology is not supported. You have to install it on the shared storage.

Vendor Clusterware

The two nodes in an active-passive topology are in a hardware cluster, which typically includes some vendor clusterware. For a list of certified clusterware, visit the Oracle Technology Network website (http://www.oracle.com/technology).

These products must be installed on both nodes (active and passive) in the topology.

Directory for Oracle Metadata Services (MDS) on the Shared Storage

On the shared storage, you also need to create a directory for Oracle Metadata Services (MDS). The MDS directory stores data about your deployed applications.

By having the MDS directory on the shared storage, you ensure that all Oracle Application Server instances access the same data and you do not have to worry about replicating the data.

The mount point for the MDS directory must be the same for all nodes in the active-passive topology (for example, /oracle/mds).

You do not need the shared storage during installation, but you need it when you deploy your applications. You specify the path to the MDS directory during application deployment.

Figures of Active-Passive Topologies

Figure 6-3 shows a diagram of an active-passive topology with the Oracle Application Server Oracle home installed on the shared storage. The Oracle home contains both Oracle HTTP Server and WebCenter Suite. Figure 6-4 shows a distributed active-passive topology, where Oracle HTTP Server, WebCenter Framework, and Oracle Content DB are installed on different Oracle home. You can use this topology if your nodes are separated into web and application tiers, and the nodes in the tiers are separated by firewalls.

The MDS directory is required to be configured on the shared storage.

Figure 6-3 Active-Passive Topology with Oracle HTTP Server and WebCenter Suite in the Same Oracle Home

Collocated Active-Passive Topology
Description of "Figure 6-3 Active-Passive Topology with Oracle HTTP Server and WebCenter Suite in the Same Oracle Home"

Figure 6-4 Active-Passive Topology with Components in Separate Oracle Homes

Distributed Active-Passive Topology
Description of "Figure 6-4 Active-Passive Topology with Components in Separate Oracle Homes"

6.4.2 Overview of Installation Steps for OracleAS Cold Failover Cluster

Follow the steps in Table 6-3 to create the OracleAS Cold Failover Cluster configuration. If you are installing Oracle HTTP Server and OC4J in the same Oracle home (Figure 6-3), perform the steps on the hardware cluster. If you are installing Oracle HTTP Server and OC4J in separate Oracle homes (Figure 6-4), perform each step on both hardware clusters.

Table 6-3 Overview of Installation Steps for OracleAS Cold Failover Cluster


Step Description

1.

Perform Preinstallation Steps


Preinstallation tasks, described in Section 6.4.3 include:

2.

Install a High Availability Oracle Database


Install a high availability database for Oracle Content DB.

3.

Set VIRTUAL_HOST_NAME Environment Variable


Set the VIRTUAL_HOST_NAME variable to the virtual hostname.

4.

Install Oracle Application Server on the Shared Disk


In this step, you run the installer from either node of the hardware cluster to install Oracle Content DB, Oracle WebCenter Framework, and Oracle HTTP Server on the shared disk.

5.

(optional) Configure the Oracle Application Server Instance for SSL


If you want the Oracle Application Server instance to use SSL, enable SSL in the Oracle Application Server installation.

6.

(optional) Create a File System on the Shared Disk for OracleAS JMS File-Based Persistence


If you are using OracleAS JMS, create a file system on the shared disk.


6.4.3 Preinstallation Steps for OracleAS Cold Failover Cluster

Before installing Oracle Application Server in an OracleAS Cold Failover Cluster, perform these procedures:

Note:

In addition to the requirements listed in this chapter, ensure that you meet the requirements described in Section 6.2, "Requirements for High Availability Configurations".

6.4.3.1 Map the Virtual Hostname and Virtual IP Address

Each node in an OracleAS Cold Failover Cluster configuration is associated with its own physical IP address. In addition, the active node in the cluster is associated with a virtual hostname and virtual IP address. This allows clients to access the OracleAS Cold Failover Cluster using the virtual hostname.

Virtual hostnames and virtual IP addresses are any valid hostname and IP address in the context of the subnet containing the hardware cluster.

Note:

Map the virtual hostname and virtual IP address only to the active node. Do not map the virtual hostname and IP address to both active and passive nodes at the same time. When you failover, only then map the virtual hostname and IP address to the passive node, which is now the active node.

Note:

Before attempting to complete this procedure, ask the system or network administrator to review all the steps required. The procedure will reconfigure the network settings on the cluster nodes and may vary with differing network implementations.

The following example configures a virtual hostname called vhost.mydomain.com, with a virtual IP of 138.1.12.191:

  1. Register the virtual hostname and IP address with DNS for the network.

    For example, register the vhost.mydomain.com/138.1.12.191 pair with DNS.

  2. Add the following line to the /etc/hosts file on the active node:

    ip_address hostname.domain hostname
    
    

    For example:

    138.1.12.191   vhost.mydomain.com   vhost
    
    
  3. Determine the primary public network interface.

    The primary public network interface for Ethernet encapsulation is typically lan0 on HP-UX PA RISC.

    /usr/bin/netstat -i
    
    

    Using this command, search for a network interface that has an Address value of the physical hostname of the node.

  4. Find an available index number for the primary public network interface.

    Using the same command in step 3, determine an available index number for an additionl IP address to the primary public network interface.

    For example, on HP-UX, if the following is the output of the /usr/bin/netstat -i command and lan0 is determined to be the primary public interface in step 3, then lan0:2 is available for an additional IP address.

    Name       Mtu      Network        Address            Ipkts       Opkts
    lan0:1     1500     datacenter1    www2.mydomain.com  1050265     734793
    lan1*      1500     none           none               0           0
    lan0       1500     datacenter1    www1.mydomain.com  39783928    41833023
    lo0        4136     loopback       localhost          1226188     1226196
    
    
  5. Add the virtual IP address to the primary public network interface by running the following command, as the root user, using the available index number from step 4:

    
    
    /usr/sbin/ifconfig primary_public_interface:available_index ip_address
    
    

    For example, enter the following command if lan0:2 is available:

    /usr/sbin/ifconfig lan0:2 138.1.12.191
    
    

    Note:

    You must use the same NETMASK and BROADCAST values for this interface as those used for the primary public network interface (lan0 in this example). Modify the ifconfig commands in this step to include the appropiate netmask and broadcast options.
  6. Check that the virtual IP address is configured correctly:

    1. Use the instructions listed in step 3 to confirm the new entry for the primary_public_interface:available_index entry created in step 5.

    2. Try to connect to the node using the virtual hostname and virtual IP address from another node. For example, entering both of the following commands from a different node should provide a login to the node you configured in this procedure:

      telnet hostname.domain
      telnet ip_address
      
      

      For example, enter:

      telnet vhost.mydomain.com
      telnet 138.1.12.191
      
      

On Failover

If the active node fails, then the passive node takes over. If you do not have a clusterware agent to map the virtual IP from the failed node to the passive node, then you have to do it manually. You have to remove the virtual IP mapping from the failed node, and map it to the passive node.

  1. On the failed node, remove the virtual IP address by running the following command as the root user:

    /usr/sbin/ifconfig configured_interface down
    
    

    For example, enter the following command if lan0:2 is configured with the virtual IP address:

    /usr/sbin/ifconfig lan0:2 down
    
    

    Note:

    Use the commands in step 3 of the previous procedure to confirm that the virtual IP address has been removed.
  2. On the passive node, add the virtual IP address.

    On the passive node, follow steps 2 to 6 of the previous procedure to add and confirm the virtual IP address on the passive node.

6.4.3.2 Set Up a File System That Can Be Mounted from Both Nodes

Although the hardware cluster has shared storage, you need to create a file system on this shared storage such that both nodes of the OracleAS Cold Failover Cluster can mount this file system. You will use this file system for the following directories:

  • Oracle home directory for the Oracle Application Server instance

  • The oraInventory directory

For disk space requirements, see Section 2.2, "System Requirements".

If you are running a volume manager on the cluster to manage the shared storage, refer to the volume manager documentation for steps to create a volume. Once a volume is created, you can create the file system on that volume.

If you do not have a volume manager, you can create a file system on the shared disk directly. Ensure that the hardware vendor supports this, that the file system can be mounted from either node of the OracleAS Cold Failover Cluster, and that the file system is repairable from either node if a node fails.

To check that the file system can be mounted from either node, do the following steps:

  1. Set up and mount the file system from node 1.

  2. Unmount the file system from node 1.

  3. Mount the file system from node 2 using the same mount point that you used in step 1.

  4. Unmount it from node 2, and mount it on node 1, because you will be running the installer from node 1.

Note:

Only one node of the OracleAS Cold Failover Cluster should mount the file system at any given time. File system configuration files on all nodes of the cluster should not include an entry for the automatic mount of the file system upon a node restart or execution of a global mount command. For example, on UNIX platforms, do not include an entry for this file system in /etc/fstab file.

6.4.4 OracleAS Cold Failover Cluster: Details of Installation Steps

This section lists the steps for installing OracleAS Cold Failover Cluster.

If you are installing Oracle HTTP Server and OC4J in separate Oracle homes, you need to perform each of these steps on both clusters.


Step 1   Perform Preinstallation Steps

Perform the preinstallation steps listed in Section 6.4.3, "Preinstallation Steps for OracleAS Cold Failover Cluster".

Step 2   Install a High Availability Oracle Database

You need a database for Oracle Content DB. In high availability topologies, you should use high availability databases, such as cold failover cluster database or Real Application Clusters database.

See Section 2.4, "Requirements for Oracle Content Database" for more information on database requirements for Oracle Content DB.

Step 3   Set VIRTUAL_HOST_NAME Environment Variable

Set the VIRTUAL_HOST_NAME environment variable to the virtual hostname on either node of the hardware cluster. You will perform the install from this node onto the shared disk in the next step. To find out more about how to set environment variables, see Section 2.9, "Environment Variables".

Step 4   Install Oracle Application Server on the Shared Disk

Install Oracle Application Server on the shared disk of the hardware cluster from the node where you set the VIRTUAL_HOST_NAME environment variable.

  • For OracleAS Cold Failover Cluster with Oracle HTTP Server and WebCenter Suite in the Same Oracle Home

    Perform a basic installation as described in Chapter 4, "Basic Installation".

  • For OracleAS Cold Failover Cluster with Components in Separate Oracle Homes

    If you are installing on the hardware cluster where you want to run Oracle HTTP Server, follow the steps in Section 5.2.5, "Installing Oracle HTTP Server". During installation, perform the following actions:

    • If you want to route all requests to OC4J through the Oracle HTTP Server, select Configure this Oracle HTTP Server instance to be part of an Oracle Application Server cluster on the "Cluster Topology Configuration" screen. Specify the IP Address and Port for the multicast address shared by all the nodes in the cluster.

    • If you do not want to route all requests to OC4J through the Oracle HTTP Server, deselect Configure this Oracle HTTP Server instance to be part of an Oracle Application Server cluster on the "Cluster Topology Configuration" screen.

    If you are installing on the hardware cluster where you want to run Oracle WebCenter Framework and Oracle Content DB, you need to perform the following two installations:

    During the installation of Oracle WebCenter Framework, perform the following actions:

    • On the "Administration Settings" screen, select Start Oracle Enterprise Manager 10g ASControl in this instance if you want to configure Application Server Control for administering the OC4J instance. Otherwise, deselect this option. You should select this option in only one Oracle home in the cluster topology.

    • If you want to route all requests to OC4J through the Oracle HTTP Server, select Configure this instance to be part of an Oracle Application Server cluster topology on the "Cluster Topology Configuration" screen. Specify the IP Address and Port for the multicast address shared by all the nodes in the cluster. Select Access this OC4J Instance from a separate Oracle HTTP Server.

    • If you do not want to route all requests to OC4J through the Oracle HTTP Server, deselect Configure this instance to be part of an Oracle Application Server cluster topology on the "Cluster Topology Configuration" screen.

    During the installation of Oracle Content DB, perform the following actions:

    • On the "Administration Settings" screen, select Start Oracle Enterprise Manager 10g ASControl in this instance if you want to configure Application Server Control for administering the OC4J instance. Otherwise, deselect this option. You should select this option in only one Oracle home in the cluster topology.

    • If you want to route all requests to OC4J through the Oracle HTTP Server, select Configure this instance to be part of an Oracle Application Server cluster topology on the "Cluster Topology Configuration" screen. Specify the IP Address and Port for the multicast address shared by all the nodes in the cluster.

    • If you do not want to route all requests to OC4J through the Oracle HTTP Server, deselect Configure this instance to be part of an Oracle Application Server cluster topology on the "Cluster Topology Configuration" screen.

Step 5   Disable Oracle HTTP Server on Nodes with Oracle Content DB (Distributed Installations Only)

This step is required only if you performed a distributed installation in step 4.

Disable Oracle HTTP Server on the Oracle homes where you installed Oracle Content DB.

  1. Set the status of Oracle HTTP Server to disabled in the ORACLE_HOME/opmn/conf/opmn.xml file:

    <ias-component id="HTTP_Server" status="disabled">
    
    
  2. Stop and restart all components in the Oracle home.

    > ORACLE_HOME/bin/opmnctl stopall
    > ORACLE_HOME/bin/opmnctl startall
    
    

Step 6   (optional) Configure the Oracle Application Server Instance for SSL

This step is required only if you are using SSL.

Edit the ORACLE_HOME/Apache/Apache/conf/httpd.conf file on each node as follows:

  1. Add the following line to the httpd.conf to load the certheaders_module.

    LoadModule certheaders_module libexec/mod_certheaders.so
    
    

    Note:

    If you installed Oracle HTTP Server 2.0 using the Oracle Application Server Companion CD, add the following line to the httpd.conf file:
    LoadModule certheaders_module modules/mod_certheaders.so
    
    
  2. Add a virtual host directive.

    <VirtualHost *:7777>
    ServerName myworkplace.com
    Port 443
    ServerAdmin you@youraddress.com
    RewriteEngine On
    RewriteOptions inherit
    SimulateHTTPS On
    </VirtualHost>
    
    
  3. Restart Oracle HTTP Server on all the nodes.

Step 7   (optional) Create a File System on the Shared Disk for OracleAS JMS File-Based Persistence

If you are using OracleAS JMS with file-based persistence, create a file system on the shared disk for the OracleAS JMS queues, and mount this file system from node 1.

6.5 Creating an OracleAS Disaster Recovery Configuration

This section describes how to install Oracle Application Server in OracleAS Disaster Recovery configurations. OracleAS Disaster Recovery is one of the high availability environments supported by Oracle Application Server.

Contents of this section:

6.5.1 OracleAS Disaster Recovery: Introduction

Use the OracleAS Disaster Recovery environment when you want to have two physically separate sites in your environment. One site is the production site, and the other site is the standby site. The production site is active, while the standby site is passive; the standby site becomes active when the production site goes down.

OracleAS Disaster Recovery supports a number of basic topologies for the configuration of the Infrastructure and middle tier on production and standby sites. OracleAS Disaster Recovery supports these basic topologies:

  • Symmetrical topologies -- strict mirror of the production site with collocated Oracle Identity Management and OracleAS Metadata Repository Infrastructure

  • Asymmetrical topologies -- simple asymmetric standby topology with collocated Oracle Identity Management and OracleAS Metadata Repository Infrastructure

  • Separate OracleAS Metadata Repository for OracleAS Portal with collocated Oracle Identity Management and OracleAS Metadata Repository Infrastructure (the Departmental Topology)

  • Distributed Application OracleAS metadata Repositories with Non collocated Oracle Identity Management and OracleAS Metadata Repository Infrastructure

  • Redundant Multiple OracleAS 10.1.3.2.0 Home J2EE Topology

  • Redundant Single OracleAS 10.1.3.2.0 Home J2EE Topology Integrated with an Existing Oracle Identity Management 10.1.2.0.2 Topology

In a symmetric topology, each node in the standby site corresponds to a node in the production site. This includes the nodes running both OracleAS Infrastructure and middle tiers. In an asymmetric topology, the number of instances required on the standby site are fewer than the number on the production site and the number of instances required on the standby site must be the minimum set of instances required to run your site in the event of a switchover or failover operation. The last two supported topologies are particularly important in OracleAS Release 10.1.3.2.0. See the Oracle Application Server High Availability Guide for a detailed description of these topologies.

As a small variation to this environment, you can set up the OracleAS Infrastructure on the production site in an OracleAS Cold Failover Cluster environment. See Section 6.5.2.4, "If You Want to Use OracleAS Cold Failover Cluster on the Production Site (OracleAS 10.1.2.n.n only)" for details.

For these supported topologies, OracleAS Guard will be installed in every Oracle home on every system that is part of your production and standby topology configured for the OracleAS Disaster Recovery solution.

OracleAS Guard can be installed as a standalone install kit located on OracleAS Companion CD #2. See Section 6.5.4, "Installing the OracleAS 10g (10.1.3.2.0) Standalone Install of OracleAS Guard into Oracle Homes" for more information about when this standalone kit should be installed.

Figure 6-5 shows an example symmetric OracleAS Disaster Recovery environment. Each site has two nodes running middle tiers and a node running OracleAS Infrastructure.

Data Synchronization

For OracleAS Disaster Recovery to work, data between the production and standby sites must be synchronized so that failover can happen very quickly. Configuration changes done at the production site must be synchronized with the standby site.

You need to synchronize two types of data. The synchronization method depends on the type of data:

  • Use Oracle Data Guard to synchronize data in the OracleAS Metadata Repository databases on the production and standby sites. You can configure Oracle Data Guard to perform the synchronization.

  • Use the backup and recovery scripts to synchronize data outside of the database (such as data stored in configuration files).

See the Oracle Application Server High Availability Guide for details on how to use Oracle Data Guard and the backup and recovery scripts.

Figure 6-5 OracleAS Disaster Recovery Environment

Description of Figure 6-5 follows
Description of "Figure 6-5 OracleAS Disaster Recovery Environment"

6.5.2 Setting up the OracleAS Disaster Recovery Environment

Before you can install Oracle Application Server in an OracleAS Disaster Recovery environment, you have to perform these steps:

6.5.2.1 Ensure Nodes Are Identical at the Operating System Level

Ensure that the nodes are identical with respect to the following items:

  • The nodes are running the same version of the operating system.

  • The nodes have the same operating system patches and packages.

  • You can install Oracle Application Server in the same directory path on all nodes.

6.5.2.2 Set Up staticports.ini File

The same component must use the same port number on the production and standby sites. For example, if Oracle HTTP Server is using port 80 on the production site, it must also use port 80 on the standby site. To ensure this is the case, create a staticports.ini file for use during installation. This file enables you to specify port numbers for each component. See Section 2.6.3, "Using Custom Port Numbers (the "Static Ports" Feature)" for details.

6.5.2.3 Set Up Identical Hostnames on Both Production and Standby Sites

The names of the corresponding nodes on the production and standby sites must be identical, so that when you synchronize data between the sites, you do not have to edit the data to fix the hostnames.

For the Infrastructure Nodes

For the node running the infrastructure, set up a virtual name. To do this, specify an alias for the node in the /etc/hosts file.

For example, on the infrastructure node on the production site, the following line in the hosts file sets the alias to asinfra:

138.1.2.111   prodinfra   asinfra

On the standby site, the following line sets the node's alias to asinfra.

213.2.2.110   standbyinfra   asinfra

When you install OracleAS Infrastructure on the production and standby sites, you specify this alias (asinfra) in the Specify Virtual Hostname screen. The configuration data will then contain this alias for the infrastructure nodes.

For the Middle-Tier Nodes

For the nodes running the middle tiers, you cannot set up aliases like you did for the infrastructure nodes because the installer does not display the Specify Virtual Hostname screen for middle-tier installations. When installing middle tiers, the installer determines the hostname automatically by calling the gethostname() function. You want to be sure that for each middle-tier node on the production site, the corresponding node on the standby site returns the same hostname.

Note:

For 10.1.3.x midtier installations onwards, you may also set the VIRTUAL_HOST_NAME environment variable appropriately before the installation for installer to pick up the virtualname correctly for the install. Hostnames may be allowed to be the same or different across primary and standby in this case.

To do this, set up a local, or internal, hostname, which could be different from the public, or external, hostname. You can change the names of the nodes on the standby site to match the names of the corresponding nodes on the production site, or you can change the names of the nodes on both production and standby sites to be the same. This depends on other applications that you might be running on the nodes, and whether changing the node name will affect those applications.

  1. Enable the other nodes in the OracleAS Disaster Recovery environment to be able to resolve the node using the new local hostname. You can do this in one of two ways:

    Method 1: Edit the /etc/hosts file on each node on both sites. This method does not involve configuring DNS servers, but you have to maintain the hosts file on each node in the OracleAS Disaster Recovery environment. For example, if an IP address changes, you have to update the files on all the nodes, and restart the nodes.

    Method 1 Details

    1. On each node on the production site, include these lines in the /etc/hosts file. The IP addresses resolve to nodes on the production site.

      Note:

      In the hosts file, be sure that the line that identifies the current node comes immediately after the localhost definition (the line with the 127.0.0.1 address).
      127.0.0.1    localhost
      138.1.2.333  asmid1.oracle.com   asmid1
      138.1.2.444  asmid2.oracle.com   asmid2
      138.1.2.111  asinfra.oracle.com  asinfra
      
      
    2. On each node on the standby site, include these lines in the hosts file. The IP addresses resolve to nodes on the standby site.

      Note:

      In the hosts file, be sure that the line that identifies the current node comes immediately after the localhost definition (the line with the 127.0.0.1 address).
      127.0.0.1    localhost
      213.2.2.330  asmid1.oracle.com   asmid1
      213.2.2.331  asmid2.oracle.com   asmid2
      213.2.2.110  asinfra.oracle.com  asinfra
      
      
    3. Ensure that the "hosts:" line in the /etc/nsswitch.conf file has "files" as the first item:

      hosts:   files nis dns
      
      

      The entry specifies the ordering of the name resolution. If another method is listed first, then the node will use the other method to resolve the hostname.

    Note:

    Restart the nodes after editing these files.

    Method 2: Set up separate internal DNS servers for the production and standby sites. This configuration allows nodes on each site (production or standby) to resolve hostnames within the site. Above the internal DNS servers are the corporate, or external, DNS servers. The internal DNS servers forward non-authoritative requests to the external DNS servers. The external DNS servers do not know about the existence of the internal DNS servers. See Figure 6-6.

    Figure 6-6 Method 2: Using DNS Servers

    Description of Figure 6-6 follows
    Description of "Figure 6-6 Method 2: Using DNS Servers"

    Method 2 Details

    1. Make sure the external DNS names are defined in the external DNS zone. Example:

      prodmid1.us.oracle.com     IN  A  138.1.2.333
      prodmid2.us.oracle.com     IN  A  138.1.2.444
      prodinf.us.oracle.com      IN  A  138.1.2.111
      standbymid1.us.oracle.com  IN  A  213.2.2.330
      standbymid2.us.oracle.com  IN  A  213.2.2.331
      standbyinf.us.oracle.com   IN  A  213.2.2.110
      
      
    2. At the production site, create a new zone at the production site using a domain name different from your external domain name. To do this, populate the zone data files with entries for each node in the OracleAS Disaster Recovery environment.

      For the infrastructure node, use the virtual name or alias.

      For the middle-tier nodes, use the node name (the value in /etc/nodename).

      The following example uses "asha" as the domain name for the new zone.

      asmid1.asha    IN  A  138.1.2.333
      asmid2.asha    IN  A  138.1.2.444
      asinfra.asha   IN  A  138.1.2.111
      
      

      Do the same for the standby site. Use the same domain name that you used for the production site.

      asmid1.asha    IN  A  213.2.2.330
      asmid1.asha    IN  A  213.2.2.331
      asinfra.asha   IN  A  213.2.2.110
      
      
    3. Configure the DNS resolver to point to the internal DNS servers instead of the external DNS server.

      In the /etc/resolv.conf file for each node on the production site, replace the existing name server IP address with the IP address of the internal DNS server for the production site.

      Do the same for the nodes on the standby site, but use the IP address of the internal DNS server for the standby site.

    4. Create a separate entry for Oracle Data Guard in the internal DNS servers. This entry is used by Oracle Data Guard to ship redo data to the database on the standby site.

      In the next example, the "remote_infra" entry points to the infrastructure node on the standby site. This name is used by the TNS entries on both the production and standby sites so that if a switchover occurs, the entry does not have to be changed.

      Figure 6-7 Entry for Oracle Data Guard in the Internal DNS Servers

      Description of Figure 6-7 follows
      Description of "Figure 6-7 Entry for Oracle Data Guard in the Internal DNS Servers"

      On the production site, the DNS entries look like this:

      asmid1.asha        IN  A  138.1.2.333
      asmid2.asha        IN  A  138.1.2.444
      asinfra.asha       IN  A  138.1.2.111
      remote_infra.asha  IN  A  213.2.2.110
      
      

      On the standby site, the DNS entries look like this:

      asmid1.asha        IN  A  213.2.2.330
      asmid2.asha        IN  A  213.2.2.331
      asinfra.asha       IN  A  213.2.2.110
      remote_infra.asha  IN  A  138.1.2.111
      
      

Verifying that the Nodes Resolve the Hostnames Correctly

After making the changes and restarting the nodes, check that the nodes resolve the hostnames properly by running the following commands:

  • On the middle-tier nodes on both sites, run the hostname command. This should return the internal hostname. For example, the command should return "asmid1" if you run it on prodmid1 and standbymid1.

    prompt> hostname
    asmid1
    
    
  • On each node, ping the other nodes in the environment using the internal hostname as well as the external hostname. The command should be successful. For example, from the first midtier node, prodmid1, run the following commands:

    prompt>  ping the production infrastructure node
    PING prodinfra: 56 data byes
    64 bytes from prodinfra.oracle.com (138.1.2.111): icmp_seq=0. time=0. ms
    ^C
    
    prompt>  ping the production infrastructure node
    PING iasinfra: 56 data byes
    64 bytes from iasinfra.oracle.com (138.1.2.111): icmp_seq=0. time=0. ms
    ^C
    
    prompt>  ping the second production midtier node
    PING iasmid2: 56 data byes
    64 bytes from iasmid2.oracle.com (138.1.2.444): icmp_seq=0. time=0. ms
    ^C
    
    prompt>  ping the second production midtier node
    PING prodmid2: 56 data byes
    64 bytes from prodmid2.oracle.com (138.1.2.444): icmp_seq=0. time=0. ms
    ^C
    
    prompt>  ping the first standby midtier node
    PING standbymid1: 56 data byes
    64 bytes from standbymid1.oracle.com (213.2.2.330): icmp_seq=0. time=0. ms
    ^C
    
    

6.5.2.4 If You Want to Use OracleAS Cold Failover Cluster on the Production Site (OracleAS 10.1.2.n.n only)

Note:

You must perform this installation in an OracleAS Release 10.1.2.n.n environment, where n.n represents 0.0 or higher. This information is presented here for informative purposes only.

On the production site of a OracleAS Disaster Recovery system, you can set up the OracleAS Infrastructure to run in a OracleAS Cold Failover Cluster configuration. In this case, you have two nodes in a hardware cluster, and you install the OracleAS Infrastructure on a shared disk. See Chapter 11, "Installing in High Availability Environments: OracleAS Cold Failover Cluster" in the Oracle Application Server Installation Guide 10g Release 2 (10.1.2) Documentation set for details.

Figure 6-8 Infrastructure in an OracleAS Cold Failover Cluster Configuration

Description of Figure 6-8 follows
Description of "Figure 6-8 Infrastructure in an OracleAS Cold Failover Cluster Configuration"

To set up OracleAS Cold Failover Cluster in this environment, use the virtual IP address (instead of the physical IP address) for asinfra.asha on the production site. The following example assumes 138.1.2.120 is the virtual IP address.

asmid1.asha          IN  A  138.1.2.333
asmid2.asha          IN  A  138.1.2.444
asinfra.asha         IN  A  138.1.2.120         this is a virtual IP address
remote_infra.asha    IN  A  213.2.2.110

On the standby site, you still use the physical IP address for asinfra.asha, but the remote_infra.asha uses the virtual IP address.

asmid1.asha          IN  A  213.2.2.330
asmid2.asha          IN  A  213.2.2.331
asinfra.asha         IN  A  213.2.2.110         physical IP address
remote_infra.asha    IN  A  138.1.2.120         virtual IP address

6.5.3 Installing Oracle Application Server in an OracleAS Disaster Recovery Environment

For OracleAS Release 10.1.3.2.0, you can only install middle tiers on the production and standby sites.

Install Oracle Application Server as follows:

Note:

For all of the installations, be sure to use staticports.ini to specify port numbers for the components. See Section 6.5.2.2, "Set Up staticports.ini File".

Install Middle Tiers (OracleAS Release 10.1.3.2.0 only)

  1. Install middle tiers on the production site.

  2. Install middle tiers on the standby site.

Install OracleAS Infrastructure and Middle Tiers (Release 10.1.2.n.n only)

Note:

You must perform this installation in an OracleAS Release 10.1.2.n.n environment, where n.n represents 0.0 or higher. This information is presented here for informative purposes only.
  1. Install OracleAS Infrastructure on the production site.

  2. Install OracleAS Infrastructure on the standby site.

  3. Start the OracleAS Infrastructure in each site before installing the middle tiers for that site.

  4. Install middle tiers on the production site.

  5. Install middle tiers on the standby site.

6.5.3.1 Installing the OracleAS Infrastructure (OracleAS Release 10.1.2.n.n Only)

Note:

You must perform this installation in an OracleAS Release 10.1.2.n.n environment, where n.n represents 0.0 or higher. This information is presented here for informative purposes only.

In an OracleAS Release 10.1.2.0.0 environment, you must install the Oracle Identity Management and the OracleAS Metadata Repository components of OracleAS Infrastructure on the same node. You cannot distribute the components over multiple nodes. In an OracleAS Release 10.1.2.0.2 environment, you can distribute the components over multiple nodes.

The installation steps are similar to that for OracleAS Cold Failover Cluster. See Section 11.3, "Installing an OracleAS Cold Failover Cluster (Infrastructure) Configuration" in the Oracle Application Server Installation Guide 10g Release 2 (10.1.2) Documentation set for the screen sequence.

Note the following points:

  • Select Configuration Options screen: be sure you select High Availability and Replication. See Table 11–5, step 2.

  • Specify Virtual Hostname screen: enter an alias as the virtual address (for example, asinfra.oracle.com). See Table 11–5, step 6.

6.5.3.2 Installing Middle Tiers (OracleAS Release 10.1.3.2.0 and 10.1.2.n.n)

Depending on your configuration, you can install OracleAS 10.1.3.2.0 middle tiers or OracleAS 10.1.2.n.n middle tiers, where n.n represents 0.0 or higher.

OracleAS Release 10.1.3.2.0

On OracleAS release 10.1.3.2.0, you can install any type of middle tier that you like:

For installing Oracle WebCenter Framework with Oracle HTTP Server, see Section 5.2.2, "Installing Oracle WebCenter Framework with Oracle HTTP Server".

For installing Oracle WebCenter Framework, see Section 5.2.4, "Installing Oracle WebCenter Framework".

For installing Oracle HTTP Server, see Section 5.2.5, "Installing Oracle HTTP Server".

OracleAS Release 10.1.2.n.n

Note:

You must perform this installation in an OracleAS Release 10.1.2.n.n environment, where n.n represents 0.0 or higher. This information is presented here for informative purposes only.

On OracleAS Release 10.1.2.n.n, you can install any type of middle tier that you like:

For installing J2EE and Web Cache, see Section 7.9 "Installing J2EE and Web Cache in a Database-Based Farm Repository and with Oracle Identity Management Access" in the Oracle Application Server Installation Guide for 10g Release 2 (10.1.2).

For installing Portal and Wireless or Business Intelligence and Forms, see Section 7.13, "Installing Portal and Wireless or Business Intelligence and Forms".

Note the following points on OracleAS 10.1.2.n.n:

When the installer prompts you to register with Oracle Internet Directory, and asks you for the Oracle Internet Directory hostname, enter the alias of the node running OracleAS Infrastructure (for example, asinfra.oracle.com).

6.5.4 Installing the OracleAS 10g (10.1.3.2.0) Standalone Install of OracleAS Guard into Oracle Homes

OracleAS 10g (10.1.3.2.0) standalone install of OracleAS Guard is located on Companion CD Disk 2. This standalone install of OracleAS Guard can be installed in the following environments:

  • In its own home in the case when you are cloning an instance or topology to a new standby system (see the section on standby site cloning in Oracle Application Server High Availability Guide for more information).

  • Oracle database server home for an OracleAS Metadata Repository configuration created using OracleAS Metadata Repository Creation Assistant.

  • OracleAS Disaster Recovery full site upgrade from OracleAS 10g (9.0.4) to OracleAS 10g (10.1.3.2.0) (see the chapter on OracleAS Disaster Recovery site upgrade procedure in Oracle Application Server High Availability Guide for more information).

  • OracleAS Guard patch upgrade from OracleAS 10g (10.1.2.0.0) to OracleAS 10g (10.1.2.0.2) (see Section 6.5.5, "Upgrading OracleAS Guard Release 10.1.2.n.n with Release 10.1.3.2.0" for more information).

If this is an upgrade installation of OracleAS Guard, make a copy of your dsa.conf configuration file to save your current settings for your OracleAS Guard environment. After running the OracleAS 10g (10.1.3.2.0) standalone install kit of OracleAS Guard, you can restore your saved dsa.conf configuration file with your settings to continue using the same settings for the upgraded OracleAS Guard environment.

To run the OracleAS 10g (10.1.3.2.0) standalone install kit of OracleAS Guard, run the kit in the following directory path:

/Disk2/asg/install/runInstaller

Choose the type of install that you want. Choose Typical for most installations. Choose Custom or Reinstall for upgrading from an older release of OracleAS Guard to the current release.

Enter the oc4jadmin account password to continue the installation.

6.5.5 Upgrading OracleAS Guard Release 10.1.2.n.n with Release 10.1.3.2.0

If you already have an OracleAS Disaster Recovery environment set up using OracleAS Guard Release 10.1.2.n.n, where n.n represents 0.0 or higher, you can upgrade OracleAS Guard in your environment to take advantage of new features and support for the topologies described in Section 6.5.1, "OracleAS Disaster Recovery: Introduction". To upgrade your OracleAS Disaster Recovery environment, follow these basic steps:

  1. Stop the OracleAS Guard server in all OracleAS 10.1.2.n.n Oracle homes on both production and standby sites using the following opmnctl command:

    <ORACLE_HOME>/opmn/bin/opmnctl stopall
    
    
  2. Install the OracleAS 10g (10.1.3.2.0) standalone install of OracleAS Guard into each Oracle home on the production and standby sites.

    If multiple Oracle homes exist on the same system, ensure that different ports are configured for each of the OracleAS Guard servers in this configuration file.

    Because this is an upgrade installation of OracleAS Guard, make a copy of your dsa.conf configuration file to save your current settings for your OracleAS Guard environment. After running the OracleAS 10g (10.1.3.2.0) standalone install kit of OracleAS Guard, you can restore your saved dsa.conf configuration file with your settings to continue using the same settings for the upgraded OracleAS Guard environment.

    <ORACLE_HOME>/dsa/dsa.conf
    
    
  3. Start the OracleAS Guard server in all OracleAS 10.1.3.2.0 Oracle homes on both production and standby sites using the following opmnctl command:

    <ORACLE_HOME>/opmn/bin/opmnctl startall
    <ORACLE_HOME>/opmn/bin/opmnctl startproc ias-component=ASG
    
    

6.5.6 What to Read Next

For information on how to manage your OracleAS Disaster Recovery environment, such as setting up Oracle Data Guard and configuring the OracleAS Metadata Repository database, see the Oracle Application Server High Availability Guide.