Skip Headers
Oracle® Communications Services Gatekeeper Concepts Guide
Release 5.0

Part Number E16613-02
Go to Documentation Home
Home
Go to Book List
Book List
Go to Table of Contents
Contents
Go to Feedback page
Contact Us

Go to previous page
Previous
Go to next page
Next
View PDF

8 Redundancy, Load Balancing, and High Availability

This chapter explains the redundancy, load balancing, and high availability functionality in Oracle Communications Services Gatekeeper. Services Gatekeeper uses both software and hardware components to support these important capabilities.

Services Gatekeeper's high-availability mechanisms are supported by the clustering mechanisms made available by Oracle WebLogic Server. For general information about Oracle WebLogic Server and clustering, see Oracle® Fusion Middleware Using Clusters for Oracle WebLogic Server at:

http://download.oracle.com/docs/cd/E15523_01/web.1111/e13709/toc.htm

Tiering

For both high-availability and security reasons, Services Gatekeeper is split into two tiers: the Access Tier and the Network Tier.

Native SMPP and native UCP operate entirely in the Network Tier. In these cases access to applications is performed by a Server Service in the Network Tier.

Each tier consists of at least one cluster, with at least two server instances per cluster, and all server instances run in active mode, independently of each other. The servers in all clusters are, in the context of Oracle WebLogic Server, Managed Servers. Together the clusters make up a single WebLogic Server administrative domain, controlled through an Administration Server.

Figure 8-1 Sample Production Domain

Surrounding text describes Figure 8-1 .

Communication between the Access Tier and the Network Tier takes place using Java RMI. Application requests are load balanced between the Access Tier and the Network Tier and failover mechanisms are present between the two. See "Traffic Management Inside Services Gatekeeper" for more information on these mechanisms in application-initiated and network-triggered traffic flows.

There is an additional tier containing the database. Within the cluster, data is made highly available using a cluster-aware storage service which ensures that all state data is made available across all Network Tier instances.

Traffic Management Inside Services Gatekeeper

Potential failure is possible at many stages in traffic workflow in Gatekeeper. The following sections detail, tier by tier, how Services Gatekeeper deals with problems that might arise in both application-initiated and network-triggered traffic.

Application-Initiated Traffic

Application-initiated traffic consists of all requests that travel from applications through Services Gatekeeper to underlying network nodes.

Figure 8-2 below follows the worst-case scenario for application-initiated traffic as it passes through Services Gatekeeper and the failover mechanisms that attempt to keep the request alive.

Figure 8-2 Failover Mechanisms in Application-Initiated Traffic

Surrounding text describes Figure 8-2 .
  1. The application sends a request to Services Gatekeeper. In a production environment, this request is routed through a hardware load balancer that is usually protocol-aware. If the request towards the initial Access Tier server fails (1.1 in Figure 8-2), either a time out or a failure is reported. The load-balancer, or the application itself, is responsible for retrying the request.

  2. The request is retried on a second server in the cluster (1.2 in Figure 8-2) and it succeeds. That server then attempts to send the request on to the Network Tier.

  3. The request either fails to reach the Network Tier or fails during the process of marshalling/unmarshalling the request as it travels to the Network Tier server (2.1 in Figure 8-2).

  4. A failover mechanism in the Access Tier sends the request to a different server in the Network Tier cluster and it succeeds (2.2 in Figure 8-2). That server then attempts to send the request on to the network node.

  5. The request is sent to a plug-in in the Network Tier that is unavailable (3.1 in Figure 8-2). An interceptor from the stack retries the remaining eligible plug-ins in the same server and succeeds (3.2 in Figure 8-2).

  6. The attempt to send the request to the telecom network node fails (4.1 in Figure 8-2).

  7. If a redundant pair of network nodes exists, the request is forwarded to the redundant node (4.2 in Figure 8-2). If this request fails, the failure is reported to the application.

    Note:

    In addition to the mechanisms described above, Services Gatekeeper also allows the creation of multiple instances of a single SMPP plug-in type, with multiple binds, which can set up redundant connections to one or more network nodes. Such mechanisms can also increase throughput, and help optimize traffic to SMSCs with small transport windows.

Network-triggered Traffic

Network-triggered traffic can consist of the following:

  • Requests that contain a payload, such as terminal location or an SMS

  • Acknowledgements from the underlying network node that an application-initiated request has been processed by the network node itself. A typical example might indicate that an SMS has reached the SMSC. From an application's perspective, this is normally processed as part of a synchronous request, although it may be asynchronous from the point of view of the network.

  • Acknowledgements from the underlying network node that the request has been processed by the destination end-user terminal; for example, an SMS delivery receipt indicating that the SMS has been delivered to the end-user terminal. From an application's perspective, this is normally handled as an incoming notification.

For network-triggered traffic, Services Gatekeeper relies on internal mechanisms in concert with the capabilities of the telecom network node or other external artifacts such as load-balancers with failover capabilities to do failover.

Some network nodes can handle the registration of multiple callback interfaces. In such cases, Services Gatekeeper registers one primary and one secondary callback interface. If the node is unable to send a request to the network plug-in registered as the primary callback interface, it is responsible for retrying the request by sending it to the plug-in that is registered as the secondary callback interface. This plug-in resides in another Network Tier instance. The plug-ins themselves are responsible for communicating with each other and making sure that both callback interfaces are registered. See "When the Network Node Supports Primary and Secondary Notification" for more information.

In the case of communication services using SMPP, all Services Gatekeeper plug-ins can function equally as receivers for any transmission from the network node.

Finally, for HTTP-based protocols, such as MM7, MLP, and PAP, Services Gatekeeper relies on an HTTP load balancer with failover functionality between the telecom network node and Services Gatekeeper. See "When the Network Node Supports Only Single Notification" for more information.

If a telecom network protocol does not support load balancing and high availability, a single point of failure is unavoidable. In this case, all traffic associated with a specific application is routed through the same Network Tier server and each plug-in has one single connection to one telecom network node.

The worst-case scenario for network-triggered traffic for medium life span notifications using a network node that supports primary and secondary callback interfaces is described in Figure 8-3.

Figure 8-3 Failover Mechanisms in Network-Triggered Traffic

Surrounding text describes Figure 8-3 .
  1. A telecom network node sends a request to the Services Gatekeeper network plug-in that has been registered as the primary. It fails (1.1 in Figure 8-3) due to either a communication or server failure.

  2. The telecom network node resends the request, this time to the plug-in that is registered as the secondary callback interface (1.2 in Figure 8-3). This plug-in is in a different server instance within the Network Tier cluster. It succeeds.

  3. The Network Tier attempts to send the message to the callback mechanism in the Access Tier. It fails (2.1 in Figure 8-3).

  4. If the request fails to reach the Access Tier, or failure occurs during the marshalling/unmarshalling process (2.1 in Figure 8-3), the Network Tier retries, targeting another server in the Access Tier. It succeeds (2.2 in Figure 8-3).

    Note:

    If, however, the failure occurs after processing has begun in the Access Tier, failover does not occur and an error is reported to the network node.
  5. The callback mechanism in the Access Tier attempts to send the request to the application (3.1 in Figure 8-3). If the application is unreachable or does not respond, the request is considered as having failed, and an error is reported to the network node.

Registering Notifications with Network Nodes

Before applications can receive network-triggered traffic, or notifications, they must register their interest in doing so with Services Gatekeeper, either by sending a request or having the operator set the notification up using OAM methods. In turn, these notifications must be registered with the underlying network node that will be supplying them. The form of this registration is dependent on the capabilities of that node.

If registration for notifications is supported by the underlying network node protocol, the communication service's network plug-in is responsible for performing it, whether the registration is the result of an application-initiated registration request or an online provisioning step in Services Gatekeeper. For example, all OSA/Parlay Gateway interfaces support such registration for notifications.

Note:

Some network protocols support some, but not all registration types. For example, in MM7 an application can register to receive notifications for delivery reports on messages sent from the application, but not to receive notifications on messages sent to the application from the network. In this case, registration for such notifications can be done as an off-line provisioning step in the MMSC.

Whether the plug-in sets up the notification in the network or it is done using OAM, Services Gatekeeper is responsible for correlating all network-triggered traffic with its corresponding application.

Notification Life Span

Notifications are placed into three categories, based on the expected life span of the notification. These categories determine the failover strategies used:

  • Short life span

    These notifications have an expected life span of a few seconds. Typically these are delivery acknowledgements for hand-off of the request to the network node, where the response to the request is reported asynchronously. For this category, a single plug-in, the originating one, is deemed sufficient to handle the response from the network node.

  • Medium life span

    These notifications have an expected life span of minutes up to a few days. Typically these are delivery acknowledgements for message delivery to an end-user terminal. For this category, the delivery notification criteria that have been registered are replicated to exactly one additional instance of the network protocol plug-in. The plug-in that receives the notification is responsible for registering a secondary notification with the network node, if possible.

  • Long life span

    These notifications have an expected life span of more than a a few days. Typically these are registrations for notifications for network-triggered SMS and MMS messages or calls that need to be handled by an application. For this category, the delivery notification criteria are replicated to all instances of the network plug-in. Each plug-in that receives the notification is responsible for registering an interface with the network node.

When the Network Node Supports Primary and Secondary Notification

Figure 8-4 illustrates how Services Gatekeeper registers both primary and secondary notifications with network nodes that support it. This capability must be supported by the network protocol in the abstract and in the implementation of the protocol as it exists in both the network node and the communication service's network plug-in.

Note:

The scenario assumes that the network node supports registration for notifications with overlapping criteria (primary/secondary).

Figure 8-4 Network Node Supports Primary/Secondary Notifications

Surrounding text describes Figure 8-4 .
  1. The request to register for notifications enters the network protocol plug-in from the application.

  2. The primary notification is registered with the telecom network node.

  3. The notification information is propagated to another instance of the network protocol plug-in.

  4. The secondary notification is registered with the telecom network node.

    Note:

    The concept of primary/secondary notification is not necessarily ordered. The most recently registered notification may, for example, be designated the primary notification.

When a network-triggered request that matches the criteria in a previously registered notification reaches the telecom network node, the node first tries the network plug-in that registered the primary notification. If that request fails, the network node has the responsibility of retrying, using the plug-in that registered the secondary notification. The secondary plug-in will have all necessary information to propagate the request through Services Gatekeeper and on to the correct application.

When the Network Node Supports Only Single Notification

Figure 8-5 illustrates the registration step in Services Gatekeeper if the underlying network node does not support primary/secondary notification registration.

Note:

The scenario assumes that the network node does not support registration for notifications with overlapping criteria. Only one notification for a given criteria is allowed.

Figure 8-5 Network Node Supports Only Single Notification

Surrounding text describes Figure 8-5 .
  1. The request to register for notifications enters the network protocol plug-in from the application.

  2. The notification is registered with the telecom network node.

  3. The notification information (matching criteria, target URL, etc.) is propagated to another instance of the network protocol plug-in. The plug-in makes the necessary arrangements to be able to receive notifications.

There are two possibilities for high-availability and failover support in this case:

  • All plug-ins can receive notifications from the network node. This is the case with SMPP, in which all plug-ins can function as receivers for any transmission from the network node.

  • A load balancer with failover support is introduced between the network protocol plug-in and the network node. This is the case with HTTP- based protocols, as in Figure 8-6.

    Note:

    Whether or not this is possible depends on the network protocol, because the load-balancer must be protocol aware.

Figure 8-6 Traffic With a Single Notification Only Node: Load-Balancer With Failover Support

Surrounding text describes Figure 8-6 .

Network Configuration

The general structure of a production Services Gatekeeper installation is also designed to support redundancy and high availability. A typical installation consists of a number of UNIX/Linux servers connected through duplicated switches. Each server has redundant network cards connected to separate switches. The servers are organized into clusters, with the number of servers in the cluster determined by the needed capacity.

As described previously, Services Gatekeeper is deployed on an Access Tier, which manages connections to applications, and a Network Tier, which manages connections to the underlying telecom network. For security, the Network Tier is usually connected only to Access Tier servers, the appropriate underlying network nodes, and the Oracle WebLogic Server Administration Server, which manages the domain. A third tier hosts the database. This tier should be hosted on dedicated, redundant servers. For physical storage, a Network Attached Storage using fibre channel controller cards is an option.

Because the different tiers perform different tasks, their servers should be optimized with different physical profiles, including amount of RAM, disk-types, and CPUs. Each tier scales individually, so the number of servers in a specific tier can be increased without affecting the other tiers.

A sample configuration is shown in Figure 8-7. Smaller systems in which the Access Tier and the Network Tier are co-located in the same physical servers are possible but only for non-production systems. Particular hardware configurations depend on the specific deployment requirements and are worked out in the dimensioning and capacity planning stage.

Figure 8-7 Sample Hardware Configuration

Surrounding text describes Figure 8-7 .

In high-availability mode, all hardware components are duplicated, eliminating any single point of failure. This means that there are at least two servers executing the same software modules, that each server has two network cards, and that each server has a fault-tolerant disk system, as, for example, RAID.

The Administration Server may have duplicate network cards, connected to each switch.

For security reasons, the servers used for the Access Tier can be separated from the Network Tier servers using firewalls. The Access Tier servers reside in a Demilitarized Zone (DMZ) while the Network Tier servers are in a trusted environment.

Geographic Redundancy

All Services Gatekeeper modules in production systems are deployed in clusters to ensure high availability. This prevents single points of failure in general usage. Within a cluster, a Budget Service cluster-local master regulates the enforcement of SLAs. The enforcement service is highly available and is migrated to another server should the cluster-local master node fail. See"Managing and Configuring Budgets” in System Administrator's Guide for more information on this mechanism.

However, to prevent service failure in the face of catastrophic events - natural disasters or massive system outages like power failures - Services Gatekeeper can also be deployed at two geographically distant sites that are designated as site pairs. Each site, which is a Services Gatekeeper domain, has another site as its peer. See Figure 8-8 for an overview. Application and service provider configuration information, including related SLAs and budget information, is replicated and enforced across sites.

Note:

Custom, Subscriber, Service Provider Node, and Global Node SLAs cannot be replicated across sites.

Figure 8-8 Overview of Geographically Redundant Site Pairs

Surrounding text describes Figure 8-8 .

Geo-Redundant Sites

In a geo-redundant setup, all sites have a geographic site name and each site is configured to have a reference to its peer site using that name. The designated set of information is synchronized between these site peers.

One site is defined as the geomaster, the other as the slave. Checks are run periodically between the site pairs to verify data consistency and an alarm is triggered if mismatches are found, at which point the administrator can force the slave to re-sync to the geomaster, using the syncFromGeoMaster operation. Any relevant configuration changes made to either site are written synchronously across the site pairs, so that a failure to write to either the geomaster or the slave causes the write to fail and an alarm to fire.

During the period in which the slave is syncing up with the geomaster, both the geomaster and the slave sites are in read-only mode. No configuration changes can be made. If a slave site becomes unavailable for any reason, the geomaster site becomes read-only either until the slave site is available and has completed all data replication, or until the slave site has been removed from the geomaster site's configuration, terminating geo-redundancy.

Note:

If a new site is then added to replace the terminated site, it must be added as a slave site. The site that is designated the geomaster site must remain the geomaster site for the lifetime of the site configuration.

If a geomaster site fails permanently, the failed site should be removed from the configuration using the GeoRedundantService. If a replacement site is added to the configuration, the remaining operating site must be reconfigured to be the geomaster and the replacement site must be added as the slave.

If a geomaster site fails permanently, the failed site should be removed from the configuration using the GeoRedundantService. If a replacement site is added to the configuration, the remaining operating site must be reconfigured to be the geomaster and the replacement site must be added as the slave.

Applications and Geo-Redundancy

For applications, geo-redundancy means that their traffic can continue to flow in the face of a catastrophic failure at an operator site. Even applications that normally use only a single site for their traffic can fail over to a peer site while maintaining ongoing SLA enforcement for their accounts. This scenario is particularly relevant for SLA aspects that have longer term impact, such as quotas.

Figure 8-9 Geographically Redundant Site Pairs and Applications

Surrounding text describes Figure 8-9 .

In many respects, the geo-redundancy mechanism is not transparent to applications. There is no single sign-on mechanism across sites, and an application must establish a session with each site it intends to use. In case of site failure, an application must manually fail over to a different site.

While application and service provider budget and configuration information are maintained across sites, state for ongoing conversations is not maintained. Conversations in this sense are defined in terms of the correlation identifiers that are returned to the applications by Services Gatekeeper or passed into Services Gatekeeper from the applications. Any state associated with a correlation identifier exists on only a single geographic site and is lost in the event of a site-wide disaster. Conversational state includes, but is not limited to, call state and registration for network-triggered notifications. This type of state is considered volatile, or transient, and is not replicated at the site level.

This means that conversations must be conducted and complete on their site of origin. If an application wishes to maintain conversational state cross-site - for example, to maintain a registration for network-triggered traffic - it must register with each site individually.

Note:

On the other hand, this type of affinity does allow load balancing between sites for different or new conversations. For example, because each request to send an SMS message constitutes a new conversation, sending SMS messages can be balanced between the sites.

Below is a high-level outline of the redundancy functionality:

  • The contractual usage relationships represented by SLAs can be enforced across geographic site domains. The mechanism covers SLAs on both the service provider group and application group level.

  • Service provider and application account configuration data, including any changes to this information, can be replicated across sites, reducing the administrative overhead in setting up geo-redundant site pairs.

  • When peer sites fail to establish connection a configurable number of times, a connection-lost alarms is raised.

  • Alarms are also generated:

    • If there is a site configuration mismatch between the two sites; for example if site A treats site B as a peer, but site B does not recognize site A as a peer

    • If the paired sites do not have identical application and service provider configuration information, including related SLAs and budget information

    • If the master site fails to complete a configuration update to the slave site