Tuesday Dec 09, 2008

Open HA Cluster agent for LDoms Guest Domains

Recently, we provided some information for the HA-xVM agent for OpenSolaris xVM Hypervisor guest domains, which can be found at http://opensolaris.org/os/project/ha-xvm.  If you were interested in that, it may be of interest to you to know that we are now planning to provide an addendum to the HA-xVM project to also include LDoms guest domains.

As such, two architectural reviews are being planned on December 11th and 18th.

- Thursday December 11th at 9:00am US/Pacific, we will be holding an Open CLARC commitment (final design) review for the Open HA Cluster (OHAC) agent for OpenSolaris xVM guest domains.

- Thursday December 18th at 10:00am US/Pacific, we will be holding an Open CLARC inception (requirements) review for the OHAC agent for LDoms guest domains.

A requirements specification for the OHAC agent for LDoms guest domains will be posted to the HA-xVM project page in the next few days, in preparation for the architectural review.

Please feel free to participate in both reviews.  Phone-in details etc., will be provided on http://wiki.genunix.org/wiki/index.php/CLARC.

Neil Garthwaite
Solaris Cluster Engineering

Thursday Sep 11, 2008

Sun Cluster agent for Informix

In August 2007, we started an OpenSolaris project to develop an Open HA Cluster (OHAC) agent for Informix Dynamic Server (IDS), providing a design specification, cheat sheets and code reviews to anyone that was interested.  Within the next few months that followed, the OHAC agent for IDS was complete and available for anyone to download and use.

However, what the OHAC agent for IDS does not have is full 24x7 support for the agent.  Today we've filled that gap by making available for download the Sun Cluster agent for Informix which has been fully tested by our QA group.  The Sun Cluster agent for Informix allows customers that require 24x7 support to purchase a support contract and differs from the OHAC agent for IDS in this regard.

We've tested the Sun Cluster agent for Informix on Solaris 10 (SPARC & x86) in global and non-global zones.  Our initial supported configuration is for IDS v11, however there are internal reports of customers successfully trying out the Sun Cluster agent for Informix with IDS v9.4. Nevertheless, today's supported configuration is for IDS v11.

I hope our development and now delivery of the Sun Cluster agent for Informix is useful and represents a path that you could follow.  The following links maybe useful,

You can find our other OHAC projects and general community information here.


Neil Garthwaite
Solaris Cluster Engineering

Wednesday Aug 27, 2008

Solaris Cluster and Open HA Cluster are ready for MySQL 5.1

If you want high availability for MySQL databases on Solaris or Open Solaris consider Solaris Cluster / Open HA Cluster. Solaris Cluster supports all available releases for MySQ 5.1. The MySQL agent for Solaris Cluster is usable with the release candidates of MySQL 5.1, and definitely with any final release of the MySQL 5.1 enterprise server or the community server. In addition to 5.1 we support all the current versions up to 5.0.51a.

With Solaris Cluster, you can achieve high availability for the MySQL database server and other applications in the same cluster. Solaris Cluster provides you with a rich framework to orchestrate any start or restart dependencies between the MySQL database server and other applications. You can even combine master and slave databases within the same Solaris Cluster. So you can achieve highly available masters and highly available slaves.

One advantage of the MySQL acquisition is that we have now earlier access to to the MySQL enterprise releases and closer access to the MySQL engineering force. So we can qualify new versions earlier than before. The next advantage in the close relationship is a faster problem resolution, because only one organization is working on customer problems.

To enable the MySQL 5.1 support of the Sun Cluster 3.2 agent for MySQL, you have to apply the following patches:

    Solaris 9 Sparc: 126031-04 or higher

    Solaris 10 Sparc: 126032-04 or higher

    Solaris 10 X64: 126033-05 or higher

For more information about setting up MySQL in Solaris Cluster, consult Sun Cluster Data Service for MySQL Guide for Solaris OS

Detlef Ulherr
Solaris Cluster Engineering 

Tuesday May 27, 2008

OSP blog for May/08


Solaris Cluster 3.2u1 or 2/08 has launched! You'll find that most of the vendors listed on the matrix are supported or in the process of supporting this latest release. You can always find the latest updates at the OSP URL, where we welcome your comments and feedback. 

We've got some significant updates this month in relation to increasing the Solaris Cluster node count on two particular configurations. The first being EMC, where we now have (16) node connectivity on Symmetrix Arrays supported on various platforms.

\* Solaris and Solaris Cluster connected to EMC Symmetrix storage is now supported in combinations between (8) and (16) nodes. Support includes the following Sun Servers platforms:  T1000/T2000, M4000/M5000/M8000 along with T5120 and T5220. Combinations of servers are supported on Symmetrix 800 and 1000 arrays. Solaris 10 11/06, 8/07 and 4/08 along with Solaris Cluster 3.2 are included.

For the details of these specific configurations, always reference the Solaris Cluster Open Storage matrix or the EMC, External Support matrix.

In other EMC related configurations, ZFS and MPxIO are now supported with EMC on Solaris 10x86 (Opteron) and Solaris 10 11/16 (Sparc).

Solaris Cluster on NetApp rounds out the server node increases with the NAS storage product being increased from (32) to (48) node attach. We've also included the following servers as updates to the matrix; M4000/5000/8000/9000 along with Sparc Enterprise T5120 and T5220, Sun Fire x4150and x4450. Sun Blade T6300, T6320, T6220 and T6250 are included.

In other related news, Solaris Cluster on HP now includes the new EVA 4400 Storage Array which is a significant addition to the Solaris Cluster storage portfolio.

Until next time.

Roger Autrand
Sr. Manager, Sun Cluster Availability

Friday Apr 18, 2008

Improving Sun Java System Application Server availability with Solaris Cluster


Introduction

Sun Java System Application Server is one of the leading middle-ware products in the market with its robust architecture, stability, and ease of use.  The design of the Application Server by itself has some  high availability (HA) features in the form of node agents (NA) which are spread on multiple nodes to avoid a single point of failure (SPoF).  A simple illustration of the design :




However, as we can notice from the above block diagram, the Domain Administration Server (DAS) is not highly available. If the DAS goes down, then the administrative tasks cannot be done.  Despite the client connections being redirected to other instances of the cluster in case of an instance or NA failure or unavailability, an automated recovery would be desirable to reduce the load on the remaining instances of the cluster.  There are also the hardware, OS and network failure scenarios that needs to be accounted for in critical deployments, in which uptime is one of the main requirements.  

Why is a High Availability Solution Required?

A high availability solution is required to handle those failures which Application Server or for that matter any user-land application, cannot recover from, like network, hardware, operating  system failures, and human errors. Apart from these, there are other scenarios like providing continuous service even when OS or hardware upgrades and/or maintenance is done.

Apart from failures, a high availability solution helps the deployment take advantage of other operating system features, like network level load distribution, link failure detection, and virtualization etc.,  to the fullest.

How to decide on the best solution?

Once customers decide that their deployment is better served by a high availability solution, they need to decide on which solution to choose from the market.  The answer to the following questions will help in the decision making:

Is the solution very mature and robust?

Does the vendor provide an Agent that is specifically designed for Sun Java System Application Server?

Is the solution very easy to use and deploy?

Is the solution cost effective?

Is the solution complete? Can it provide high availability for associated components like
Message Queue?

And importantly, can they get very good support in the form of documentation, customer service and a single point of support?

Why Solaris Cluster?


Solaris Cluster is the best high availability solution for the Solaris platform available. It offers excellent integration with the Solaris Operating System and helps customers make use of new features introduced in Solaris without making modifications on their deployments.  Solaris Cluster supports applications running in containers, offers a very good choice of file systems that can be used, choices of processor architecture, etc.  Some of the  highlights include:

Kernel level integration to make use of Solaris features like containers, ZFS, FMA, etc.

A wide portfolio of agents to support the most widely used applications in the market.

Very robust and quick failure detection mechanism and stability even during very high loads.

IPMP - based network failure detection and load balancing.

The same agent can be used for both Sun Java Application Server and Glassfish.

Data Services Configuration Wizards for most common Solaris Cluster tasks.

Sophisticated fencing mechanism to avoid data corruption.

Detect loss of access to storage by monitoring the disk paths.

How does Solaris Cluster Provide High Availability?

Solaris Cluster provides high availability by using redundant components.  The storage, server and network card are redundant.  The following figure illustrates a simple two-node cluster which has the recommended redundant interconnects, storage accessible to both nodes, and public network interfaces each. It is important to note that this is the recommended configuration and the minimal configuration can have just one shared storage, interconnect, and public network interface.  Solaris Cluster even provides the flexibility of having a single-node cluster as well based on individual needs.

LH =  Logical hostname, type of virtual IP used for moving IP addresses across NICs.

RAID =  any suitable software or hardware based RAID mechanism that provides both redundancy and performance.

One can opt to provide high availability just for the DAS alone or for the node agent as well. The choice is based on the environment. Scalability of the node agents is not a problem with high availability deployments, since multiple node agents can be deployed on a single Solaris Cluster installation. These node agents are configured in multiple resource groups, with each resource group having a single logical host, HAStoragePlus and agent node resource. Since node agents are spread over multiple nodes in a normal deployment, there is no need for additional hardware just because a  highly available architecture is being used.  Storage can be made redundant either with software or hardware based RAID.

Solaris Cluster Failover Steps in Case of a Failure

Solaris Cluster provides a set of sophisticated algorithms that are applied to determine whether to restart an application or to failover to the redundant node. Typically the IP address, the file system on which the application binaries and data reside, and the application resource itself are grouped into a logical entity called resource group (RG).  As the name implies, the IP address, file system, and application itself are viewed as resources and each one of them is identified by a resource type (RT) typically referred to as an agent. The recovery mechanism, i.e restart or fail over to another node is, determined based on a combination of time outs, number of restarts, and history of failovers. An agent typically has start, stop, and validate methods that are used to start, stop, and verify prerequisites every time the application changes state.  It also includes a probe which is executed at a predetermined period of time to determine application availability.

Solaris Cluster has two RTs or agents for the Sun Java System Application Server.  The Resource Type SUNW.jsas is used for DAS, and SUNW.jsas_na for node agent. The probe mechanism involves executing the “asadmin list-domain” and “asadmin list-node-agents” command and interpreting the output to determine if the DAS and the node agents are in the desired state or not.  The Application Server software, file system, and  IP address are moved to the redundant node in case of a failover. Please refer to the Sun Cluster Data Service guide (http://docs.sun.com/app/docs/doc/819-2988) for more details.

The following is a simple illustration of a failover in case of a server crash.
 

In the previously mentioned setup, Application Server is not failed over to the second node if
one of the NICs alone fails. The redundant NIC, which is part of the same IPMP group hosts the logical host to which the DAS and NA make use. A temporary network delay will be noticed for until the logical host is moved from nic1 to nic2.

The Global File System (GFS) is recommended for Application Server deployments since there is very little write activity other than logs on the file system in which the configuration files and in specific deployments, binaries are installed. Because GFS is always mounted on all nodes, it results in better fail over times and quicker startup of Application server in case of a node crash or similar problems.

Maintenance and Upgrades

The same properties that help Solaris Cluster provide recovery during failures can be used to provide service continuity in case of maintenance and upgrade work. 

During any planned OS maintenance or upgrade, the RGs are switched over to the redundant node and the node that needs maintenance is rebooted into non-cluster mode. The planned actions are performed and the node is then rebooted into the cluster.  The same procedure can be repeated for all the remaining nodes of the cluster.

Application Server maintenance or upgrade depends on the way in which the binaries and the data and  configuration files are stored. 

1.)Storing the binaries on the node's internal hard disk and storing the domain and node agent related files on the shared storage.  This method is preferable for environments in which frequent updates are necessary. The downside is the possibility of inconsistency in the application binaries, due to differences in patches or upgrades

2.)Storing both the binaries and the data in the shared storage.   This method provides consistent data during all times but makes upgrades and maintenance without outages difficult.

The choice has to be made by taking into account the procedures and processes followed in the organization.

Other Features

Solaris Cluster also provides features that can be used for co-locating services based on the concept of affinities. For example, you can use negative affinity to evacuate the test environment when a production environment is switched to a node or use positive affinity to move the Application Server resources to the same node on which database server is hosted for better performance etc.

Solaris Cluster has an easy-to-use and intuitive GUI  management tool called Sun Cluster Manager, which can be used to perform most management taks.

Solaris Cluster has an inbuilt telemetry feature that can be used to monitor the usage of resources like CPU, memory, etc.


Sun Java Application server doesn't require any modification for Solaris Cluster as the agent is designed with this scenario in mind.

The same agent can be used for Glassfish as well.

The Message Queue Broker can be made highly available as well with the HA  for Sun Java Message Queue agent.

Consistent with Sun's philosophy, the product is being open sourced in phases and the agents are already available under the CDDL license.

An open source product based on the same code base is available for OpenSolaris releases called Open High Availability Cluster.  For more details on the product and community, please visit http://www.opensolaris.org/os/communities/ohac .

The open-source product also has a comprehensive test suite that serves helps users test their deployment satisfactorily.  For more details, please read http://opensolaris.org/os/community/ha-clusters/ohac/Documentation/Tests/.


Summary

For mission-critical environments, availability against all types of failures is a very important criterion.  Solaris Cluster is best designed to provide the highest availability for  Application Server by virtue of its integration with Solaris OS, stability, and having an agent specifically designed for Sun Java System Application Server.
 

Madhan Kumar
Solaris Cluster Engineering

About

Oracle Solaris Cluster Engineering Blog

Search

Archives
« April 2015
SunMonTueWedThuFriSat
   
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
  
       
Today