Use this procedure to quickly configure an HACMP cluster, consisting of 2 nodes and disk heart-beating.


Make sure you have the following in place:

  • Have the IP addresses and host names of both nodes, and for a service IP label. Add these into the /etc/hosts files on both nodes of the new HACMP cluster.
  • Make sure you have the HACMP software installed on both nodes. Just install all the filesets of the HACMP CD-ROM, and you should be good.
  • Make sure you have this entry in /etc/inittab (as one of the last entries):

clinit:a:wait:/bin/touch /usr/es/sbin/cluster/.telinit

  • In case you’re using EMC SAN storage, make sure you configure you’re disks correctly as hdiskpower devices. Or, if you’re using a mksysb image, you may want to follow this procedure EMC ODM cleanup.


  • Create the cluster and its nodes:
  • # smitty hacmp
  • Initialization and Standard Configuration

Configure an HACMP Cluster and Nodes

Enter a cluster name and select the nodes you’re going to use. It is vital here to have the hostnames and IP address correctly entered in the /etc/hosts file of both nodes.

  • Create an IP service label:
  • # smitty hacmp
  • Initialization and Standard Configuration
  • Configure Resources to Make Highly Available
  • Configure Service IP Labels/Addresses

Add a Service IP Label/Address

Enter an IP Label/Address (press F4 to select one), and enter a Network name (again, press F4 to select one).

  • Set up a resource group:
  • # smitty hacmp
  • Initialization and Standard Configuration
  • Configure HACMP Resource Groups

Add a Resource Group

Enter the name of the resource group. It’s a good habit to make sure that a resource group name ends with “rg”, so you can recognize it as a resource group. Also, select the participating nodes. For the “Fallback Policy”, it is a good idea to change it to “Never Fallback”. This way, when the primary node in the cluster comes up, and the resource group is up-and-running on the secondary node, you won’t see a failover occur from the secondary to the primary node.

Note: The order of the nodes is determined by the order you select the nodes here. If you put in “node01 node02” here, then “node01” is the primary node. If you want to have this any other way, now is a good time to correctly enter the order of node priority.

  • Add the Servie IP/Label to the resource group:
  • # smitty hacmp
  • Initialization and Standard Configuration
  • Configure HACMP Resource Groups

Change/Show Resources for a Resource Group (standard)

Select the resource group you’ve created earlier, and add the Service IP/Label.

  • Run a verification/synchronization:
  • # smitty hacmp
  • Extended Configuration

Extended Verification and Synchronization

Just hit [ENTER] here. Resolve any issues that may come up from this synchronization attempt. Repeat this process until the verification/synchronization process returns “Ok”. It’s a good idea here to select to “Automatically correct errors”.

  • Start the HACMP cluster:
  • # smitty hacmp
  • System Management (C-SPOC)
  • Manage HACMP Services

Start Cluster Services

Select both nodes to start. Make sure to also start the Cluster Information Daemon.

  • Check the status of the cluster:
  • # clstat -o

# cldump

Wait until the cluster is stable and both nodes are up.

Basically, the cluster is now up-and-running. However, during the Verification & Synchronization step, it will complain about not having a non-IP network. The next part is for setting up a disk heartbeat network, that will allow the nodes of the HACMP cluster to exchange disk heartbeat packets over a SAN disk. We’re assuming here, you’re using EMC storage. The process on other types of SAN storage is more or less similar, except for some differences, e.g. SAN disks on EMC storage are called “hdiskpower” devices, and they’re called “vpath” devices on IBM SAN storage. 

First, look at the available SAN disk devices on your nodes, and select a small disk, that won’t be used to store any data on, but only for the purpose of doing the disk heartbeat. It is a good habit, to request your SAN storage admin to zone a small LUN as a disk heartbeating device to both nodes of the HACMP cluster. Make a note of the PVID of this disk device, for example, if you choose to use device hdiskpower4:

# lspv | grep hdiskpower4

hdiskpower4   000a807f6b9cc8e5    None

So, we’re going to set up the disk heartbeat network on device hdiskpower4, with PVID 000a807f6b9cc8e5:

  • Create an concurrent volume group:
  • # smitty hacmp
  • System Management (C-SPOC)
  • HACMP Concurrent Logical Volume Management
  • Concurrent Volume Groups

Create a Concurrent Volume Group

Select both nodes to create the concurrent volume group on by pressing F7 for each node. Then select the correct PVID. Give the new volume group a name, for example “hbvg”.

  • Set up the disk heartbeat network:
  • # smitty hacmp
  • Extended Configuration
  • Extended Topology Configuration
  • Configure HACMP Networks

Add a Network to the HACMP Cluster

Select “diskhb” and accept the default Network Name.

  • Run a discovery:
  • # smitty hacmp
  • Extended Configuration

Discover HACMP-related Information from Configured Nodes

  • Add the disk device:
  • # smitty hacmp
  • Extended Configuration
  • Extended Topology Configuration
  • Configure HACMP Communication Interfaces/Devices
  • Add Communication Interfaces/Devices
  • Add Discovered Communication Interface and Devices

Communication Devices

Select the disk device on both nodes by selecting the same disk on each node by pressing F7.

  • Run a Verification & Synchronization again, as described earlier above. Then check with clstat and/or cldump again, to check if the disk heartbeat network comes online.


    Leave a Reply

    Your email address will not be published. Required fields are marked *