How To Create a Linux Cluster in Red Hat/Centos 7

During the OS upgrade cycle to version 7 the team over at Red Hat made some changes under the hood that affected not only the basic packages but also how clustering is done in Linux. These changes resulted from the migration to the new Pacemaker/Corosync engine that powers the tokens and heartbeats needed to maintain quorum as well as the consolidation of the userspace programs used to configure and monitor the cluster services behind the newly introduced pcs utility.

Fortunately, the basics and scenarios behind our previous linux clustering tutorials are sound and they can still be followed with only a few modifications. Rather than making the changes directly to Part 1, Part 2 and Part 3 we will leave those articles unmodified as they are accurate and can be followed verbatim when on Red Hat Enterprise Linux 6 or CentOS Linux 6.

Instead, what is produced here is something similar to what Toki Winter has written on creating an Apache HA cluster. Below you will find a power user's guide to the commands and surface changes brought to you in RCS version 7.


Our first note is on the installation steps. There is now an additional package to add this time around pcs.

yum install pcs fence-agents-all

Also, note that ccs, ricci and luci are deprecated. All cluster configuration, operation and monitoring is either performed from the command line using pcs or from the desktop using pcs-gui. From version 7 onwards all Fedora based distributions including CentOS Linux and Red Hat Enterprise Linux now used Systemd as the init process and for run control. This means that we no longer use service and chkconfig but instead rely on systemctl as the manager for our start up processes.

What this really means when clustering is that you no longer have to remember to start the individual cman,clvmd,rgmanager processes and in the correct sequence. Now to start/stop the cluster you just start a single service at the required milestone.

chkconfig cman on;chkconfig clvmd on;chkconfig rgmanager on;
service cman start;service clvmd start;service rgmanager start;

becomes simply.

systemctl enable pcsd.service
systemctl start pcsd.service
pcs cluster start -all

The firewall has now been migrated from pure iptables to firewalld and the rules for the cluster can be added by including the high-availablity service to the rules table.

firewall-cmd --permanent --add-service=high-availability
firewall-cmd --add-service=high-availability
firewall-cmd --list-service


The configuration utility ccs is no longer available and has been replaced by the pcs command.

pcs will now do both the synchronization between nodes and the addition of members and services. Before configuration syncing was done with ricci which has been removed. Now you are required to create a common system user called hacluster. This user will be used by both pcs and the pcsd service to manager cluster changes.

Configuration has also been split out from the monolithic cluster.conf XML file into two separate files. Cluster configuration in release 7 is in /etc/corosync/corosync.conf for membership and quorum configuration and /var/lib/heartbeat/crm/cib.xml for cluster node and resource configuration.

There is no need to edit these manually as like in release 6 all modifications can be done in pcs.

pcs cluster setup <option> <member> ...


ccs -h <master member> --<option> <member> ...

A cluster can also now be destroyed by issuing the command

pcs cluster destroy <cluster>


To check cluster status do

pcs status

as opposed to clustat. You can also print the full cluster configuration with

pcs config

Resource Relocation

Resources and services are managed using

pcs resource

This replaces the previous clusvcadm command that was use to relocate, enable and disable service groups. The new way to do this is by issuing a move.

pcs resource move <resource>

to move back

pcs resource clear <resource>

Note, resource allocation and movement can also be affected by setting up constraints.

pcs contraint <type> <option>

Starting and Stopping is done by calling the pcs cluster with the start --all flags.


pcs cluster start --all


pcs cluster start <nodename>

While stopping is performed by swapping out start with stop from the above.

Cluster members can be monitored using

pcs status corosync


Putting a node in standby will no longer necessitate its graceful exit from the cluster by doing as in version 6:

service rgmanager stop
service clvmd stop
service cman stop

the correct procedure is now to mark the node as being in standby mode.

pcs cluster standby <node-undergoing-maintenance>

Once the maintenance is complete you simple unstandby the node.

pcs cluster unstandby <node-exiting-maintenance>

For further reading on pacemaker and for more additional examples try reviewing Clusters From Scratch by ClusterLabs.