During the OS upgrade cycle to version 7 the team over at Red Hat made some changes under the hood that affected not only the basic packages but also how clustering is done in Linux. These changes resulted from the migration to the new Pacemaker/Corosync engine that powers the tokens and heartbeats needed to maintain quorum as well as the consolidation of the userspace programs used to configure and monitor the cluster services behind the newly introduced pcs utility.
Fortunately, the basics and scenarios behind our previous linux clustering tutorials are sound and they can still be followed with only a few modifications. Rather than making the changes directly to Part 1, Part 2 and Part 3 we will leave those articles unmodified as they are accurate and can be followed verbatim when on Red Hat Enterprise Linux 6 or CentOS Linux 6.
Instead, what is produced here is something similar to what Toki Winter has written on creating an Apache HA cluster. Below you will find a power user's guide to the commands and surface changes brought to you in RCS version 7.
Our first note is on the installation steps. There is now an additional package to add this time around pcs.
yum install pcs fence-agents-all
Also, note that
luci are deprecated. All cluster configuration, operation and monitoring is either performed from the command line using
pcs or from the desktop using
pcs-gui. From version 7 onwards all Fedora based distributions including CentOS Linux and Red Hat Enterprise Linux now used Systemd as the init process and for run control. This means that we no longer use
chkconfig but instead rely on
systemctl as the manager for our start up processes.
What this really means when clustering is that you no longer have to remember to start the individual cman,clvmd,rgmanager processes and in the correct sequence. Now to start/stop the cluster you just start a single service at the required milestone.
chkconfig cman on;chkconfig clvmd on;chkconfig rgmanager on; service cman start;service clvmd start;service rgmanager start;
systemctl enable pcsd.service systemctl start pcsd.service pcs cluster start -all
The firewall has now been migrated from pure
firewalld and the rules for the cluster can be added by including the high-availablity service to the rules table.
firewall-cmd --permanent --add-service=high-availability firewall-cmd --add-service=high-availability firewall-cmd --list-service
The configuration utility
ccs is no longer available and has been replaced by the
pcs will now do both the synchronization between nodes and the addition of members and services. Before configuration syncing was done with ricci which has been removed. Now you are required to create a common system user called hacluster. This user will be used by both pcs and the pcsd service to manager cluster changes.
Configuration has also been split out from the monolithic cluster.conf XML file into two separate files. Cluster configuration in release 7 is in
/etc/corosync/corosync.conf for membership and quorum configuration and
/var/lib/heartbeat/crm/cib.xml for cluster node and resource configuration.
There is no need to edit these manually as like in release 6 all modifications can be done in pcs.
pcs cluster setup <option> <member> ...
ccs -h <master member> --<option> <member> ...
A cluster can also now be destroyed by issuing the command
pcs cluster destroy <cluster>
To check cluster status do
as opposed to
clustat. You can also print the full cluster configuration with
Resources and services are managed using
This replaces the previous
clusvcadm command that was use to relocate, enable and disable service groups. The new way to do this is by issuing a move.
pcs resource move <resource>
to move back
pcs resource clear <resource>
Note, resource allocation and movement can also be affected by setting up
pcs contraint <type> <option>
Starting and Stopping is done by calling the
pcs cluster with the
start --all flags.
pcs cluster start --all
pcs cluster start <nodename>
While stopping is performed by swapping out
stop from the above.
Cluster members can be monitored using
pcs status corosync
Putting a node in standby will no longer necessitate its graceful exit from the cluster by doing as in version 6:
service rgmanager stop service clvmd stop service cman stop
the correct procedure is now to mark the node as being in standby mode.
pcs cluster standby <node-undergoing-maintenance>
Once the maintenance is complete you simple unstandby the node.
pcs cluster unstandby <node-exiting-maintenance>
For further reading on pacemaker and for more additional examples try reviewing Clusters From Scratch by ClusterLabs.