Red Hat Enterprise Linux 7 High Availability Add

Red Hat Enterprise Linux 7
High Availability Add-On
Administration
Configuring and Managing the High Availability Add-On
Red Hat Enterprise Linux 7 High Availability Add-On Administration
Configuring and Managing the High Availability Add-On
Legal Notice
Co pyright © 20 14 Red Hat, Inc. and o thers.
This do cument is licensed by Red Hat under the Creative Co mmo ns Attributio n-ShareAlike 3.0
Unpo rted License. If yo u distribute this do cument, o r a mo dified versio n o f it, yo u must pro vide
attributio n to Red Hat, Inc. and pro vide a link to the o riginal. If the do cument is mo dified, all Red
Hat trademarks must be remo ved.
Red Hat, as the licenso r o f this do cument, waives the right to enfo rce, and agrees no t to assert,
Sectio n 4 d o f CC-BY-SA to the fullest extent permitted by applicable law.
Red Hat, Red Hat Enterprise Linux, the Shado wman lo go , JBo ss, MetaMatrix, Fedo ra, the Infinity
Lo go , and RHCE are trademarks o f Red Hat, Inc., registered in the United States and o ther
co untries.
Linux ® is the registered trademark o f Linus To rvalds in the United States and o ther co untries.
Java ® is a registered trademark o f Oracle and/o r its affiliates.
XFS ® is a trademark o f Silico n Graphics Internatio nal Co rp. o r its subsidiaries in the United
States and/o r o ther co untries.
MySQL ® is a registered trademark o f MySQL AB in the United States, the Euro pean Unio n and
o ther co untries.
No de.js ® is an o fficial trademark o f Jo yent. Red Hat So ftware Co llectio ns is no t fo rmally
related to o r endo rsed by the o fficial Jo yent No de.js o pen so urce o r co mmercial pro ject.
The OpenStack ® Wo rd Mark and OpenStack Lo go are either registered trademarks/service
marks o r trademarks/service marks o f the OpenStack Fo undatio n, in the United States and o ther
co untries and are used with the OpenStack Fo undatio n's permissio n. We are no t affiliated with,
endo rsed o r spo nso red by the OpenStack Fo undatio n, o r the OpenStack co mmunity.
All o ther trademarks are the pro perty o f their respective o wners.
Abstract
High Availability Add-On Administratio n describes the co nfiguratio n and management o f the
High Availability Add-On fo r Red Hat Enterprise Linux 7.
T able of Cont ent s
T able of Contents
. .hapt
⁠C
. . . .er
. .1. .. Creat
. . . . .ing
. . . a. .Red
. . . .Hat
. . . High. . . . . Availabilit
. . . . . . . . .y. Clust
. . . . . er
. . wit
. . .h. .Pacemaker
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2. . . . . . . . . .
⁠1.1. Clus ter So ftware Ins tallatio n
2
⁠1.2. Clus ter Creatio n
3
⁠1.3. Fenc ing Co nfig uratio n
4
. .hapt
⁠C
. . . .er
. .2. .. Configuring
. . . . . . . . . . . an
. . .Apache
. . . . . . .Web
. . . .Server
. . . . . .in
. . a. .Red
. . . .Hat
. . . High
. . . . .Availabilit
. . . . . . . . .y. Clust
. . . . .er
. .wit
. . .h. t. he
. . .pcs
.............
Command
6
⁠2 .1. Co nfig uring an LVM Vo lume with an ext4 File Sys tem
7
⁠2 .2. Web Server Co nfig uratio n
8
⁠2 .3. Exc lus ive Ac tivatio n o f a Vo lume G ro up in a Clus ter
8
⁠2 .4. Creating the Res o urc es and Res o urc e G ro up s with the p c s Co mmand
10
⁠2 .5. Tes ting the Res o urc e Co nfig uratio n
12
⁠2 .6 . Clus ter p c s Co mmand Summary
13
. . . . . . . . .Hist
Revision
. . . ory
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1. 5. . . . . . . . . .
1
Red Hat Ent erprise Linux 7 High Availabilit y Add- O n Administ rat ion
Chapter 1. Creating a Red Hat High-Availability Cluster with
Pacemaker
This chapter describes the procedure for creating a Red Hat High Availability two-node cluster using
pcs. After you have created a cluster, you can configure the resources and resource groups that you
require.
Configuring the cluster provided in this chapter requires that your system include the following
components:
2 nodes, which will be used to create the cluster. In this example, the nodes used are
z1. exampl e. co m and z2. exampl e. co m.
Network switches for the private network, required for communication among the cluster nodes
and other cluster hardware such as network power switches and Fibre Channel switches.
A power fencing device for each node of the cluster. This example uses two ports of the APC
power switch with a host name of zapc. exampl e. co m.
This chapter is divided into three sections.
Section 1.1, “ Cluster Software Installation” provides the procedure for installing the cluster
software.
Section 1.2, “ Cluster Creation” provides the procedure for configuring a two-node cluster.
Section 1.3, “ Fencing Configuration” provides the procedure for configuring fencing devices for
each node of the cluster.
1.1. Clust er Soft ware Inst allat ion
The procedure for installing and configuring a cluster is as follows.
1. On each node in the cluster, install the Red Hat High Availability Add-On software packages
along with all available fence agents from the High Availability channel.
# yum i nstal l pcs fence-ag ents-al l
2. If you are running the fi rewal l d daemon, execute the following commands to enable the
ports that are required by the Red Hat High Availability Add-On.
Note
You can determine whether the fi rewal l d daemon is installed on your system with
the rpm -q fi rewal l d command. If the fi real l d daemon is installed, you can
determine whether it is running with the fi rewal l -cmd --state command.
# fi rewal l -cmd --permanent --ad d -servi ce= hi g h-avai l abi l i ty
# fi rewal l -cmd --ad d -servi ce= hi g h-avai l abi l i ty
2
⁠Chapt er 1 . Creat ing a Red Hat High- Availabilit y Clust er wit h Pacemaker
3. In order to use pcs to configure the cluster and communicate among the nodes, you must set
a password on each node for the user ID hacl uster, which is the the pcs administration
account. It is recommended that the password for user hacl uster be the same on each
node.
# passwd hacl uster
Changing password for user hacluster.
New password:
Retype new password:
passwd: all authentication tokens updated successfully.
4. Before the cluster can be configured, the pcsd daemon must be started and enabled to boot
on startup on each node. This daemon works with the pcs command to manage
configuration across the nodes in the cluster.
On each node in the cluster, execute the following commands to start the pcsd service and to
enable pcsd at system start.
# systemctl start pcsd . servi ce
# systemctl enabl e pcsd . servi ce
5. Authenticate the pcs user hacl uster for each node in the cluster on the node from which
you will be running pcs.
The following command authenticates user hacl uster on z1. exampl e. co m for both of the
nodes in the example two-node cluster, z1. exampl e. co m and z2. exampl e. co m.
[email protected] z1 ~]# pcs cl uster auth z1. exampl e. co m z2. exampl e. co m
Username: hacluster
Password:
z1.example.com: Authorized
z2.example.com: Authorized
1.2. Clust er Creat ion
This procedure creates a Red Hat High Availability Add-On cluster that consists of the nodes
z1. exampl e. co m and z2. exampl e. co m.
1. Execute the following command from z1. exampl e. co m to create the two-node cluster
mycl uster that consists of nodes z1. exampl e. co m and z2. exampl e. co m. This will
propagate the cluster configuration files to both nodes in the cluster. This command includes
the --start option, which will start the cluster services on both nodes in the cluster.
[[email protected] z1 ~]# pcs cl uster setup --start --name my_cl uster \
z1. exampl e. co m z2. exampl e. co m
z1.example.com: Succeeded
z1.example.com: Starting Cluster...
z2.example.com: Succeeded
z2.example.com: Starting Cluster...
2. Enable the cluster services to run on each node in the cluster when the node is booted.
3
Red Hat Ent erprise Linux 7 High Availabilit y Add- O n Administ rat ion
Note
For your particular environment, you may choose to leave the cluster services disabled
by skipping this step. This allows you to ensure that if a node goes down, any issues
with your cluster or your resources are resolved before the node rejoins the cluster. If
you leave the cluster services disabled, you will need to manually start the services
when you reboot a node by executing the pcs cl uster start command on that
node.
# pcs cl uster enabl e --al l
You can display the current status of the cluster with the pcs cl uster status command.
[[email protected] z1 ~]# pcs cl uster status
Cluster Status:
Last updated: Thu Jul 25 13:01:26 2013
Last change: Thu Jul 25 13:04:45 2013 via crmd on z2.example.com
Stack: corosync
Current DC: z2.example.com (2) - partition with quorum
Version: 1.1.10-5.el7-9abe687
2 Nodes configured
0 Resources configured
1.3. Fencing Configurat ion
You must configure a fencing device for each node in the cluster. For general information about
configuring fencing devices, see the Red Hat Enterprise Linux 7 High Availability Add-On Reference.
Note
When configuring a fencing device, you should ensure that your fencing device does not
share power with the node that it controls.
This example uses the APC power switch with a host name of zapc. exampl e. co m to fence the
nodes, and it uses the fence_apc_snmp fencing agent. Because both nodes will be fenced by the
same fencing agent, you can configure both fencing devices as a single resource, using the
pcmk_ho st_map and pcmk_ho st_l i st options.
You create a fencing device by configuring the device as a sto ni th resource with the pcs sto ni th
create command. The following command configures a sto ni th resource named myapc that uses
the fence_apc_snmp fencing agent for nodes z1. exampl e. co m and z2. exampl e. co m. The
pcmk_ho st_map option maps z1. exampl e. co m to port 1, and z2. exampl e. co m to port 2. The
login value and password for the APC device are both apc. By default, this device will use a monitor
interval of 60s for each node.
Note that you can use an IP address when specifying the host name for the nodes.
[[email protected] z1 ~]# pcs sto ni th create myapc fence_apc_snmp params \
i pad d r= "zapc. exampl e. co m"
pcmk_ho st_map= "z1. exampl e. co m: 1;z2. exampl e. co m: 2" \
4
⁠Chapt er 1 . Creat ing a Red Hat High- Availabilit y Clust er wit h Pacemaker
pcmk_ho st_check= "stati c-l i st"
pcmk_ho st_l i st= "z1. exampl e. co m,z2. exampl e. co m" \
l o g i n= "apc" passwd = "apc"
Note
When you create a fence_apc_snmp sto ni th device, you may see the following warning
message, which you can safely ignore:
Warning: missing required option(s): 'port, action' for resource
type: stonith:fence_apc_snmp
The following command displays the parameters of an existing STONITH device.
[[email protected] rh7-1 ~]# pcs sto ni th sho w myapc
Resource: myapc (class=stonith type=fence_apc_snmp)
Attributes: ipaddr=zapc.example.com
pcmk_host_map=z1.example.com:1;z2.example.com:2 pcmk_host_check=staticlist pcmk_host_list=z1.example.com,z2.example.com login=apc passwd=apc
Operations: monitor interval=60s (myapc-monitor-interval-60s)
5
Red Hat Ent erprise Linux 7 High Availabilit y Add- O n Administ rat ion
Chapter 2. Configuring an Apache Web Server in a Red Hat High
Availability Cluster with the pcs Command
This chapter describes how to configure an Apache web server in a two-node Red Hat Enterprise
Linux High Availability Add-On cluster using pcs to configure cluster resources. In this use case,
clients access the Apache web server through a floating IP address. The web server runs on one of
two nodes in the cluster. If the node on which the web server is running becomes inoperative, the web
server starts up again on the second node of the cluster with minimal service interruption.
Figure 2.1, “ Apache Web Server in a Red Hat High Availability Two-Node Cluster” shows a high-level
overview of the cluster. The cluster is a two-node Red Hat High Availability cluster which is
configured with a network power switch and with shared storage. The cluster nodes are connected to
a public network, for client access to the Apache web server through a virtual IP. The Apache server
runs on either Node 1 or Node 2, each of which has access to the storage on which the Apache data
is kept.
Fig u re 2.1. Ap ach e Web Server in a R ed H at H ig h Availab ilit y T wo - N o d e C lu st er
This use case requires that your system include the following components:
A 2-node Red Hat High Availability cluster with power fencing configured for each node. This
procedure uses the cluster example provided in Chapter 1, Creating a Red Hat High-Availability
Cluster with Pacemaker.
A public virtual IP address, required for the Apache web server.
Shared storage for the nodes in the cluster, using iSCSI or Fibre Channel.
The cluster is configured with an Apache resource group, which contains the cluster components
that the web server requires: an LVM resource, a file system resource, an IP address resource, and a
web server resource. This resource group can fail over from one node of the cluster to the other,
allowing either node to run the web server. Before creating the resource group for this cluster, you will
perform the following procedures:
6
⁠Chapt er 2 . Configuring an Apache Web Server in a Red Hat High Availabilit y Clust er wit h t he pcs Command
1. Configure an ext4 file system mounted on the logical volume my_l v, as described in
Section 2.1, “ Configuring an LVM Volume with an ext4 File System” .
2. Configure a web server, as described in Section 2.2, “ Web Server Configuration” .
3. Ensure that only the cluster is capable of activating the volume group that contains my_l v,
and that the volume group will not be activated outside of the cluster on startup, as described
in Section 2.3, “ Exclusive Activation of a Volume Group in a Cluster” .
After performing these procedures, you create the resource group and the resources it contains, as
described in Section 2.4, “ Creating the Resources and Resource Groups with the pcs Command” .
2.1. Configuring an LVM Volume wit h an ext 4 File Syst em
This use case requires that you create an LVM logical volume on storage that is shared between the
nodes of the cluster.
The following procedure creates an LVM logical volume and then creates an ext4 file system on that
volume. In this example, the shared partition /d ev/sd b1 is used to store the LVM physical volume
from which the LVM logical volume will be created.
Note
LVM volumes and the corresponding partitions and devices used by cluster nodes must be
connected to the cluster nodes only.
Since the /d ev/sd b1 partition is storage that is shared, you perform this procedure on one node
only,
1. Create an LVM physical volume on partition /d ev/sd b1.
# pvcreate /d ev/sd b1
Physical volume "/dev/sdb1" successfully created
2. Create the volume group my_vg that consists of the physical volume /d ev/sd b1.
# vg create my_vg /d ev/sd b1
Volume group "my_vg" successfully created
3. Create a logical volume using the volume group my_vg .
# l vcreate -L4 50 -n my_l v my_vg
Rounding up size to full physical extent 452.00 MiB
Logical volume "my_lv" created
You can use the l vs command to display the logical volume.
# l vs
LV
VG
Copy% Convert
my_lv
my_vg
...
Attr
LSize
Pool Origin Data%
Move Log
-wi-a---- 452.00m
7
Red Hat Ent erprise Linux 7 High Availabilit y Add- O n Administ rat ion
4. Create an ext4 file system on the logical volume my_l v.
# mkfs. ext4 /d ev/my_vg /my_l v
mke2fs 1.42.7 (21-Jan-2013)
Filesystem label=
OS type: Linux
...
2.2. Web Server Configurat ion
The following procedure configures an Apache web server.
1. Ensure that the Apache HTTPD server is installed on each node in the cluster. You also need
the wg et tool installed on the cluster to be able to check the status of the Apache web server.
On each node, execute the following command.
# yum i nstal l -y httpd wg et
2. In order for the Apache resource agent to get the status of the Apache web server, ensure that
the following text is present in the /etc/httpd /co nf/httpd . co nf file on each node in the
cluster, and ensure that it has not been commented out. If this text is not already present, add
the text to the end of the file.
<Location /server-status>
SetHandler server-status
Order deny,allow
Deny from all
Allow from 127.0.0.1
</Location>
3. Create a web page for Apache to serve up. On one node in the cluster, mount the file system
you created in Section 2.1, “ Configuring an LVM Volume with an ext4 File System” , create the
file i nd ex. html on that file system, then unmount the file system.
# mo unt /d ev/my_vg /my_l v /var/www/
# mkd i r /var/www/html
# mkd i r /var/www/cg i -bi n
# mkd i r /var/www/erro r
# resto reco n -R /var/www
# cat <<-END >/var/www/html /i nd ex. html
<html >
<bo d y>Hel l o </bo d y>
</html >
END
# umo unt /var/www
2.3. Exclusive Act ivat ion of a Volume Group in a Clust er
The following procedure configures the volume group in a way that will ensure that only the cluster is
capable of activating the volume group, and that the volume group will not be activated outside of
the cluster on startup. If the volume group is activated by a system outside of the cluster, there is a
8
⁠Chapt er 2 . Configuring an Apache Web Server in a Red Hat High Availabilit y Clust er wit h t he pcs Command
risk of corrupting the volume group's metadata.
This procedure modifies the vo l ume_l i st entry in the /etc/l vm/l vm. co nf configuration file.
Volume groups listed in the vo l ume_l i st entry are allowed to automatically activate on the local
node outside of the cluster manager's control. Volume groups related to the node's local root and
home directories should be included in this list. All volume groups managed by the cluster manager
must be excluded from the vo l ume_l i st entry. Note that this procedure does not require the use of
cl vmd .
Perform the following procedure on each node in the cluster.
1. D etermine which volume groups are currently configured on your local storage with the
following command. This will output a list of the currently-configured volume groups. If you
have space allocated in separate volume groups for root and for your home directory on this
node, you will see those volumes in the output, as in this example.
# vg s --no head i ng s -o vg _name
my_vg
rhel_home
rhel_root
2. Add the volume groups other than my_vg (the volume group you have just defined for the
cluster) as entries to vo l ume_l i st in the /etc/l vm/l vm. co nf configuration file. For
example, if you have space allocated in separate volume groups for root and for your home
directory, you would uncomment the vo l ume_l i st line of the l vm. co nf file and add these
volume groups as entries to vo l ume_l i st as follows:
volume_list = [ "rhel_root", "rhel_home" ]
Note
If no local volume groups are present on a node to be activated outside of the cluster
manager, you must still initialize the vo l ume_l i st entry as vo l ume_l i st = [].
3. Rebuild the i ni trd boot image to guarantee that the boot image will not try to activate a
volume group controlled by the cluster. Update the i ni trd device cluster with the following
command. This command may take up to a minute to complete.
# d racut -H -f /bo o t/i ni tramfs-$(uname -r). i mg $(uname -r)
4. Reboot the node.
Note
If you have installed a new Linux kernel since booting the node on which you created
the boot image, the new i ni trd image will be for the kernel that was running when
you created it and not for the new kernel that is running when you reboot the node.
You can ensure that the correct i ni trd device is in use by running the uname -r
command before and after the reboot to determine the kernel release that is running. If
the releases are not the same, update the i ni trd file after rebooting with the new
kernel and then reboot the node.
9
Red Hat Ent erprise Linux 7 High Availabilit y Add- O n Administ rat ion
5. When the node has rebooted, check whether the cluster services have started up again on
that node by executing the pcs cl uster status command on that node. If this yields the
message Erro r: cl uster i s no t currentl y runni ng o n thi s no d e then run the
following command.
# pcs cl uster start
Alternately, you can wait until you have rebooted each node in the cluster and start cluster
services on each of the nodes with the following command.
# pcs cl uster start --al l
2.4 . Creat ing t he Resources and Resource Groups wit h t he pcs
Command
This use case requires that you create four cluster resources. To ensure these resources all run on
the same node, they are configured as part of the resource group apacheg ro up. The resources to
create are as follows, listed in the order in which they will start.
1. An LVM resource named my_l vm that uses the LVM volume group you created in Section 2.1,
“ Configuring an LVM Volume with an ext4 File System” .
2. A Fi l esystem resource named my_fs, that uses the filesystem device /d ev/my_vg /my_l v
you created in Section 2.1, “ Configuring an LVM Volume with an ext4 File System” .
3. An IP ad d r2 resource, which is a floating IP address for the apacheg ro up resource group.
The IP address must not be one already associated with a physical node. If the IP ad d r2
resource's NIC device is not specified, the floating IP must reside on the same network as the
statically assigned IP addresses used by the cluster nodes, otherwise the NIC device to
assign the floating IP address can not be properly detected.
4. An apache resource named Websi te that uses the i nd ex. html file and the Apache
configuration you defined in Section 2.2, “ Web Server Configuration” .
The following procedure creates the resource group apacheg ro up and the resources that the group
contains. The resources will start in the order in which you add them to the group, and they will stop
in the reverse order in which they are added to the group. Run this procedure from one node of the
cluster only.
1. The following command creates the LVM resource my_l vm. This command specifies the
excl usi ve= true parameter to ensure that only the cluster is capable of activating the LVM
logical volume. Because the resource group apacheg ro up does not yet exist, this command
creates the resource group.
[[email protected] z1 ~]# pcs reso urce create my_l vm LVM vo l g rpname= my_vg \
excl usi ve= true --g ro up apacheg ro up
When you create a resource, the resource is started automatically. You can use the following
command to confirm that the resource was created and has started.
# pcs reso urce sho w
Resource Group: apachegroup
my_lvm (ocf::heartbeat:LVM): Started
10
⁠Chapt er 2 . Configuring an Apache Web Server in a Red Hat High Availabilit y Clust er wit h t he pcs Command
You can manually stop and start an individual resource with the pcs reso urce d i sabl e
and pcs reso urce enabl e commands.
2. The following commands create the remaining resources for the configuration, adding them
to the existing resource group apacheg ro up.
[[email protected] z1 ~]# pcs reso urce create my_fs Fi l esystem \
d evi ce= "/d ev/my_vg /my_l v" d i recto ry= "/var/www" fstype= "ext4 " -g ro up \
apacheg ro up
[[email protected] z1 ~]# pcs reso urce create Vi rtual IP IP ad d r2 i p= 19 8. 51. 10 0 . 3
\
ci d r_netmask= 24 --g ro up apacheg ro up
[[email protected] z1 ~]# pcs reso urce create Websi te apache \
co nfi g fi l e= "/etc/httpd /co nf/httpd . co nf" \
statusurl = "http: //127. 0 . 0 . 1/server-status" --g ro up apacheg ro up
3. After creating the resources and the resource group that contains them, you can check the
status of the cluster. Note that all four resources are running on the same node.
[[email protected] z1 ~]# pcs status
Cluster name: my_cluster
Last updated: Wed Jul 31 16:38:51 2013
Last change: Wed Jul 31 16:42:14 2013 via crm_attribute on
z1.example.com
Stack: corosync
Current DC: z2.example.com (2) - partition with quorum
Version: 1.1.10-5.el7-9abe687
2 Nodes configured
6 Resources configured
Online: [ z1.example.com z2.example.com ]
Full list of resources:
myapc (stonith:fence_apc_snmp): Started z1.example.com
Resource Group: apachegroup
my_lvm (ocf::heartbeat:LVM): Started z1.example.com
my_fs (ocf::heartbeat:Filesystem): Started z1.example.com
VirtualIP (ocf::heartbeat:IPaddr2): Started z1.example.com
Website (ocf::heartbeat:apache): Started z1.example.com
Note that if you have not configured a fencing device for your cluster, as described in
Section 1.3, “ Fencing Configuration” , by default the resources do not start.
4. Once the cluster is up and running, you can point a browser to the IP address you defined as
the IP ad d r2 resource to view the sample display, consisting of the simple word " Hello" .
Hello
11
Red Hat Ent erprise Linux 7 High Availabilit y Add- O n Administ rat ion
If you find that the resources you configured are not running, you can run the pcs
reso urce d ebug -start resource command to test the resource configuration. For
information on the pcs reso urce d ebug -start command, see the High Availability Add-On
Reference manual.
2.5. T est ing t he Resource Configurat ion
In the cluster status display shown in Section 2.4, “ Creating the Resources and Resource Groups
with the pcs Command” , all of the resources are running on node z1. exampl e. co m. You can test
whether the resource group fails over to node z2. exampl e. co m by using the following procedure
to put the first node in stand by mode, after which the node will no longer be able to host resources.
1. The following command puts node z1. exampl e. co m in stand by mode.
[email protected] z1 ~]# pcs cl uster stand by z1. exampl e. co m
2. After putting node z1 in standby mode, check the cluster status. Note that the resources
should now all be running on z2.
[[email protected] z1 ~]# pcs status
Cluster name: my_cluster
Last updated: Wed Jul 31 17:16:17 2013
Last change: Wed Jul 31 17:18:34 2013 via crm_attribute on
z1.example.com
Stack: corosync
Current DC: z2.example.com (2) - partition with quorum
Version: 1.1.10-5.el7-9abe687
2 Nodes configured
6 Resources configured
Node z1.example.com (1): standby
Online: [ z2.example.com ]
Full list of resources:
myapc (stonith:fence_apc_snmp): Started z1.example.com
Resource Group: apachegroup
my_lvm (ocf::heartbeat:LVM): Started z2.example.com
my_fs (ocf::heartbeat:Filesystem): Started z2.example.com
VirtualIP (ocf::heartbeat:IPaddr2): Started z2.example.com
Website (ocf::heartbeat:apache): Started z2.example.com
The web site at the defined IP address should still display, without interruption.
3. To remove z1 from stand by mode, run the following command.
[email protected] z1 ~]# pcs cl uster unstand by z1. exampl e. co m
12
⁠Chapt er 2 . Configuring an Apache Web Server in a Red Hat High Availabilit y Clust er wit h t he pcs Command
Note
Removing a node from stand by mode does not in itself cause the resources to fail
back over to that node. For information on controlling which node resources can run
on, see the chapter on configuring cluster resources in the Red Hat High Availability AddOn Reference.
2.6. Clust er pcs Command Summary
For a quick summary of the cluster configuration procedure, this section provides a listing of the pcs
commands for this use case that create the Apache web server in a cluster, including the
configuration commands that created the cluster itself.
After you have set a password for user hacl uster on both nodes and and started the pcsd service,
the commands to create the cluster and configure fencing for the cluster nodes are as follows.
[[email protected] z1 ~]# pcs cl uster auth z1. exampl e. co m z2. exampl e. co m
[[email protected] z1 ~]# pcs cl uster setup --start --name my_cl uster z1. exampl e. co m \
z2. exampl e. co m
[[email protected] z1 ~]# pcs sto ni th create myapc fence_apc_snmp params \
i pad d r= "zapc. exampl e. co m"
pcmk_ho st_map= "z1. exampl e. co m: 1;z2. exampl e. co m: 2" \
pcmk_ho st_check= "stati c-l i st"
pcmk_ho st_l i st= "z1. exampl e. co m,z2. exampl e. co m" \
l o g i n= "apc" passwd = "apc"
Note
When you create a fence_apc_snmp sto ni th device, you may see the following warning
message, which you can safely ignore:
Warning: missing required option(s): 'port, action' for resource
type: stonith:fence_apc_snmp
After you have set up the initial LVM volume and Apache web server, the following commands
configure the resources and resource groups for the cluster.
[[email protected] z1 ~]# pcs reso urce create my_l vm LVM vo l g rpname= my_vg
excl usi ve= true \
--g ro up apacheg ro up
[[email protected] z1 ~]# pcs reso urce create my_fs Fi l esystem
d evi ce= "/d ev/my_vg /my_l v" \
d i recto ry= "/var/www" fstype= "ext4 " --g ro up apacheg ro up
[[email protected] z1 ~]# pcs reso urce create Vi rtual IP IP ad d r2 i p= 19 8. 51. 10 0 . 3 \
ci d r_netmask= 24 --g ro up apacheg ro up
13
Red Hat Ent erprise Linux 7 High Availabilit y Add- O n Administ rat ion
[[email protected] z1 ~]# pcs reso urce create Websi te apache \
co nfi g fi l e= "/etc/httpd /co nf/httpd . co nf" \
statusurl = "http: //127. 0 . 0 . 1/server-status" --g ro up apacheg ro up
14
Revision Hist ory
Revision History
R evisio n 0.1- 34
Fri D ec 5 2014
St even Levin e
Updating to implement new sort order on the Red Hat Enterprise Linux splash page.
R evisio n 0.1- 33
Version for 7.0 GA release
Mo n Ju n 2 2014
St even Levin e
R evisio n 0.1- 31
Resolves: #886235
D ocument volume_list usage
Wed May 21 2014
St even Levin e
R evisio n 0.1- 29
T u e May 20 2014
Rebuild for style changes and updated draft
St even Levin e
R evisio n 0.1- 20
Updated Beta draft
Wed Ap r 9 2014
St even Levin e
R evisio n 0.1- 8
Beta draft
Fri D ec 6 2013
St even Levin e
R evisio n 0.0- 1
Wed Jan 16 2013
First version for Red Hat Enterprise Linux 7
St even Levin e
15