diff --git a/doc/Clusters_from_Scratch/en-US/Ch-Intro.xml b/doc/Clusters_from_Scratch/en-US/Ch-Intro.xml index c2734a7832..173f6e78d1 100644 --- a/doc/Clusters_from_Scratch/en-US/Ch-Intro.xml +++ b/doc/Clusters_from_Scratch/en-US/Ch-Intro.xml @@ -1,175 +1,172 @@ Read-Me-First
The Scope of this Document - The purpose of this document is to definitively explain the concepts used to configure Pacemaker. - To achieve this best, it will focus exclusively on the XML syntax used to configure the CIB. + Computer clusters can be used to provide highly available services or resources. The redundancy of multiple machines is used to guard against failures of many types. - For those that are allergic to XML, Pacemaker comes with a cluster shell and a Python based GUI exists, however these tools will not be covered at all in this document - - It is hoped however, that having understood the concepts explained here, that the functionality of these tools will also be more readily understood. - - , precisely because they hide the XML. + This document will walk through the installation and setup of simple clusters using the Fedora distribution, version 14. The clusters described here will use Pacemaker and Corosync to provide resource management and messaging. Required packages and modifications to their configuration files are described along with the use of the Pacemaker command line tool for generating the XML used for cluster control. - Additionally, this document is NOT a step-by-step how-to guide for configuring a specific clustering scenario. - Although such guides exist, the purpose of this document is to provide an understanding of the building blocks that can be used to construct any type of Pacemaker cluster. + Pacemaker is a central component and provides the resource management required in these systems. This management includes detecting and recovering from the failure of various nodes, resources and services under its control. + + + When more in depth information is required and for real world usage, please refer to the Pacemaker Explained manual.
What Is Pacemaker? - Pacemaker is a cluster resource manager. - It achieves maximum availability for your cluster services (aka. resources) by detecting and recovering from node and resource-level failures by making use of the messaging and membership capabilities provided by your preferred cluster infrastructure (either Corosync or Heartbeat). + Pacemaker is a cluster resource manager. It achieves maximum availability for your cluster services (aka. resources) by detecting and recovering from node and resource-level failures by making use of the messaging and membership capabilities provided by your preferred cluster infrastructure (either Corosync or Heartbeat). Pacemaker's key features include: Detection and recovery of node and service-level failures Storage agnostic, no requirement for shared storage Resource agnostic, anything that can be scripted can be clustered Supports STONITH for ensuring data integrity Supports large and small clusters Supports both quorate and resource driven clusters Supports practically any redundancy configuration Automatically replicated configuration that can be updated from any node Ability to specify cluster-wide service ordering, colocation and anti-colocation - Support for advanced services type + Support for advanced service types Clones: for services which need to be active on multiple nodes Multi-state: for services with multiple modes (eg. master/slave, primary/secondary) Unified, scriptable, cluster shell
-
- Types of Pacemaker Clusters - Pacemaker makes no assumptions about your environment, this allows it to support practically any redundancy configuration including Active/Active, Active/Passive, N+1, N+M, N-to-1 and N-to-N. - - -
- Active/Passive Redundancy - - - - - Two-node Active/Passive clusters using Pacemaker and DRBD are a cost-effective solution for many High Availability situations. - -
-
- -
- Shared Failover - - - - - By supporting many nodes, Pacemaker can dramatically reduce hardware costs by allowing several active/passive clusters to be combined and share a common backup node - -
-
- -
- N to N Redundancy - - - - - - When shared storage is available, every node can potentially be used for failover. - Pacemaker can even run multiple copies of services to spread out the workload. - - -
-
-
Pacemaker Architecture At the highest level, the cluster is made up of three pieces: + - Core cluster infrastructure providing messaging and membership functionality (illustrated in red) + Non-cluster aware components (illustrated in green). + + + These pieces include the resources themselves, scripts that start, stop and monitor them, and also a local daemon that masks the differences between the different standards these scripts implement. + - Non-cluster aware components (illustrated in blue). - In a Pacemaker cluster, these pieces include not only the scripts that knows how to start, stop and monitor resources, but also a local daemon that masks the differences between the different standards these scripts implement. + Resource management + + + Pacemaker provides the brain (illustrated in blue) that processes and reacts to events regarding the cluster. These events include nodes joining or leaving the cluster; resource events caused by failures, maintenance, scheduled activities; and other administrative actions. Pacemaker will compute the ideal state of the cluster and plot a path to achieve it after any of these events. This may include moving resources, stopping nodes and even forcing them offline with remote power switches. + - A brain (illustrated in green) that processes and reacts to events from the cluster (nodes leaving or joining) and resources (eg. monitor failures) as well as configuration changes from the administrator. - In response to all of these events, Pacemaker will compute the ideal state of the cluster and plot a path to achieve it. - This may include moving resources, stopping nodes and even forcing them offline with remote power switches. + Low level infrastructure + + + Corosync provides reliable messaging, membership and quorum information about the cluster (illustrated in red). +
Conceptual Stack Overview - + Conceptual overview of the cluster stack
When combined with Corosync, Pacemaker also supports popular open source cluster filesystems Even though Pacemaker also supports Heartbeat, the filesystems need to use the stack for messaging and membership and Corosync seems to be what they're standardizing on. Technically it would be possible for them to support Heartbeat as well, however there seems little interest in this. Due to recent standardization within the cluster filesystem community, they make use of a common distributed lock manager which makes use of Corosync for its messaging capabilities and Pacemaker for its membership (which nodes are up/down) and fencing services.
The Pacemaker Stack - + The Pacemaker stack when running on Corosync
Internal Components Pacemaker itself is composed of four key components (illustrated below in the same color scheme as the previous diagram): CIB (aka. Cluster Information Base) CRMd (aka. Cluster Resource Management daemon) PEngine (aka. PE or Policy Engine) STONITHd
Internal Components - + Subsystems of a Pacemaker cluster running on Corosync
The CIB uses XML to represent both the cluster's configuration and current state of all resources in the cluster. The contents of the CIB are automatically kept in sync across the entire cluster and are used by the PEngine to compute the ideal state of the cluster and how it should be achieved. This list of instructions is then fed to the DC (Designated Co-ordinator). Pacemaker centralizes all cluster decision making by electing one of the CRMd instances to act as a master. Should the elected CRMd process, or the node it is on, fail... a new one is quickly established. The DC carries out the PEngine's instructions in the required order by passing them to either the LRMd (Local Resource Management daemon) or CRMd peers on other nodes via the cluster messaging infrastructure (which in turn passes them on to their LRMd process). The peer nodes all report the results of their operations back to the DC and based on the expected and actual results, will either execute any actions that needed to wait for the previous one to complete, or abort processing and ask the PEngine to recalculate the ideal cluster state based on the unexpected results. In some cases, it may be necessary to power off nodes in order to protect shared data or complete resource recovery. For this Pacemaker comes with STONITHd. STONITH is an acronym for Shoot-The-Other-Node-In-The-Head and is usually implemented with a remote power switch. In Pacemaker, STONITH devices are modeled as resources (and configured in the CIB) to enable them to be easily monitored for failure, however STONITHd takes care of understanding the STONITH topology such that its clients simply request a node be fenced and it does the rest.
+
+ Types of Pacemaker Clusters + Pacemaker makes no assumptions about your environment, this allows it to support practically any redundancy configuration including Active/Active, Active/Passive, N+1, N+M, N-to-1 and N-to-N. + + In this document we will focus on the setup of a highly available Apache web server with an Active/Passive cluster using DRBD and Ext4 to store data. Then, we will upgrade this cluster to Active/Active using GFS2. + + +
+ Active/Passive Redundancy + + + + + Two-node Active/Passive clusters using Pacemaker and DRBD are a cost-effective solution for many High Availability situations. + +
+
+ +
+ N to N Redundancy + + + + + + When shared storage is available, every node can potentially be used for failover. + Pacemaker can even run multiple copies of services to spread out the workload. + + +
+
+
diff --git a/doc/Clusters_from_Scratch/en-US/images/pcmk-active-active.svg b/doc/Clusters_from_Scratch/en-US/images/pcmk-active-active.svg index c377cce410..f07a20b451 100644 --- a/doc/Clusters_from_Scratch/en-US/images/pcmk-active-active.svg +++ b/doc/Clusters_from_Scratch/en-US/images/pcmk-active-active.svg @@ -1,1398 +1,1419 @@ + gradientUnits="userSpaceOnUse" + gradientTransform="translate(2.1390374,453.4317)" /> image/svg+xml + transform="matrix(0.43829706,0,0,0.49424167,332.72274,539.50866)" /> + transform="matrix(0.43829706,0,0,0.49424167,596.24792,538.51856)" /> + transform="matrix(0.43829706,0,0,0.49424167,200.96015,581.0134)" /> + transform="matrix(0.43829706,0,0,0.49424167,465.55485,581.17224)" /> D'base + transform="matrix(0.43829706,0,0,0.49424167,333.71284,622.67698)" /> + transform="matrix(0.43829706,0,0,0.49424167,597.23802,621.68688)" /> + transform="matrix(0.43829706,0,0,0.49424167,201.95025,663.19162)" /> Web Site Web Site + transform="matrix(0.43829706,0,0,0.49424167,466.54495,663.35046)" /> Web Site Web Site GFS2 + transform="matrix(0.43829706,0,0,0.49424167,333.71284,704.8552)" /> + transform="matrix(0.43829706,0,0,0.49424167,597.23802,703.8651)" /> Host Host Host Host Shared Storage CoroSync Pacemaker Active / Active + y="524.63947">Active / Active Hardware - ClusterServicesSoftware - Services + id="tspan4847" + x="203.22443" + y="714.25586">ClusterSoftwareHardware GFS2 GFS2 GFS2 URL Mail D'base URL Mail D'base URL Mail D'base URL Mail diff --git a/doc/Clusters_from_Scratch/en-US/images/pcmk-active-passive.svg b/doc/Clusters_from_Scratch/en-US/images/pcmk-active-passive.svg index 2c33fd0db2..2e0b0b01e9 100644 --- a/doc/Clusters_from_Scratch/en-US/images/pcmk-active-passive.svg +++ b/doc/Clusters_from_Scratch/en-US/images/pcmk-active-passive.svg @@ -1,1027 +1,1079 @@ + + + - - + gradientTransform="matrix(1.4992949,0,0,1.4260558,437.29747,348.66482)" /> + + + + + + + + + + + + + + + + + + inkscape:guide-bbox="true" + inkscape:snap-grids="false" + inkscape:snap-to-guides="false" /> image/svg+xml URL + transform="matrix(0.43829706,0,0,0.49424167,338.46686,549.05022)" /> + transform="matrix(0.43829706,0,0,0.49424167,207.36122,590.79415)" /> Web Site + transform="matrix(0.43829706,0,0,0.49424167,339.04439,631.38821)" /> + transform="matrix(0.43829706,0,0,0.49424167,600.43053,630.39811)" /> + transform="matrix(0.43829706,0,0,0.49424167,208.35132,672.97237)" /> Files Storage - - + transform="matrix(0.86942866,0,0,0.48920271,434.41441,910.33146)" /> Host Host - Host - Host CoroSync Pacemaker Active / Passive + id="tspan2970">Active / Passive Hardware - ClusterServicesSoftware - Services + id="tspan3084" + x="230.71913" + y="643.65417" />ClusterSoftwareHardware URL D/base Storage D/base Synch Synch + inkscape:connection-end="#rect3922" + inkscape:connection-start="#rect3918" />