Page Menu
Home
ClusterLabs Projects
Search
Configure Global Search
Log In
Files
F1842092
No One
Temporary
Actions
View File
Edit File
Delete File
View Transforms
Subscribe
Mute Notifications
Flag For Later
Award Token
Size
19 KB
Referenced Files
None
Subscribers
None
View Options
diff --git a/cts/README b/cts/README
index b86fc942f9..ced3dfd428 100644
--- a/cts/README
+++ b/cts/README
@@ -1,450 +1,507 @@
BASIC REQUIREMENTS BEFORE STARTING:
Three machines: one test exerciser and two test cluster machines.
The two test cluster machines need to be on the same subnet.
and they should have journalling filesytems for
all their filesystems
You also need two free IP addresses on that subnet to test
mutual IP address takeover
The test exerciser machine doesn't need to be on the same subnet
as the test machines. Minimal demands are made on the exerciser
machine - it just has to stay up during the tests ;-).
However, it does need to have a current copy of the cts test
scripts. It is worth noting that these scripts are coordinated
with particular versions of linux-ha, so that in general you
have to the same version of test scripts as the rest of linux-ha.
Install heartbeat, heartbeat-pils, and heartbeat-stonith on all three
machines. Set up the configuration on the cluster machines *and make
a copy of on the test exerciser machine*. These are the necessary files:
/etc/ha.d/ha.cf
/etc/ha.d/haresources
/etc/ha.d/authkeys
Note that wherever machine names are mentioned in these configuration files,
they must match the machines' `uname -n` name. This may or may not match
the machines' FQDN (fully qualified domain name) - it depends on how
you (and your OS) have named the machines. It helps a lot in tracking
problems if the three machines' clocks are closely synchronized. xntpd
does this, but you can do it by hand if you want.
Make sure the at service is enabled on the test cluster machines.
(this is normally the 'atd' service started by /etc/init.d/atd).
This doesn't mean just start it, it means enable it to start on every boot into
your default init state (probably either 3 or 5). Enabling it for both state
3 and 5 is a good minimum. We don't need this in production - just for these
tests.
Make sure all your filesystems are journalling filesystems (/boot can be
ext2 if you want)... This means jfs, ext3, or reiserfs.
Here's what you need to do to run CTS
Configure the two "cluster" machines with their logging of heartbeat
messages redirected via syslog to the third machine. Let's call it the
exerciser... The exerciser doesn't have to be the same OS as the others
but it needs to be one that supports a lot of the other things
(like ssh and remote syslog logging).
You may want to configure the cluster machines to boot into run level 3,
that is without Xdm logins - particularly if they're behind a KVM switch.
Some distros refuse to boot correctly without knowing what kind of mouse
is present, and the kvm switch will likely make it impossible to figure
out without manual intervention. And since some of the tests cause the
machine to reboot without manual intervention this would be a problem.
Configure syslog on the cluster machines accordingly.
(see the mini-HOWTOs at the end for more details)
The exerciser needs to be able to ssh over to the cluster nodes as root
without a password challenge. Configure ssh accordingly.
(see the mini-HOWTOs at the end for more details)
The exerciser needs to have the IP addresses of the test machines available
to it - either by DNS or by /etc/hosts. It uses this to validate configuration
information.
The "heartbeat" service (init script) needs to be enabled to
automatically start in the default run level on the cluster machines.
This typically means you need a symlink for /etc/rc.d/rc3.d/S*heartbeat
to /etc/init.d/heartbeat, and one in /etc/rc.d/rc5.d/S*heartbeat.
If you don't do this, then things will look fine until you run the STONITH
test - and it will always fail...
The test software is called cts and is in the (surprise!) cts directory.
It's in the tarball, and (for later versions) is installed in
/usr/lib/heartbeat/cts.
The cts system consists of the following files:
CM_fs.py - ignore this - it's for failsafe
CM_hb.py - interacts with heartbeat
CTS.py - the core common code for testing
CTSaudits.py - performs audits at the end of each test
CTSlab.py - defines the "lab" (test) environment
CTStests.py - contains the definitions of the tests
You probably should look at the CTSlab.py file, but you should no longer
need to modify it.
OK. Now assuming you did all this and the stuff described below, what you
need to do is run CTSlab.py. This is the same as the file you modified above.
If you run any other file, it won't test your cluster ;-)
Depending on permissions, etc., this may be either done as:
./CTSlab.py number-of-tests-to-run
or as
python ./CTSlab.py number-of-tests-to-run
The test output goes to standard error, so you'll probably want to catch stderr
with the usual 2>&1 construct like this:
./CTSlab.py > outputfile 2>&1 &
followed by a
tail -f outputfile
Options for CTSlab:
--suppressmonitoring
Don't "monitor" resources as part of the audits
--directory dirname
Directory to find config info in. Defaults to
/etc/ha.d
--logfile
Directory to find logging information in defaults to
/var/log/ha-log-local7
--stonith (yes|no)
Enable/disable STONITH tests
--standby (yes|no)
Enable/disable standby tests
-
-
+ --v2
+ Test the release 2.0 ( include 1.99.*)
+ (see the mini-HOWTOs at the end for more details)
+
==============
Mini HOWTOs:
==============
--------------------------------------------------------------------------------
How to redirect linux-HA logging the way CTS wants it using syslog
--------------------------------------------------------------------------------
NOTE: There have been reports of messages being lost with vanilla syslog.
At least one of the developers recommend using syslog-ng or indead
anything else to avoid these problems.
1) Redirect each machines to go (at least) to syslog local7:
Change /etc/ha.d/ha.cf on each test machine to say this:
logfacility local7
(you can also log to a dedicated local file with logfile if you want)
2) Change /etc/syslog.conf to redirect local7 on each of your slave
machines to redirect to your testmonitor machine by adding this line
somewhere near the top of /etc/syslog.conf
local7.* @testmonitor-machine
3) Change syslog on the testmonitor-machine to accept remote
logging requests. You do this by making sure it gets invoked with
the "-r" option On SuSE Linux you need to change /etc/rc.config
to put have this line for SYSLOGD_PARAMS:
If you're on a recent version of SuSE/UL, this parameter has
moved into /etc/sysconfig/syslog. You'll have to restart syslog
after putting these parameters into effect.
SYSLOGD_PARAMS="-r"
4) Change on the testmonitor-machine to redirect messages
from local7 into /var/log/ha-log-local7 by adding this line to
/etc/syslog.conf
local7.* -/var/log/ha-log-local7
and then (on SuSE) run this command:
/etc/rc.d/syslog restart
Use the corresponding function for your distro.
--------------------------------------------------------------------------------
How to make OpenSSH allow you to login as root across the network without
a password.
--------------------------------------------------------------------------------
All our scripts run ssh -l root, so you don't have to do any of your testing
logged in as root on the test machine
1) Grab your key from the testmonitor-machine:
take the single line out of ~/.ssh/identity.pub
and put it into root's authorized_keys file.
[This has changed to: copying the line from ~/.ssh/id_dsa.pub into
root's authorized_keys file ]
Run this command on each of the "test" machines as root:
ssh -v -l myid testmonitor-machine cat /home/myid/.ssh/identity.pub \
>> ~root/.ssh/authorized_keys
[For most people, this has changed to:
ssh -v -l myid testmonitor-machine cat /home/myid/.ssh/id_dsa.pub \
>> ~root/.ssh/authorized_keys
]
You will probably have to provide your password, and possibly say
"yes" to some questions about accepting the identity of the
test machines
You must also do the corresponding update for the testmonitor machine itself
as root:
cat /home/myid/.ssh/identity.pub >> ~root/.ssh/authorized_keys
To test this, try this command from the testmonitor-machine for each
of your testmachines, and for the testmonitor-machine itself.
ssh -l root othermachine
If this works, without prompting for a password, you're in business...
If not, you need to look at the ssh/openssh documentation and the output from
the -v options above...
--------------------------------------------------------------------------------
How to redirect linux-HA logging the way CTS wants it using syslog-ng
--------------------------------------------------------------------------------
why syslong-ng:
Syslog-ng can use tcp and guarantee availability of logs. It is important
we get every log message or our test may fail with lost of some important messages.
The following instructions apply to RedHat systems:
1) Redirect each machines to go (at least) to syslog local7:
Change /etc/ha.d/ha.cf on each test machine to say this:
logfacility local7
2) On all machines: download syslog-ng rpm and install it.
Or you can download its source at http://www.balabit.com/products/syslog_ng/,
compile and install it.
3) On all machines: stop syslog and disable it on reboot
$ /etc/init.d/syslog stop
$ chkconfig --level 2345 syslog off
4) In testmonitor-machine, add the source and destination for remote log in the
file /etc/syslog-ng/syslong-ng.conf. You can change the port number to
other number, but it must be the same with the port number in slave machines
(see 4)
source s_tcp { tcp(port(15789) max-connections(10)); };
destination d_ha { file("/var/log/ha-log-local7"); };
log { source(s_tcp); destination(d_ha); };
5) In slave machines, send log out to a remote machine
add the following lines to /etc/syslog-ng/syslong-ng.conf
after the definition of f_boot.
Note the port number must be the same with 3)
Change the ip to the testmonitor machine's ip
filter f_boot { facility(local7); }; /* this line is already in file*/
destination d_tcp { tcp("192.168.0.3" port(15789));};
log { source(s_sys); filter(f_boot); destination(d_tcp); };
6) start syslog-ng and enable it in reboot in all machines
$/etc/init.d/syslog-ng start
$chkconfig --level 2345 syslog-ng on
--------------------------------------------------------------------------------
How to redirect linux-HA logging CTS wants with evlog
--------------------------------------------------------------------------------
Related background introduction
evlog is a new logging system. It's open source, and its source/binary is
licensed under GPL/LGPL. its web site is as below.
http://evlog.sourceforge.net/
evlog is compliance with draft POSIX Standard 1003.25. It can provide more
advanced logging capacities (please refer to its web site for more details).
Among its serveral important features, when comparing with syslog, the remote
logging with tcp protocol is preferred here.
Why? Because when testing linux-ha as described above, you may have to need
remote logging support. Of course you can use syslog to get it via suitable
setting as the steps descibed above. But, syslog itself only supports remote
logging with udp protocol. As you know, sometimes udp protocol is not reliable
enough, especially under heavy workload, may lose some udp packages, cause
cts' log to become mess and difficult to analyse. Evlog is a good way to
resolve this issue.
Briefly, we can locally forward syslog message to evlog, then continue
forwarding the log message to remote machine with evlog's tcp remote logging
capacity. This don't require to rewrite related applications, such as heartbeat.
It's a big advantage for us. Since by default evlog isn't configurated to
support tcp remote logging, so need to configure it. The following is the brief
steps. Some of them are abstracted from evlog documents.
1) Get the evlog, build binary if needed.
-----------------------------------------
You can download evlog binary or source from evlog project page:
http://sourceforge.net/projects/evlog
Some linux distributions, such as SLES, includes evlog, but normally it
doesn't contain remote tcp logging module named as tcp_rmtlog_be. So you may
need to get additional package from there.
If luck you can get the suitable binary packages for your system from there.
As for rpm package, you need two packages as below.
evlog -- Standard package, including most functions
tcp_rmtlog_be -- Module to support remote tcp logging
If you have to build binary for yourself, the simple steps is as below.
Here suppose you begin from evlog-1.5.3 tarball.
a. Log in as root
b. Download evlog-1.5.3.tar.gz
c. Untar the tarball
tar -xzvf evlog-1.5.3.tar.gz
d. cd to evlog-1.5.3
e. To run configuration scripts.
./autogen.sh
./configure
f. Build and install.
Normal way.
make
make install
make startall
Or
Build rpm do the following:
make rpm
make rpm-tcp
make rpm-udp
Then you can see the evlog and tcp_rmtlog_be in top build directory.
you can install them with rpm command.
When install is successful, you will see messages like these...
/etc/rc.d/init.d/evlog start
Starting enterprise event logger: [ OK ]
sleep 1
/etc/init.d/evlogrmt start
Starting remote event logger: [ OK ]
sleep 1
/etc/rc.d/init.d/evlnotify start
Starting enterprise event log notification: [ OK ]
sleep 1
/etc/rc.d/init.d/evlaction start
Starting notification action daemon: [ OK ]
2) Configure remote event consolidator, which normally run CTS test scripts.
----------------------------------------------------------------------------
This procedure configures the evlogrmtd daemon to accept events from other two
hosts running heartbeat testing hosts on the network. So that events from
multiple hosts can be consolidated into a single log file.
a. Log in as root
b. Edit /etc/evlog.d/evlhosts to add an entry for each of two testing host that
run heartbeat. Each entry must specify host name, either simple name or
fqdn, and also a unique identifier for each host. This identifier can be up
to 2 bytes, but cannot be equal to 0 (it will be ignored).
For example, the following are all valid entries:
(identifier) (hostname)
1 hatest1
2 hatest2
c. There is also a configuration file, /etc/evlog.d/evlogrmtd.conf which
contains the following as default:
Password=password
TCPPort=12000
"Password" is used only by TCP clients to authenticate remote hosts when
attempting to connect.
"TCPPort" must match the TCP port used by other two test machines for
sending events to the event consolidator.
d. Restart the evlogrmtd daemon...
/etc/init.d/evlogrmt restart
If evlogrmtd cannot resolve any of the hosts listed in evlhosts, or there
are no entries in /etc/evlog.d/evlhosts, then the evlogmrtd will exit.
3) Configure the two test machine on which hearbeat will run.
-------------------------------------------------------------
This procedure installs and configures an event tcp logging plugin for
forwarding events to a remote event consolidator.
The local logging software must be installed.
a. Log in as root.
b. If have installed tcp_rmtlog_be, then skip to next step. Or execute the
following command (shown for i386 rpm):
rpm -i tcp_rmtlog_be-1.5.3-1.i386.rpm
c. cd to /etc/evlog.d/plugins, then edit tcp_rmtlog_be.conf.
you need to specify the following items.
IP address, or hostname - for the event consolidator.
Port number - should match the port number used by the event consolidator.
Disable=no - to send events using TCP
Password - must match password expected by the event consolidator when the
TCP connection is attempted.
BufferLenInKbytes - Specifies the size of the memory buffer for events being
transmitted via TCP. This reduces the chances of losing events during
temporary loss of connection. Default size=128. Recommended range = 32
to 1024.
A sample tcp_rmtlog_be.conf may like as below.
Remote Host=172.30.1.180
Password=password
Port=12000
BufferLenInKbytes=128
Disable=no
d. Restart the evlogd daemon to load the plugin...
/etc/init.d/evlog restart
4) Configure syslog on the pair of HA machines.
-----------------------------------------------
For forwarding syslog messages to the evlog on the same machine. Issue this
command, which is from evlog package.
/sbin/slog_fwd
This will forward syslog messages immediately, and after every subsequent
reboot. To disable syslog forwarding:
/sbin/slog_fwd -r
5) Test your configure work.
----------------------------
For example, on you one of the pair of HA machines, issue this command:
logger -p local7.info "logging hello from hatest1"
Then go to remote event consolidator, which run CTS test scripts. issue this
command, which is from evlog package.
evlview -m | grep hatest1
you should see its result.
Apr 7 13:32:04 hadev1 logging hello from hatest1
Notes, by default event log message of evlog is store in file
/var/evlog/eventlog
It's a file containing binary structure messages, so you should use evlview
to read them.
Enjoy evlog ;). The end.
+
+--------------------------------------------------------------------------------
+How to configure release 2.*(1.99.*) for cts testing --draft
+--------------------------------------------------------------------------------
+To test release 2.X(1.99.*), we need do following additional work.
+
+1. Please build the source code with --enable-crm option and install it.
+
+2. Configure the ha.cf, and make it same on the test exerciser and all test
+cluster machines. Please add following lines to ha.cf:
+ apiauth default uid=root
+ apiauth ccm uid=hacluster
+ respawn root /usr/lib/heartbeat/lrmd
+ respawn hacluster /usr/lib/heartbeat/ccm
+ respawn root /usr/lib/heartbeat/crmd -VVVV
+
+3. Add resources and constraints to the cluster:
+ 3.1 Start up the heartbeat on all test machines:
+ /usr/lib/heartbeat/heartbeat -M
+
+ 3.2 Wait for the heartbeat start complete:
+ Using "/usr/lib/crmadmin -S node" on one test machine to get the states
+ of nodes. Wait for all test machines are in S_IDLE or S_NOT_DC state.
+
+ 3.3 Add resources and constraints to cluster.
+ Using "/usr/lib/cibadmin" to add resources and constraints.
+ Please notice you need do this only once on one test node.
+ For examples,
+
+ To add a constraint:
+ /usr/lib/heartbeat/cibadmin -C -o constraint -X \
+ '<rsc_location id="ad66bef2-5a34-49b2-85a9-3fc6dc69e6fb" rsc="rsc1"> \
+ <rule id="2deac01b-35ee-4148-97ab-17970fb710df" result="can"/> \
+ <rule id="84bc3cb2-fcab-4297-bd39-e6587c492fb7" score="1000" boolean_op="or"> \
+ <expression attribute="uname" operation="eq" value="hadev1"/> \
+ </rule> \
+ </rsc_location>'
+
+ To add a resource:
+ /usr/lib/heartbeat/cibadmin -C -o resource -X \
+ '<resource id="rsc1" class="heartbeat" type="IPaddr" > \
+ <instance_attributes> \
+ <rsc_parameters> \
+ <nvpair name="1" value="172.30.1.241"/> \
+ </rsc_parameters> \
+ </instance_attributes> \
+ </resource>'
+
+ 3.4 Shutdown the heartbeat on all test machines:
+ /usr/lib/heartbeat/heartbeat -k
+
+4. Now we can tart cts test on the exerciser node:
+ ./CTSlab.py -v2 number-of-tests-to-run
+ or as
+ python ./CTSlab.py -v2 number-of-tests-to-run
File Metadata
Details
Attached
Mime Type
text/x-diff
Expires
Sat, Nov 23, 11:58 AM (22 h, 27 m)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
1018696
Default Alt Text
(19 KB)
Attached To
Mode
rP Pacemaker
Attached
Detach File
Event Timeline
Log In to Comment