diff --git a/.gitignore b/.gitignore index e2b7c039c..0ac9c6b76 100644 --- a/.gitignore +++ b/.gitignore @@ -1,156 +1,157 @@ *.swp Makefile.in aclocal.m4 autoconf autoheader autom4te.cache automake autoscan.log compile configure configure.scan config.guess config.log config.sub config.status Makefile depcomp install-sh libtoolize ltmain.sh libtool make/stamp-h1 make/clusterautoconfig.h* missing resource-agents.spec *.pc .deps .libs *.o *.la *.lo *.loT rgmanager/src/resources/* !rgmanager/src/resources/*.in rgmanager/src/resources/Makefile.in rgmanager/src/resources/utils/config-utils.sh resource-agents-* .version # generated by ./autogen.sh && ./configure doc/man/*.7 doc/man/*.xml heartbeat/ocf-binaries heartbeat/ocf-directories heartbeat/ocf-shellfuncs heartbeat/send_ua heartbeat/shellfuncs heartbeat/*.pyc heartbeat/AoEtarget heartbeat/CTDB heartbeat/ManageRAID heartbeat/ManageVE heartbeat/Squid heartbeat/SysInfo heartbeat/aws-vpc-route53 heartbeat/azure-events heartbeat/azure-events-az heartbeat/clvm heartbeat/conntrackd heartbeat/dnsupdate heartbeat/dummypy heartbeat/eDir88 heartbeat/fio heartbeat/galera heartbeat/gcp-pd-move heartbeat/gcp-vpc-move-ip heartbeat/gcp-vpc-move-route heartbeat/gcp-vpc-move-vip heartbeat/iSCSILogicalUnit heartbeat/iSCSITarget heartbeat/jira heartbeat/kamailio heartbeat/lxc heartbeat/lxd-info heartbeat/machine-info heartbeat/mariadb heartbeat/mpathpersist heartbeat/nfsnotify heartbeat/openstack-info +heartbeat/powervs-subnet heartbeat/rabbitmq-cluster heartbeat/redis heartbeat/rsyslog heartbeat/sg_persist heartbeat/slapd heartbeat/smb-share heartbeat/storage-mon heartbeat/sybaseASE heartbeat/syslog-ng heartbeat/vsftpd include/agent_config.h include/config.h include/config.h.in include/stamp-h1 include/stamp-h2 ldirectord/ldirectord ldirectord/ldirectord.8 ldirectord/OCF/ldirectord ldirectord/init.d/ldirectord ldirectord/init.d/ldirectord.debian ldirectord/init.d/ldirectord.debian.default ldirectord/systemd/ldirectord.service systemd/resource-agents.conf tools/findif tools/nfsconvert tools/ocf-tester tools/send_arp tools/storage_mon tools/tickle_tcp tools/ocft/README tools/ocft/README.zh_CN tools/ocft/caselib tools/ocft/ocft *.cache *.upgrade.xml py-compile ylwrap __pycache__ # BEAM Entries *.beam parser-messages MISC_ERRORS cscope.files cscope.out patches updates logs # OS and Editor Artifacts .DS_Store .bomb *.rej *.bz2 *.gz *.xz *.sed *.diff *.patch *.gres *~ # Misc HTML TAGS GPATH GRTAGS GSYMS GTAGS .gres.* *.orig .gdb_history *~ \#* .changes pacemaker.tar.gz diff --git a/configure.ac b/configure.ac index e05de3a0e..002b34dce 100644 --- a/configure.ac +++ b/configure.ac @@ -1,1085 +1,1092 @@ dnl dnl autoconf for Agents dnl dnl License: GNU General Public License (GPL) dnl =============================================== dnl Bootstrap dnl =============================================== AC_PREREQ(2.63) dnl Suggested structure: dnl information on the package dnl checks for programs dnl checks for libraries dnl checks for header files dnl checks for types dnl checks for structures dnl checks for compiler characteristics dnl checks for library functions dnl checks for system services AC_INIT([resource-agents], m4_esyscmd([make/git-version-gen .tarball-version]), [developers@clusterlabs.org]) AC_USE_SYSTEM_EXTENSIONS CRM_DTD_VERSION="1.0" AC_CONFIG_AUX_DIR(.) AC_CONFIG_MACRO_DIR([m4]) AC_CANONICAL_HOST dnl Where #defines go (e.g. `AC_CHECK_HEADERS' below) dnl dnl Internal header: include/config.h dnl - Contains ALL defines dnl - include/config.h.in is generated automatically by autoheader dnl - NOT to be included in any header files except lha_internal.h dnl (which is also not to be included in any other header files) dnl dnl External header: include/agent_config.h dnl - Contains a subset of defines checked here dnl - Manually edit include/agent_config.h.in to have configure include new defines dnl - Should not include HAVE_* defines dnl - Safe to include anywhere AM_CONFIG_HEADER(include/config.h include/agent_config.h) ALL_LINGUAS="en fr" AC_ARG_WITH(version, [ --with-version=version Override package version (if you're a packager needing to pretend) ], [ PACKAGE_VERSION="$withval" ]) AC_ARG_WITH(pkg-name, [ --with-pkg-name=name Override package name (if you're a packager needing to pretend) ], [ PACKAGE_NAME="$withval" ]) dnl dnl AM_INIT_AUTOMAKE([1.11.1 foreign dist-bzip2 dist-xz]) dnl AM_INIT_AUTOMAKE([1.10.1 foreign dist-bzip2]) AC_DEFINE_UNQUOTED(AGENTS_VERSION, "$PACKAGE_VERSION", Current agents version) CC_IN_CONFIGURE=yes export CC_IN_CONFIGURE LDD=ldd dnl ======================================================================== dnl Compiler characteristics dnl ======================================================================== # check stolen from gnulib/m4/gnu-make.m4 if ! ${MAKE-make} --version /cannot/make/this >/dev/null 2>&1; then AC_MSG_ERROR([you don't seem to have GNU make; it is required]) fi AC_PROG_CC dnl Can force other with environment variable "CC". AM_PROG_CC_C_O AC_PROG_CC_STDC AC_PROG_CPP AC_PROG_AWK AC_PROG_LN_S AC_PROG_INSTALL AC_PROG_MAKE_SET AC_C_STRINGIZE AC_C_INLINE AC_TYPE_SIZE_T AC_TYPE_SSIZE_T AC_TYPE_UID_T AC_TYPE_UINT16_T AC_TYPE_UINT8_T AC_TYPE_UINT32_T AC_CHECK_SIZEOF(char) AC_CHECK_SIZEOF(short) AC_CHECK_SIZEOF(int) AC_CHECK_SIZEOF(long) AC_CHECK_SIZEOF(long long) AC_STRUCT_TIMEZONE dnl =============================================== dnl Helpers dnl =============================================== cc_supports_flag() { local CPPFLAGS="$@" AC_MSG_CHECKING(whether $CC supports "$@") AC_PREPROC_IFELSE([AC_LANG_PROGRAM([])], [RC=0; AC_MSG_RESULT([yes])], [RC=1; AC_MSG_RESULT([no])]) return $RC } extract_header_define() { AC_MSG_CHECKING(for $2 in $1) Cfile=$srcdir/extract_define.$2.${$} printf "#include \n" > ${Cfile}.c printf "#include <%s>\n" $1 >> ${Cfile}.c printf "int main(int argc, char **argv) { printf(\"%%s\", %s); return 0; }\n" $2 >> ${Cfile}.c $CC $CFLAGS ${Cfile}.c -o ${Cfile} value=`${Cfile}` AC_MSG_RESULT($value) printf $value rm -f ${Cfile}.c ${Cfile} } AC_MSG_NOTICE(Sanitizing prefix: ${prefix}) case $prefix in NONE) prefix=/usr dnl Fix default variables - "prefix" variable if not specified if test "$localstatedir" = "\${prefix}/var"; then localstatedir="/var" fi if test "$sysconfdir" = "\${prefix}/etc"; then sysconfdir="/etc" fi ;; esac # ordering is important, PKG_PROG_PKG_CONFIG is to be invoked before any other PKG_* related stuff PKG_PROG_PKG_CONFIG(0.18) # PKG_CHECK_MODULES will fail if systemd is not found by default, so make sure # we set the proper vars and deal with it PKG_CHECK_MODULES([systemd], [systemd], [HAS_SYSTEMD=yes], [HAS_SYSTEMD=no]) if test "x$HAS_SYSTEMD" = "xyes"; then PKG_CHECK_VAR([SYSTEMD_UNIT_DIR], [systemd], [systemdsystemunitdir]) if test "x$SYSTEMD_UNIT_DIR" = "x"; then AC_MSG_ERROR([Unable to detect systemd unit dir automatically]) fi PKG_CHECK_VAR([SYSTEMD_TMPFILES_DIR], [systemd], [tmpfilesdir]) if test "x$SYSTEMD_TMPFILES_DIR" = "x"; then AC_MSG_ERROR([Unable to detect systemd tmpfiles directory automatically]) fi # sanitize systed vars when using non standard prefix if test "$prefix" != "/usr"; then SYSTEMD_UNIT_DIR="$prefix/$SYSTEMD_UNIT_DIR" AC_SUBST([SYSTEMD_UNIT_DIR]) SYSTEMD_TMPFILES_DIR="$prefix/$SYSTEMD_TMPFILES_DIR" AC_SUBST([SYSTEMD_TMPFILES_DIR]) fi fi AM_CONDITIONAL(HAVE_SYSTEMD, [test "x$HAS_SYSTEMD" = xyes ]) dnl =============================================== dnl Configure Options dnl =============================================== dnl Some systems, like Solaris require a custom package name AC_ARG_WITH(pkgname, [ --with-pkgname=name name for pkg (typically for Solaris) ], [ PKGNAME="$withval" ], [ PKGNAME="LXHAhb" ], ) AC_SUBST(PKGNAME) AC_ARG_ENABLE([ansi], [ --enable-ansi force GCC to compile to ANSI/ANSI standard for older compilers. [default=no]]) AC_ARG_ENABLE([fatal-warnings], [ --enable-fatal-warnings very pedantic and fatal warnings for gcc [default=yes]]) INITDIR="" AC_ARG_WITH(initdir, [ --with-initdir=DIR directory for init (rc) scripts [${INITDIR}]], [ INITDIR="$withval" ]) OCF_ROOT_DIR="${prefix}/lib/ocf" AC_ARG_WITH(ocf-root, [ --with-ocf-root=DIR directory for OCF scripts [${OCF_ROOT_DIR}]], [ OCF_ROOT_DIR="$withval" ]) HA_RSCTMPDIR=${localstatedir}/run/resource-agents AC_ARG_WITH(rsctmpdir, [ --with-rsctmpdir=DIR directory for resource agents state files [${HA_RSCTMPDIR}]], [ HA_RSCTMPDIR="$withval" ]) AC_ARG_ENABLE([libnet], [ --enable-libnet Use libnet for ARP based functionality, [default=try]], [enable_libnet="$enableval"], [enable_libnet=try]) BUILD_RGMANAGER=0 BUILD_LINUX_HA=0 RASSET=linux-ha AC_ARG_WITH(ras-set, [ --with-ras-set=SET build/install only linux-ha, rgmanager or all resource-agents [default: linux-ha]], [ RASSET="$withval" ]) if test x$RASSET = xyes || test x$RASSET = xall ; then BUILD_RGMANAGER=1 BUILD_LINUX_HA=1 fi if test x$RASSET = xlinux-ha; then BUILD_LINUX_HA=1 fi if test x$RASSET = xrgmanager; then BUILD_RGMANAGER=1 fi if test $BUILD_LINUX_HA -eq 0 && test $BUILD_RGMANAGER -eq 0; then AC_MSG_ERROR([Are you really sure you want this package?]) exit 1 fi AM_CONDITIONAL(BUILD_LINUX_HA, test $BUILD_LINUX_HA -eq 1) AM_CONDITIONAL(BUILD_RGMANAGER, test $BUILD_RGMANAGER -eq 1) AC_ARG_WITH(compat-habindir, [ --with-compat-habindir use HA_BIN directory with compatibility for the Heartbeat stack [${libexecdir}]], [], [with_compat_habindir=no]) AM_CONDITIONAL(WITH_COMPAT_HABINDIR, test "x$with_compat_habindir" != "xno") dnl =============================================== dnl General Processing dnl =============================================== echo Our Host OS: $host_os/$host AC_MSG_NOTICE(Sanitizing exec_prefix: ${exec_prefix}) case $exec_prefix in dnl For consistency with Heartbeat, map NONE->$prefix NONE) exec_prefix=$prefix;; prefix) exec_prefix=$prefix;; esac AC_MSG_NOTICE(Sanitizing INITDIR: ${INITDIR}) case $INITDIR in prefix) INITDIR=$prefix;; "") AC_MSG_CHECKING(which init (rc) directory to use) for initdir in /etc/init.d /etc/rc.d/init.d /sbin/init.d \ /usr/local/etc/rc.d /etc/rc.d do if test -d $initdir then INITDIR=$initdir break fi done if test -z $INITDIR then INITDIR=${sysconfdir}/init.d fi AC_MSG_RESULT($INITDIR);; esac AC_SUBST(INITDIR) if test "${prefix}" = "/usr"; then INITDIRPREFIX="$INITDIR" else INITDIRPREFIX="${prefix}/$INITDIR" fi AC_SUBST(INITDIRPREFIX) AC_MSG_NOTICE(Sanitizing libdir: ${libdir}) case $libdir in dnl For consistency with Heartbeat, map NONE->$prefix *prefix*|NONE) AC_MSG_CHECKING(which lib directory to use) for aDir in lib64 lib do trydir="${exec_prefix}/${aDir}" if test -d ${trydir} then libdir=${trydir} break fi done AC_MSG_RESULT($libdir); ;; esac if test "x$with_compat_habindir" != "xno" ; then libexecdir=${libdir} fi dnl Expand autoconf variables so that we dont end up with '${prefix}' dnl in #defines and python scripts dnl NOTE: Autoconf deliberately leaves them unexpanded to allow dnl make exec_prefix=/foo install dnl No longer being able to do this seems like no great loss to me... eval prefix="`eval echo ${prefix}`" eval exec_prefix="`eval echo ${exec_prefix}`" eval bindir="`eval echo ${bindir}`" eval sbindir="`eval echo ${sbindir}`" eval libexecdir="`eval echo ${libexecdir}`" eval datadir="`eval echo ${datadir}`" eval sysconfdir="`eval echo ${sysconfdir}`" eval sharedstatedir="`eval echo ${sharedstatedir}`" eval localstatedir="`eval echo ${localstatedir}`" eval libdir="`eval echo ${libdir}`" eval includedir="`eval echo ${includedir}`" eval oldincludedir="`eval echo ${oldincludedir}`" eval infodir="`eval echo ${infodir}`" eval mandir="`eval echo ${mandir}`" dnl docdir is a recent addition to autotools eval docdir="`eval echo ${docdir}`" if test "x$docdir" = "x"; then docdir="`eval echo ${datadir}/doc`" fi AC_SUBST(docdir) dnl Home-grown variables eval INITDIR="${INITDIR}" for j in prefix exec_prefix bindir sbindir libexecdir datadir sysconfdir \ sharedstatedir localstatedir libdir includedir oldincludedir infodir \ mandir INITDIR docdir do dirname=`eval echo '${'${j}'}'` if test ! -d "$dirname" then AC_MSG_WARN([$j directory ($dirname) does not exist!]) fi done dnl This OS-based decision-making is poor autotools practice; dnl feature-based mechanisms are strongly preferred. dnl dnl So keep this section to a bare minimum; regard as a "necessary evil". REBOOT_OPTIONS="-f" POWEROFF_OPTIONS="-f" case "$host_os" in *bsd*) LIBS="-L/usr/local/lib" CPPFLAGS="$CPPFLAGS -I/usr/local/include" ;; *solaris*) REBOOT_OPTIONS="-n" POWEROFF_OPTIONS="-n" LDFLAGS+=" -lssp -lssp_nonshared" ;; *linux*) AC_DEFINE_UNQUOTED(ON_LINUX, 1, Compiling for Linux platform) POWEROFF_OPTIONS="-nf" REBOOT_OPTIONS="-nf" ;; darwin*) AC_DEFINE_UNQUOTED(ON_DARWIN, 1, Compiling for Darwin platform) LIBS="$LIBS -L${prefix}/lib" CFLAGS="$CFLAGS -I${prefix}/include" ;; esac AC_DEFINE_UNQUOTED(HA_LOG_FACILITY, LOG_DAEMON, Default logging facility) AC_MSG_NOTICE(Host CPU: $host_cpu) case "$host_cpu" in ppc64|powerpc64) case $CFLAGS in *powerpc64*) ;; *) if test "$GCC" = yes; then CFLAGS="$CFLAGS -m64" fi ;; esac esac AC_MSG_CHECKING(which format is needed to print uint64_t) case "$host_cpu" in s390x)U64T="%lu";; *64*) U64T="%lu";; *) U64T="%llu";; esac AC_MSG_RESULT($U64T) AC_DEFINE_UNQUOTED(U64T, "$U64T", Correct printf format for logging uint64_t) dnl Variables needed for substitution AC_CHECK_HEADERS(heartbeat/glue_config.h) if test "$ac_cv_header_heartbeat_glue_config_h" != "yes"; then enable_libnet=no fi AC_DEFINE_UNQUOTED(OCF_ROOT_DIR,"$OCF_ROOT_DIR", OCF root directory - specified by the OCF standard) AC_SUBST(OCF_ROOT_DIR) GLUE_STATE_DIR=${localstatedir}/run AC_DEFINE_UNQUOTED(GLUE_STATE_DIR,"$GLUE_STATE_DIR", Where to keep state files and sockets) AC_SUBST(GLUE_STATE_DIR) AC_DEFINE_UNQUOTED(HA_VARRUNDIR,"$GLUE_STATE_DIR", Where Heartbeat keeps state files and sockets - old name) HA_VARRUNDIR="$GLUE_STATE_DIR" AC_SUBST(HA_VARRUNDIR) # Expand $prefix eval HA_RSCTMPDIR="`eval echo ${HA_RSCTMPDIR}`" AC_DEFINE_UNQUOTED(HA_RSCTMPDIR,"$HA_RSCTMPDIR", Where Resource agents keep state files) AC_SUBST(HA_RSCTMPDIR) dnl Eventually move out of the heartbeat dir tree and create symlinks when needed HA_VARLIBHBDIR=${localstatedir}/lib/heartbeat AC_DEFINE_UNQUOTED(HA_VARLIBHBDIR,"$HA_VARLIBHBDIR", Whatever this used to mean) AC_SUBST(HA_VARLIBHBDIR) OCF_RA_DIR="${OCF_ROOT_DIR}/resource.d" AC_DEFINE_UNQUOTED(OCF_RA_DIR,"$OCF_RA_DIR", Location for OCF RAs) AC_SUBST(OCF_RA_DIR) OCF_RA_DIR_PREFIX="$OCF_RA_DIR" AC_SUBST(OCF_RA_DIR_PREFIX) OCF_LIB_DIR="${OCF_ROOT_DIR}/lib" AC_DEFINE_UNQUOTED(OCF_LIB_DIR,"$OCF_LIB_DIR", Location for shared code for OCF RAs) AC_SUBST(OCF_LIB_DIR) OCF_LIB_DIR_PREFIX="$OCF_LIB_DIR" AC_SUBST(OCF_LIB_DIR_PREFIX) dnl =============================================== dnl rgmanager ras bits dnl =============================================== LOGDIR=${localstatedir}/log/cluster CLUSTERDATA=${datadir}/cluster AC_SUBST([LOGDIR]) AC_SUBST([CLUSTERDATA]) dnl =============================================== dnl Program Paths dnl =============================================== PATH="$PATH:/sbin:/usr/sbin:/usr/local/sbin:/usr/local/bin" export PATH AC_CHECK_PROGS(MAKE, gmake make) AC_CHECK_PROGS(SHELLCHECK, shellcheck) AM_CONDITIONAL(CI_CHECKS, test "x$SHELLCHECK" != "x" ) AC_PATH_PROGS(BASH_SHELL, bash) if test x"${BASH_SHELL}" = x""; then AC_MSG_ERROR(You need bash installed in order to build ${PACKAGE}) fi AC_PATH_PROGS(XSLTPROC, xsltproc) AM_CONDITIONAL(BUILD_DOC, test "x$XSLTPROC" != "x" ) if test "x$XSLTPROC" = "x"; then AC_MSG_WARN([xsltproc not installed, unable to (re-)build manual pages]) fi AC_SUBST(XSLTPROC) AC_PATH_PROGS(XMLCATALOG, xmlcatalog) AC_PATH_PROGS(SSH, ssh, /usr/bin/ssh) AC_PATH_PROGS(SCP, scp, /usr/bin/scp) AC_PATH_PROGS(TAR, tar) AC_PATH_PROGS(MD5, md5) AC_PATH_PROGS(TEST, test) AC_PATH_PROGS(PING, ping, /bin/ping) AC_PATH_PROGS(IFCONFIG, ifconfig, /sbin/ifconfig) AC_PATH_PROGS(MAILCMD, mailx mail, mail) AC_PATH_PROGS(EGREP, egrep) AC_PATH_PROGS(RM, rm) AC_SUBST(BASH_SHELL) AC_SUBST(MAILCMD) AC_SUBST(EGREP) AC_SUBST(SHELL) AC_SUBST(PING) AC_SUBST(RM) AC_SUBST(TEST) +AM_PATH_PYTHON([3.6]) +if test -z "$PYTHON"; then + echo "*** Essential program python not found" 1>&2 + exit 1 +fi + dnl Ensure PYTHON is an absolute path AC_PATH_PROG([PYTHON], [$PYTHON]) AM_PATH_PYTHON if test -z "$PYTHON"; then echo "*** Essential program python not found" 1>&2 fi AC_PYTHON_MODULE(json) AC_PYTHON_MODULE(pyroute2) -AS_VERSION_COMPARE([$PYTHON_VERSION], [2.7], [BUILD_OCF_PY=0], [BUILD_OCF_PY=1], [BUILD_OCF_PY=1]) +AS_VERSION_COMPARE([$PYTHON_VERSION], [3.6], [BUILD_OCF_PY=0], [BUILD_OCF_PY=1], [BUILD_OCF_PY=1]) BUILD_AZURE_EVENTS=1 if test -z "$PYTHON" || test $BUILD_OCF_PY -eq 0; then BUILD_AZURE_EVENTS=0 AC_MSG_WARN("Not building azure-events") fi AM_CONDITIONAL(BUILD_AZURE_EVENTS, test $BUILD_AZURE_EVENTS -eq 1) BUILD_AZURE_EVENTS_AZ=1 if test -z "$PYTHON" || test $BUILD_OCF_PY -eq 0; then BUILD_AZURE_EVENTS_AZ=0 AC_MSG_WARN("Not building azure-events-az") fi AM_CONDITIONAL(BUILD_AZURE_EVENTS_AZ, test $BUILD_AZURE_EVENTS_AZ -eq 1) BUILD_GCP_PD_MOVE=1 if test -z "$PYTHON" || test $BUILD_OCF_PY -eq 0; then BUILD_GCP_PD_MOVE=0 AC_MSG_WARN("Not building gcp-pd-move") fi AM_CONDITIONAL(BUILD_GCP_PD_MOVE, test $BUILD_GCP_PD_MOVE -eq 1) BUILD_GCP_VPC_MOVE_ROUTE=1 if test -z "$PYTHON" || test "x${HAVE_PYMOD_PYROUTE2}" != xyes || test $BUILD_OCF_PY -eq 0; then BUILD_GCP_VPC_MOVE_ROUTE=0 AC_MSG_WARN("Not building gcp-vpc-move-route") fi AM_CONDITIONAL(BUILD_GCP_VPC_MOVE_ROUTE, test $BUILD_GCP_VPC_MOVE_ROUTE -eq 1) BUILD_GCP_VPC_MOVE_VIP=1 if test -z "$PYTHON" || test $BUILD_OCF_PY -eq 0; then BUILD_GCP_VPC_MOVE_VIP=0 AC_MSG_WARN("Not building gcp-vpc-move-vip") fi AM_CONDITIONAL(BUILD_GCP_VPC_MOVE_VIP, test $BUILD_GCP_VPC_MOVE_VIP -eq 1) AC_PATH_PROGS(ROUTE, route) AC_DEFINE_UNQUOTED(ROUTE, "$ROUTE", path to route command) AC_MSG_CHECKING(ifconfig option to list interfaces) for IFCONFIG_A_OPT in "-A" "-a" "" do $IFCONFIG $IFCONFIG_A_OPT > /dev/null 2>&1 if test "$?" = 0 then AC_DEFINE_UNQUOTED(IFCONFIG_A_OPT, "$IFCONFIG_A_OPT", option for ifconfig command) AC_MSG_RESULT($IFCONFIG_A_OPT) break fi done AC_SUBST(IFCONFIG_A_OPT) if test x"${MAKE}" = x""; then AC_MSG_ERROR(You need (g)make installed in order to build ${PACKAGE}) fi STYLESHEET_PREFIX="" if test x"${XSLTPROC}" != x""; then AC_MSG_CHECKING(docbook to manpage transform) # first try to figure out correct template using xmlcatalog query, # resort to extensive (semi-deterministic) file search if that fails DOCBOOK_XSL_URI='http://docbook.sourceforge.net/release/xsl/current' DOCBOOK_XSL_PATH='manpages/docbook.xsl' STYLESHEET_PREFIX=$(${XMLCATALOG} "" ${DOCBOOK_XSL_URI} \ | sed -n 's|^file://||p;q') if test x"${STYLESHEET_PREFIX}" = x""; then DIRS=$(find "${datadir}" -name $(basename $(dirname ${DOCBOOK_XSL_PATH})) \ -type d | LC_ALL=C sort) if test x"${DIRS}" = x""; then # when datadir is not standard OS path, we cannot find docbook.xsl # use standard OS path as backup DIRS=$(find "/usr/share" "/usr/local/share" -name $(basename $(dirname ${DOCBOOK_XSL_PATH})) \ -type d | LC_ALL=C sort) fi XSLT=$(basename ${DOCBOOK_XSL_PATH}) for d in ${DIRS}; do if test -f "${d}/${XSLT}"; then STYLESHEET_PREFIX=$(echo "${d}" | sed 's/\/manpages//') break fi done fi if test x"${STYLESHEET_PREFIX}" = x""; then AC_MSG_ERROR(You need docbook-style-xsl installed in order to build ${PACKAGE}) fi fi AC_MSG_RESULT($STYLESHEET_PREFIX) AC_SUBST(STYLESHEET_PREFIX) dnl =============================================== dnl Libraries dnl =============================================== AC_CHECK_LIB(socket, socket) AC_CHECK_LIB(gnugetopt, getopt_long) dnl if available if test "x${enable_thread_safe}" = "xyes"; then GPKGNAME="gthread-2.0" else GPKGNAME="glib-2.0" fi PKG_CHECK_MODULES([GLIB], [$GPKGNAME]) CPPFLAGS="$CPPFLAGS $GLIB_CFLAGS" LIBS="$LIBS $GLIB_LIBS" PKG_CHECK_MODULES([LIBQB], "libqb") dnl ======================================================================== dnl Headers dnl ======================================================================== AC_HEADER_STDC AC_CHECK_HEADERS(sys/socket.h) AC_CHECK_HEADERS(sys/sockio.h) AC_CHECK_HEADERS([arpa/inet.h]) AC_CHECK_HEADERS([fcntl.h]) AC_CHECK_HEADERS([limits.h]) AC_CHECK_HEADERS([malloc.h]) AC_CHECK_HEADERS([netdb.h]) AC_CHECK_HEADERS([netinet/in.h]) AC_CHECK_HEADERS([sys/file.h]) AC_CHECK_HEADERS([sys/ioctl.h]) AC_CHECK_HEADERS([sys/param.h]) AC_CHECK_HEADERS([sys/time.h]) AC_CHECK_HEADERS([syslog.h]) dnl ======================================================================== dnl Functions dnl ======================================================================== AC_FUNC_FORK AC_FUNC_STRNLEN AC_CHECK_FUNCS([alarm gettimeofday inet_ntoa memset mkdir socket uname]) AC_CHECK_FUNCS([strcasecmp strchr strdup strerror strrchr strspn strstr strtol strtoul]) AC_PATH_PROGS(REBOOT, reboot, /sbin/reboot) AC_SUBST(REBOOT) AC_SUBST(REBOOT_OPTIONS) AC_DEFINE_UNQUOTED(REBOOT, "$REBOOT", path to the reboot command) AC_DEFINE_UNQUOTED(REBOOT_OPTIONS, "$REBOOT_OPTIONS", reboot options) AC_PATH_PROGS(POWEROFF_CMD, poweroff, /sbin/poweroff) AC_SUBST(POWEROFF_CMD) AC_SUBST(POWEROFF_OPTIONS) AC_DEFINE_UNQUOTED(POWEROFF_CMD, "$POWEROFF_CMD", path to the poweroff command) AC_DEFINE_UNQUOTED(POWEROFF_OPTIONS, "$POWEROFF_OPTIONS", poweroff options) AC_PATH_PROGS(POD2MAN, pod2man) AM_CONDITIONAL(BUILD_POD_DOC, test "x$POD2MAN" != "x" ) if test "x$POD2MAN" = "x"; then AC_MSG_WARN([pod2man not installed, unable to (re-)build ldirector manual page]) fi AC_SUBST(POD2MAN) dnl ======================================================================== dnl Functions dnl ======================================================================== AC_CHECK_FUNCS(getopt, AC_DEFINE(HAVE_DECL_GETOPT, 1, [Have getopt function])) dnl ======================================================================== dnl sfex dnl ======================================================================== build_sfex=no case $host_os in *Linux*|*linux*) if test "$ac_cv_header_heartbeat_glue_config_h" = "yes"; then build_sfex=yes fi ;; esac AM_CONDITIONAL(BUILD_SFEX, test "$build_sfex" = "yes" ) dnl ======================================================================== dnl tickle (needs port to BSD platforms) dnl ======================================================================== AC_CHECK_MEMBERS([struct iphdr.saddr],,,[[#include ]]) AM_CONDITIONAL(BUILD_TICKLE, test "$ac_cv_member_struct_iphdr_saddr" = "yes" ) dnl ======================================================================== dnl libnet dnl ======================================================================== libnet="" libnet_version="none" LIBNETLIBS="" LIBNETDEFINES="" AC_MSG_CHECKING(if libnet is required) libnet_fatal=$enable_libnet case $enable_libnet in no) ;; yes|libnet10|libnet11|10|11) libnet_fatal=yes;; try) case $host_os in *Linux*|*linux*) libnet_fatal=no;; *) libnet_fatal=yes;; dnl legacy behavior esac ;; *) libnet_fatal=yes; enable_libnet=try;; esac AC_MSG_RESULT($libnet_fatal) if test "x$enable_libnet" != "xno"; then AC_PATH_PROGS(LIBNETCONFIG, libnet-config) AC_CHECK_LIB(nsl, t_open) dnl -lnsl AC_CHECK_LIB(socket, socket) dnl -lsocket AC_CHECK_LIB(net, libnet_get_hwaddr, LIBNETLIBS=" -lnet", []) fi AC_MSG_CHECKING(for libnet) if test "x$LIBNETLIBS" != "x" -o "x$enable_libnet" = "xlibnet11"; then LIBNETDEFINES="" if test "$ac_cv_lib_nsl_t_open" = yes; then LIBNETLIBS="-lnsl $LIBNETLIBS" fi if test "$ac_cv_lib_socket_socket" = yes; then LIBNETLIBS="-lsocket $LIBNETLIBS" fi libnet=net libnet_version="libnet1.1" fi if test "x$enable_libnet" = "xtry" -o "x$enable_libnet" = "xlibnet10"; then if test "x$LIBNETLIBS" = x -a "x${LIBNETCONFIG}" != "x" ; then LIBNETDEFINES="`$LIBNETCONFIG --defines` `$LIBNETCONFIG --cflags`"; LIBNETLIBS="`$LIBNETCONFIG --libs`"; libnet_version="libnet1.0 (old)" case $LIBNETLIBS in *-l*) libnet=`echo $LIBNETLIBS | sed 's%.*-l%%'`;; *) libnet_version=none;; esac CPPFLAGS="$CPPFLAGS $LIBNETDEFINES" AC_CHECK_HEADERS(libnet.h) if test "$ac_cv_header_libnet_h" = no; then libnet_version=none fi fi fi AC_MSG_RESULT(found $libnet_version) if test "$libnet_version" = none; then LIBNETLIBS="" LIBNETDEFINES="" if test $libnet_fatal = yes; then AC_MSG_ERROR(libnet not found) fi else AC_CHECK_LIB($libnet,libnet_init, [new_libnet=yes; AC_DEFINE(HAVE_LIBNET_1_1_API, 1, Libnet 1.1 API)], [new_libnet=no; AC_DEFINE(HAVE_LIBNET_1_0_API, 1, Libnet 1.0 API)],$LIBNETLIBS) AC_SUBST(LIBNETLIBS) fi if test "$new_libnet" = yes; then AC_MSG_CHECKING(for libnet API 1.1.4: ) save_CFLAGS="$CFLAGS" CFLAGS="$CFLAGS -fgnu89-inline -Wall -Werror" AC_COMPILE_IFELSE([ AC_LANG_SOURCE(#include int main(){libnet_t *l=NULL; libnet_pblock_record_ip_offset(l, l->total_size); return(0); })], [AC_MSG_RESULT(no)], [AC_DEFINE(HAVE_LIBNET_1_1_4_API, 1, Libnet 1.1.4 API) AC_MSG_RESULT(yes)]) CFLAGS="$save_CFLAGS" fi sendarp_linux=0 case $host_os in *Linux*|*linux*) sendarp_linux=1;; esac redhat_based=0 AC_CHECK_FILE(/etc/redhat-release, [redhat_based=1]) AC_SUBST(LIBNETLIBS) AC_SUBST(LIBNETDEFINES) AM_CONDITIONAL(SENDARP_LINUX, test $sendarp_linux = 1 ) AM_CONDITIONAL(USE_LIBNET, test "x$libnet_version" != "xnone" ) AM_CONDITIONAL(NFSCONVERT, test $redhat_based = 1 ) dnl ************************************************************************ dnl * Check for netinet/icmp6.h to enable the IPv6addr resource agent AC_CHECK_HEADERS(netinet/icmp6.h,[],[],[#include ]) AM_CONDITIONAL(USE_IPV6ADDR_AGENT, test "$ac_cv_header_netinet_icmp6_h" = yes && test "$ac_cv_header_heartbeat_glue_config_h" = yes) AM_CONDITIONAL(IPV6ADDR_COMPATIBLE, test "$ac_cv_header_netinet_icmp6_h" = yes) dnl ======================================================================== dnl Compiler flags dnl ======================================================================== dnl Make sure that CFLAGS is not exported. If the user did dnl not have CFLAGS in their environment then this should have dnl no effect. However if CFLAGS was exported from the user's dnl environment, then the new CFLAGS will also be exported dnl to sub processes. CC_ERRORS="" CC_EXTRAS="" if export -p | fgrep " CFLAGS=" > /dev/null; then SAVED_CFLAGS="$CFLAGS" unset CFLAGS CFLAGS="$SAVED_CFLAGS" unset SAVED_CFLAGS fi if test "$GCC" != yes; then CFLAGS="$CFLAGS -g" enable_fatal_warnings=no else CFLAGS="$CFLAGS -ggdb3" # We had to eliminate -Wnested-externs because of libtool changes # Also remove -Waggregate-return because we use one libnet # call which returns a struct EXTRA_FLAGS="-fgnu89-inline -fstack-protector-all -Wall -Wbad-function-cast -Wcast-qual -Wdeclaration-after-statement -Wendif-labels -Wfloat-equal -Wformat=2 -Wformat-security -Wformat-nonliteral -Winline -Wmissing-prototypes -Wmissing-declarations -Wmissing-format-attribute -Wnested-externs -Wno-long-long -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wunsigned-char -Wwrite-strings -Wno-maybe-uninitialized" # Additional warnings it might be nice to enable one day # -Wshadow # -Wunreachable-code for j in $EXTRA_FLAGS do if cc_supports_flag $j then CC_EXTRAS="$CC_EXTRAS $j" fi done dnl In lib/ais/Makefile.am there's a gcc option available as of v4.x GCC_MAJOR=`gcc -v 2>&1 | awk 'END{print $3}' | sed 's/[.].*//'` AM_CONDITIONAL(GCC_4, test "${GCC_MAJOR}" = 4) dnl System specific options case "$host_os" in *linux*|*bsd*) if test "${enable_fatal_warnings}" = "unknown"; then enable_fatal_warnings=yes fi ;; esac if test "x${enable_fatal_warnings}" != xno && cc_supports_flag -Werror ; then enable_fatal_warnings=yes else enable_fatal_warnings=no fi if test "x${enable_ansi}" != xno && cc_supports_flag -std=iso9899:199409 ; then AC_MSG_NOTICE(Enabling ANSI Compatibility) CC_EXTRAS="$CC_EXTRAS -ansi -D_GNU_SOURCE -DANSI_ONLY" fi AC_MSG_NOTICE(Activated additional gcc flags: ${CC_EXTRAS}) fi CFLAGS="$CFLAGS $CC_EXTRAS" NON_FATAL_CFLAGS="$CFLAGS" AC_SUBST(NON_FATAL_CFLAGS) dnl dnl We reset CFLAGS to include our warnings *after* all function dnl checking goes on, so that our warning flags don't keep the dnl AC_*FUNCS() calls above from working. In particular, -Werror will dnl *always* cause us troubles if we set it before here. dnl dnl if test "x${enable_fatal_warnings}" = xyes ; then AC_MSG_NOTICE(Enabling Fatal Warnings) CFLAGS="$CFLAGS -Werror" fi AC_SUBST(CFLAGS) dnl This is useful for use in Makefiles that need to remove one specific flag CFLAGS_COPY="$CFLAGS" AC_SUBST(CFLAGS_COPY) AC_SUBST(LOCALE) AC_SUBST(CC) AC_SUBST(MAKE) dnl The Makefiles and shell scripts we output AC_CONFIG_FILES(Makefile \ resource-agents.pc \ include/Makefile \ heartbeat/Makefile \ heartbeat/ocf-binaries \ heartbeat/ocf-directories \ heartbeat/ocf-shellfuncs \ heartbeat/shellfuncs \ systemd/Makefile \ systemd/resource-agents.conf \ tools/Makefile \ tools/nfsconvert \ tools/ocf-tester \ tools/ocft/Makefile \ tools/ocft/ocft \ tools/ocft/caselib \ tools/ocft/README \ tools/ocft/README.zh_CN \ ldirectord/Makefile \ ldirectord/ldirectord \ ldirectord/init.d/Makefile \ ldirectord/init.d/ldirectord \ ldirectord/init.d/ldirectord.debian \ ldirectord/init.d/ldirectord.debian.default \ ldirectord/systemd/Makefile \ ldirectord/systemd/ldirectord.service \ ldirectord/logrotate.d/Makefile \ ldirectord/OCF/Makefile \ ldirectord/OCF/ldirectord \ doc/Makefile \ doc/man/Makefile \ rgmanager/Makefile \ rgmanager/src/Makefile \ rgmanager/src/resources/Makefile \ rgmanager/src/resources/ocf-shellfuncs \ rgmanager/src/resources/svclib_nfslock \ rgmanager/src/resources/lvm_by_lv.sh \ rgmanager/src/resources/lvm_by_vg.sh \ rgmanager/src/resources/utils/Makefile \ rgmanager/src/resources/utils/fs-lib.sh \ rgmanager/src/resources/utils/messages.sh \ rgmanager/src/resources/utils/config-utils.sh \ rgmanager/src/resources/utils/member_util.sh \ rgmanager/src/resources/utils/ra-skelet.sh \ ) dnl Files we output that need to be executable AC_CONFIG_FILES([heartbeat/azure-events], [chmod +x heartbeat/azure-events]) AC_CONFIG_FILES([heartbeat/azure-events-az], [chmod +x heartbeat/azure-events-az]) AC_CONFIG_FILES([heartbeat/AoEtarget], [chmod +x heartbeat/AoEtarget]) AC_CONFIG_FILES([heartbeat/ManageRAID], [chmod +x heartbeat/ManageRAID]) AC_CONFIG_FILES([heartbeat/ManageVE], [chmod +x heartbeat/ManageVE]) AC_CONFIG_FILES([heartbeat/Squid], [chmod +x heartbeat/Squid]) AC_CONFIG_FILES([heartbeat/SysInfo], [chmod +x heartbeat/SysInfo]) AC_CONFIG_FILES([heartbeat/aws-vpc-route53], [chmod +x heartbeat/aws-vpc-route53]) AC_CONFIG_FILES([heartbeat/clvm], [chmod +x heartbeat/clvm]) AC_CONFIG_FILES([heartbeat/conntrackd], [chmod +x heartbeat/conntrackd]) AC_CONFIG_FILES([heartbeat/dnsupdate], [chmod +x heartbeat/dnsupdate]) AC_CONFIG_FILES([heartbeat/dummypy], [chmod +x heartbeat/dummypy]) AC_CONFIG_FILES([heartbeat/eDir88], [chmod +x heartbeat/eDir88]) AC_CONFIG_FILES([heartbeat/fio], [chmod +x heartbeat/fio]) AC_CONFIG_FILES([heartbeat/galera], [chmod +x heartbeat/galera]) AC_CONFIG_FILES([heartbeat/gcp-pd-move], [chmod +x heartbeat/gcp-pd-move]) AC_CONFIG_FILES([heartbeat/gcp-vpc-move-ip], [chmod +x heartbeat/gcp-vpc-move-ip]) AC_CONFIG_FILES([heartbeat/gcp-vpc-move-vip], [chmod +x heartbeat/gcp-vpc-move-vip]) AC_CONFIG_FILES([heartbeat/gcp-vpc-move-route], [chmod +x heartbeat/gcp-vpc-move-route]) AC_CONFIG_FILES([heartbeat/iSCSILogicalUnit], [chmod +x heartbeat/iSCSILogicalUnit]) AC_CONFIG_FILES([heartbeat/iSCSITarget], [chmod +x heartbeat/iSCSITarget]) AC_CONFIG_FILES([heartbeat/jira], [chmod +x heartbeat/jira]) AC_CONFIG_FILES([heartbeat/kamailio], [chmod +x heartbeat/kamailio]) AC_CONFIG_FILES([heartbeat/lxc], [chmod +x heartbeat/lxc]) AC_CONFIG_FILES([heartbeat/lxd-info], [chmod +x heartbeat/lxd-info]) AC_CONFIG_FILES([heartbeat/machine-info], [chmod +x heartbeat/machine-info]) AC_CONFIG_FILES([heartbeat/mariadb], [chmod +x heartbeat/mariadb]) AC_CONFIG_FILES([heartbeat/mpathpersist], [chmod +x heartbeat/mpathpersist]) AC_CONFIG_FILES([heartbeat/nfsnotify], [chmod +x heartbeat/nfsnotify]) AC_CONFIG_FILES([heartbeat/openstack-info], [chmod +x heartbeat/openstack-info]) +AC_CONFIG_FILES([heartbeat/powervs-subnet], [chmod +x heartbeat/powervs-subnet]) AC_CONFIG_FILES([heartbeat/rabbitmq-cluster], [chmod +x heartbeat/rabbitmq-cluster]) AC_CONFIG_FILES([heartbeat/redis], [chmod +x heartbeat/redis]) AC_CONFIG_FILES([heartbeat/rsyslog], [chmod +x heartbeat/rsyslog]) AC_CONFIG_FILES([heartbeat/smb-share], [chmod +x heartbeat/smb-share]) AC_CONFIG_FILES([heartbeat/sg_persist], [chmod +x heartbeat/sg_persist]) AC_CONFIG_FILES([heartbeat/slapd], [chmod +x heartbeat/slapd]) AC_CONFIG_FILES([heartbeat/storage-mon], [chmod +x heartbeat/storage-mon]) AC_CONFIG_FILES([heartbeat/sybaseASE], [chmod +x heartbeat/sybaseASE]) AC_CONFIG_FILES([heartbeat/syslog-ng], [chmod +x heartbeat/syslog-ng]) AC_CONFIG_FILES([heartbeat/vsftpd], [chmod +x heartbeat/vsftpd]) AC_CONFIG_FILES([heartbeat/CTDB], [chmod +x heartbeat/CTDB]) AC_CONFIG_FILES([rgmanager/src/resources/ASEHAagent.sh], [chmod +x rgmanager/src/resources/ASEHAagent.sh]) AC_CONFIG_FILES([rgmanager/src/resources/apache.sh], [chmod +x rgmanager/src/resources/apache.sh]) AC_CONFIG_FILES([rgmanager/src/resources/bind-mount.sh], [chmod +x rgmanager/src/resources/bind-mount.sh]) AC_CONFIG_FILES([rgmanager/src/resources/clusterfs.sh], [chmod +x rgmanager/src/resources/clusterfs.sh]) AC_CONFIG_FILES([rgmanager/src/resources/db2.sh], [chmod +x rgmanager/src/resources/db2.sh]) AC_CONFIG_FILES([rgmanager/src/resources/drbd.sh], [chmod +x rgmanager/src/resources/drbd.sh]) AC_CONFIG_FILES([rgmanager/src/resources/fs.sh], [chmod +x rgmanager/src/resources/fs.sh]) AC_CONFIG_FILES([rgmanager/src/resources/ip.sh], [chmod +x rgmanager/src/resources/ip.sh]) AC_CONFIG_FILES([rgmanager/src/resources/lvm.sh], [chmod +x rgmanager/src/resources/lvm.sh]) AC_CONFIG_FILES([rgmanager/src/resources/mysql.sh], [chmod +x rgmanager/src/resources/mysql.sh]) AC_CONFIG_FILES([rgmanager/src/resources/named.sh], [chmod +x rgmanager/src/resources/named.sh]) AC_CONFIG_FILES([rgmanager/src/resources/netfs.sh], [chmod +x rgmanager/src/resources/netfs.sh]) AC_CONFIG_FILES([rgmanager/src/resources/nfsclient.sh], [chmod +x rgmanager/src/resources/nfsclient.sh]) AC_CONFIG_FILES([rgmanager/src/resources/nfsexport.sh], [chmod +x rgmanager/src/resources/nfsexport.sh]) AC_CONFIG_FILES([rgmanager/src/resources/nfsserver.sh], [chmod +x rgmanager/src/resources/nfsserver.sh]) AC_CONFIG_FILES([rgmanager/src/resources/openldap.sh], [chmod +x rgmanager/src/resources/openldap.sh]) AC_CONFIG_FILES([rgmanager/src/resources/oracledb.sh], [chmod +x rgmanager/src/resources/oracledb.sh]) AC_CONFIG_FILES([rgmanager/src/resources/oradg.sh], [chmod +x rgmanager/src/resources/oradg.sh]) AC_CONFIG_FILES([rgmanager/src/resources/orainstance.sh], [chmod +x rgmanager/src/resources/orainstance.sh]) AC_CONFIG_FILES([rgmanager/src/resources/oralistener.sh], [chmod +x rgmanager/src/resources/oralistener.sh]) AC_CONFIG_FILES([rgmanager/src/resources/postgres-8.sh], [chmod +x rgmanager/src/resources/postgres-8.sh]) AC_CONFIG_FILES([rgmanager/src/resources/samba.sh], [chmod +x rgmanager/src/resources/samba.sh]) AC_CONFIG_FILES([rgmanager/src/resources/script.sh], [chmod +x rgmanager/src/resources/script.sh]) AC_CONFIG_FILES([rgmanager/src/resources/service.sh], [chmod +x rgmanager/src/resources/service.sh]) AC_CONFIG_FILES([rgmanager/src/resources/smb.sh], [chmod +x rgmanager/src/resources/smb.sh]) AC_CONFIG_FILES([rgmanager/src/resources/tomcat-5.sh], [chmod +x rgmanager/src/resources/tomcat-5.sh]) AC_CONFIG_FILES([rgmanager/src/resources/tomcat-6.sh], [chmod +x rgmanager/src/resources/tomcat-6.sh]) AC_CONFIG_FILES([rgmanager/src/resources/vm.sh], [chmod +x rgmanager/src/resources/vm.sh]) dnl Now process the entire list of files added by previous dnl calls to AC_CONFIG_FILES() AC_OUTPUT() dnl ***************** dnl Configure summary dnl ***************** AC_MSG_RESULT([]) AC_MSG_RESULT([$PACKAGE configuration:]) AC_MSG_RESULT([ Version = ${VERSION}]) AC_MSG_RESULT([ Build Version = $Format:%H$]) AC_MSG_RESULT([]) AC_MSG_RESULT([ Prefix = ${prefix}]) AC_MSG_RESULT([ Executables = ${sbindir}]) AC_MSG_RESULT([ Man pages = ${mandir}]) AC_MSG_RESULT([ Libraries = ${libdir}]) AC_MSG_RESULT([ Header files = ${includedir}]) AC_MSG_RESULT([ Arch-independent files = ${datadir}]) AC_MSG_RESULT([ Documentation = ${docdir}]) AC_MSG_RESULT([ State information = ${localstatedir}]) AC_MSG_RESULT([ System configuration = ${sysconfdir}]) AC_MSG_RESULT([ HA_BIN directory prefix = ${libexecdir}]) AC_MSG_RESULT([ RA state files = ${HA_RSCTMPDIR}]) AC_MSG_RESULT([ AIS Plugins = ${LCRSODIR}]) AC_MSG_RESULT([]) AC_MSG_RESULT([ CPPFLAGS = ${CPPFLAGS}]) AC_MSG_RESULT([ CFLAGS = ${CFLAGS}]) AC_MSG_RESULT([ Libraries = ${LIBS}]) AC_MSG_RESULT([ Stack Libraries = ${CLUSTERLIBS}]) diff --git a/doc/man/Makefile.am b/doc/man/Makefile.am index d14b6c53c..e577e6357 100644 --- a/doc/man/Makefile.am +++ b/doc/man/Makefile.am @@ -1,264 +1,265 @@ # # doc: Linux-HA resource agents # # Copyright (C) 2009 Florian Haas # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # MAINTAINERCLEANFILES = Makefile.in EXTRA_DIST = $(doc_DATA) $(REFENTRY_STYLESHEET) \ mkappendix.sh ralist.sh CLEANFILES = $(man_MANS) $(xmlfiles) metadata-*.xml STYLESHEET_PREFIX ?= http://docbook.sourceforge.net/release/xsl/current MANPAGES_STYLESHEET ?= $(STYLESHEET_PREFIX)/manpages/docbook.xsl HTML_STYLESHEET ?= $(STYLESHEET_PREFIX)/xhtml/docbook.xsl FO_STYLESHEET ?= $(STYLESHEET_PREFIX)/fo/docbook.xsl REFENTRY_STYLESHEET ?= ra2refentry.xsl XSLTPROC_OPTIONS ?= --xinclude XSLTPROC_MANPAGES_OPTIONS ?= $(XSLTPROC_OPTIONS) XSLTPROC_HTML_OPTIONS ?= $(XSLTPROC_OPTIONS) XSLTPROC_FO_OPTIONS ?= $(XSLTPROC_OPTIONS) radir = $(abs_top_builddir)/heartbeat # required for out-of-tree build symlinkstargets = \ ocf-distro ocf.py ocf-rarun ocf-returncodes \ findif.sh apache-conf.sh http-mon.sh mysql-common.sh \ nfsserver-redhat.sh openstack-common.sh ora-common.sh preptree: for i in $(symlinkstargets); do \ if [ ! -f $(radir)/$$i ]; then \ rm -rf $(radir)/$$i; \ ln -sf $(abs_top_srcdir)/heartbeat/$$i $(radir)/$$i; \ fi; \ done $(radir)/%: $(abs_top_srcdir)/heartbeat/% if [ ! -f $@ ]; then \ ln -sf $< $@; \ fi # OCF_ROOT=. is necessary due to a sanity check in ocf-shellfuncs # (which tests whether $OCF_ROOT points to a directory metadata-%.xml: $(radir)/% preptree OCF_ROOT=. OCF_FUNCTIONS_DIR=$(radir) $< meta-data > $@ metadata-IPv6addr.xml: $(radir)/IPv6addr OCF_ROOT=. OCF_FUNCTIONS_DIR=$(radir) $< meta-data > $@ clean-local: find $(radir) -type l -exec rm -rf {} \; # Please note: we can't name the man pages # ocf:heartbeat:. Believe me, I've tried. It looks like it # works, but then it doesn't. While make can deal correctly with # colons in target names (when properly escaped), it royally messes up # when it is deals with _dependencies_ that contain colons. See Bug # 12126 on savannah.gnu.org. But, maybe it gets fixed soon, it was # first reported in 1995 and added to Savannah in in 2005... if BUILD_DOC man_MANS = ocf_heartbeat_AoEtarget.7 \ ocf_heartbeat_AudibleAlarm.7 \ ocf_heartbeat_ClusterMon.7 \ ocf_heartbeat_CTDB.7 \ ocf_heartbeat_Delay.7 \ ocf_heartbeat_Dummy.7 \ ocf_heartbeat_EvmsSCC.7 \ ocf_heartbeat_Evmsd.7 \ ocf_heartbeat_Filesystem.7 \ ocf_heartbeat_ICP.7 \ ocf_heartbeat_IPaddr.7 \ ocf_heartbeat_IPaddr2.7 \ ocf_heartbeat_IPsrcaddr.7 \ ocf_heartbeat_LVM.7 \ ocf_heartbeat_LVM-activate.7 \ ocf_heartbeat_LinuxSCSI.7 \ ocf_heartbeat_MailTo.7 \ ocf_heartbeat_ManageRAID.7 \ ocf_heartbeat_ManageVE.7 \ ocf_heartbeat_NodeUtilization.7 \ ocf_heartbeat_Pure-FTPd.7 \ ocf_heartbeat_Raid1.7 \ ocf_heartbeat_Route.7 \ ocf_heartbeat_SAPDatabase.7 \ ocf_heartbeat_SAPInstance.7 \ ocf_heartbeat_SendArp.7 \ ocf_heartbeat_ServeRAID.7 \ ocf_heartbeat_SphinxSearchDaemon.7 \ ocf_heartbeat_Squid.7 \ ocf_heartbeat_Stateful.7 \ ocf_heartbeat_SysInfo.7 \ ocf_heartbeat_VIPArip.7 \ ocf_heartbeat_VirtualDomain.7 \ ocf_heartbeat_WAS.7 \ ocf_heartbeat_WAS6.7 \ ocf_heartbeat_WinPopup.7 \ ocf_heartbeat_Xen.7 \ ocf_heartbeat_Xinetd.7 \ ocf_heartbeat_ZFS.7 \ ocf_heartbeat_aliyun-vpc-move-ip.7 \ ocf_heartbeat_anything.7 \ ocf_heartbeat_apache.7 \ ocf_heartbeat_asterisk.7 \ ocf_heartbeat_aws-vpc-move-ip.7 \ ocf_heartbeat_aws-vpc-route53.7 \ ocf_heartbeat_awseip.7 \ ocf_heartbeat_awsvip.7 \ ocf_heartbeat_azure-lb.7 \ ocf_heartbeat_clvm.7 \ ocf_heartbeat_conntrackd.7 \ ocf_heartbeat_corosync-qnetd.7 \ ocf_heartbeat_crypt.7 \ ocf_heartbeat_db2.7 \ ocf_heartbeat_dhcpd.7 \ ocf_heartbeat_docker.7 \ ocf_heartbeat_docker-compose.7 \ ocf_heartbeat_dovecot.7 \ ocf_heartbeat_dnsupdate.7 \ ocf_heartbeat_dummypy.7 \ ocf_heartbeat_eDir88.7 \ ocf_heartbeat_ethmonitor.7 \ ocf_heartbeat_exportfs.7 \ ocf_heartbeat_fio.7 \ ocf_heartbeat_galera.7 \ ocf_heartbeat_garbd.7 \ ocf_heartbeat_gcp-ilb.7 \ ocf_heartbeat_gcp-vpc-move-ip.7 \ ocf_heartbeat_iSCSILogicalUnit.7 \ ocf_heartbeat_iSCSITarget.7 \ ocf_heartbeat_iface-bridge.7 \ ocf_heartbeat_iface-macvlan.7 \ ocf_heartbeat_iface-vlan.7 \ ocf_heartbeat_ipsec.7 \ ocf_heartbeat_ids.7 \ ocf_heartbeat_iscsi.7 \ ocf_heartbeat_jboss.7 \ ocf_heartbeat_jira.7 \ ocf_heartbeat_kamailio.7 \ ocf_heartbeat_lvmlockd.7 \ ocf_heartbeat_lxc.7 \ ocf_heartbeat_lxd-info.7 \ ocf_heartbeat_machine-info.7 \ ocf_heartbeat_mariadb.7 \ ocf_heartbeat_mdraid.7 \ ocf_heartbeat_minio.7 \ ocf_heartbeat_mpathpersist.7 \ ocf_heartbeat_mysql.7 \ ocf_heartbeat_mysql-proxy.7 \ ocf_heartbeat_nagios.7 \ ocf_heartbeat_named.7 \ ocf_heartbeat_nfsnotify.7 \ ocf_heartbeat_nfsserver.7 \ ocf_heartbeat_nginx.7 \ ocf_heartbeat_nvmet-subsystem.7 \ ocf_heartbeat_nvmet-namespace.7 \ ocf_heartbeat_nvmet-port.7 \ ocf_heartbeat_openstack-info.7 \ ocf_heartbeat_ocivip.7 \ ocf_heartbeat_openstack-cinder-volume.7 \ ocf_heartbeat_openstack-floating-ip.7 \ ocf_heartbeat_openstack-virtual-ip.7 \ ocf_heartbeat_oraasm.7 \ ocf_heartbeat_oracle.7 \ ocf_heartbeat_oralsnr.7 \ ocf_heartbeat_osceip.7 \ ocf_heartbeat_ovsmonitor.7 \ ocf_heartbeat_pgagent.7 \ ocf_heartbeat_pgsql.7 \ ocf_heartbeat_pingd.7 \ ocf_heartbeat_podman.7 \ ocf_heartbeat_portblock.7 \ ocf_heartbeat_postfix.7 \ ocf_heartbeat_pound.7 \ + ocf_heartbeat_powervs-subnet.7 \ ocf_heartbeat_proftpd.7 \ ocf_heartbeat_rabbitmq-cluster.7 \ ocf_heartbeat_rabbitmq-server-ha.7 \ ocf_heartbeat_redis.7 \ ocf_heartbeat_rkt.7 \ ocf_heartbeat_rsyncd.7 \ ocf_heartbeat_rsyslog.7 \ ocf_heartbeat_scsi2reservation.7 \ ocf_heartbeat_sfex.7 \ ocf_heartbeat_slapd.7 \ ocf_heartbeat_smb-share.7 \ ocf_heartbeat_sybaseASE.7 \ ocf_heartbeat_sg_persist.7 \ ocf_heartbeat_storage-mon.7 \ ocf_heartbeat_symlink.7 \ ocf_heartbeat_syslog-ng.7 \ ocf_heartbeat_tomcat.7 \ ocf_heartbeat_varnish.7 \ ocf_heartbeat_vdo-vol.7 \ ocf_heartbeat_vmware.7 \ ocf_heartbeat_vsftpd.7 \ ocf_heartbeat_zabbixserver.7 if USE_IPV6ADDR_AGENT man_MANS += ocf_heartbeat_IPv6addr.7 endif if BUILD_AZURE_EVENTS man_MANS += ocf_heartbeat_azure-events.7 endif if BUILD_AZURE_EVENTS_AZ man_MANS += ocf_heartbeat_azure-events-az.7 endif if BUILD_GCP_PD_MOVE man_MANS += ocf_heartbeat_gcp-pd-move.7 endif if BUILD_GCP_VPC_MOVE_ROUTE man_MANS += ocf_heartbeat_gcp-vpc-move-route.7 endif if BUILD_GCP_VPC_MOVE_VIP man_MANS += ocf_heartbeat_gcp-vpc-move-vip.7 endif xmlfiles = $(man_MANS:.7=.xml) %.1 %.5 %.7 %.8: %.xml $(XSLTPROC) \ $(XSLTPROC_MANPAGES_OPTIONS) \ $(MANPAGES_STYLESHEET) $< ocf_heartbeat_%.xml: metadata-%.xml $(srcdir)/$(REFENTRY_STYLESHEET) $(XSLTPROC) --novalid \ --stringparam package $(PACKAGE_NAME) \ --stringparam version $(VERSION) \ --output $@ \ $(srcdir)/$(REFENTRY_STYLESHEET) $< ocf_resource_agents.xml: $(xmlfiles) mkappendix.sh ./mkappendix.sh $(xmlfiles) > $@ %.html: %.xml $(XSLTPROC) \ $(XSLTPROC_HTML_OPTIONS) \ --output $@ \ $(HTML_STYLESHEET) $< xml: ocf_resource_agents.xml endif diff --git a/heartbeat/Makefile.am b/heartbeat/Makefile.am index f7e1703af..ff73a15aa 100644 --- a/heartbeat/Makefile.am +++ b/heartbeat/Makefile.am @@ -1,250 +1,251 @@ # Makefile.am for OCF RAs # # Author: Sun Jing Dong # Copyright (C) 2004 IBM # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # MAINTAINERCLEANFILES = Makefile.in EXTRA_DIST = $(ocf_SCRIPTS) $(ocfcommon_DATA) \ $(common_DATA) $(hb_DATA) $(dtd_DATA) \ README README.galera AM_CPPFLAGS = -I$(top_srcdir)/include -I$(top_srcdir)/linux-ha halibdir = $(libexecdir)/heartbeat ocfdir = $(OCF_RA_DIR_PREFIX)/heartbeat dtddir = $(datadir)/$(PACKAGE_NAME) dtd_DATA = ra-api-1.dtd metadata.rng ocf_PROGRAMS = if USE_IPV6ADDR_AGENT ocf_PROGRAMS += IPv6addr endif halib_PROGRAMS = if IPV6ADDR_COMPATIBLE halib_PROGRAMS += send_ua endif IPv6addr_SOURCES = IPv6addr.c IPv6addr_utils.c IPv6addr_LDADD = -lplumb $(LIBNETLIBS) send_ua_SOURCES = send_ua.c IPv6addr_utils.c send_ua_LDADD = $(LIBNETLIBS) ocf_SCRIPTS = AoEtarget \ AudibleAlarm \ ClusterMon \ CTDB \ Delay \ Dummy \ EvmsSCC \ Evmsd \ Filesystem \ ICP \ IPaddr \ IPaddr2 \ IPsrcaddr \ LVM \ LinuxSCSI \ lvmlockd \ LVM-activate \ MailTo \ ManageRAID \ ManageVE \ NodeUtilization \ Pure-FTPd \ Raid1 \ Route \ SAPDatabase \ SAPInstance \ SendArp \ ServeRAID \ SphinxSearchDaemon \ Squid \ Stateful \ SysInfo \ VIPArip \ VirtualDomain \ WAS \ WAS6 \ WinPopup \ Xen \ Xinetd \ ZFS \ aliyun-vpc-move-ip \ anything \ apache \ asterisk \ aws-vpc-move-ip \ aws-vpc-route53 \ awseip \ awsvip \ azure-lb \ clvm \ conntrackd \ corosync-qnetd \ crypt \ db2 \ dhcpd \ dnsupdate \ dummypy \ docker \ docker-compose \ dovecot \ eDir88 \ ethmonitor \ exportfs \ fio \ galera \ garbd \ gcp-ilb \ gcp-vpc-move-ip \ iSCSILogicalUnit \ iSCSITarget \ ids \ iface-bridge \ iface-macvlan \ iface-vlan \ ipsec \ iscsi \ jboss \ jira \ kamailio \ lxc \ lxd-info \ machine-info \ mariadb \ mdraid \ minio \ mysql \ mysql-proxy \ nagios \ named \ nfsnotify \ nfsserver \ nginx \ nvmet-subsystem \ nvmet-namespace \ nvmet-port \ ocivip \ openstack-cinder-volume \ openstack-floating-ip \ openstack-info \ openstack-virtual-ip \ oraasm \ oracle \ oralsnr \ osceip \ ovsmonitor \ pgagent \ pgsql \ pingd \ podman \ portblock \ postfix \ pound \ + powervs-subnet \ proftpd \ rabbitmq-cluster \ rabbitmq-server-ha \ redis \ rkt \ rsyncd \ rsyslog \ scsi2reservation \ sfex \ sg_persist \ mpathpersist \ slapd \ smb-share \ storage-mon \ sybaseASE \ symlink \ syslog-ng \ tomcat \ varnish \ vdo-vol \ vmware \ vsftpd \ zabbixserver if BUILD_AZURE_EVENTS ocf_SCRIPTS += azure-events endif if BUILD_AZURE_EVENTS_AZ ocf_SCRIPTS += azure-events-az endif if BUILD_GCP_PD_MOVE ocf_SCRIPTS += gcp-pd-move endif if BUILD_GCP_VPC_MOVE_ROUTE ocf_SCRIPTS += gcp-vpc-move-route endif if BUILD_GCP_VPC_MOVE_VIP ocf_SCRIPTS += gcp-vpc-move-vip endif ocfcommondir = $(OCF_LIB_DIR_PREFIX)/heartbeat ocfcommon_DATA = ocf-shellfuncs \ ocf-binaries \ ocf-directories \ ocf-returncodes \ ocf-rarun \ ocf-distro \ apache-conf.sh \ http-mon.sh \ sapdb-nosha.sh \ sapdb.sh \ lvm-clvm.sh \ lvm-plain.sh \ lvm-tag.sh \ openstack-common.sh \ ora-common.sh \ mysql-common.sh \ nfsserver-redhat.sh \ findif.sh \ ocf.py # Legacy locations hbdir = $(sysconfdir)/ha.d hb_DATA = shellfuncs check: $(ocf_SCRIPTS:=.check) %.check: % OCF_ROOT=$(abs_srcdir) OCF_FUNCTIONS_DIR=$(abs_srcdir) ./$< meta-data | xmllint --path $(abs_srcdir) --noout --relaxng $(abs_srcdir)/metadata.rng - do_spellcheck = printf '[%s]\n' "$(agent)"; \ OCF_ROOT=$(abs_srcdir) OCF_FUNCTIONS_DIR=$(abs_srcdir) \ ./$(agent) meta-data 2>/dev/null \ | xsltproc $(top_srcdir)/make/extract_text.xsl - \ | aspell pipe list -d en_US --ignore-case \ --home-dir=$(top_srcdir)/make -p spellcheck-ignore \ | sed -n 's|^&\([^:]*\):.*|\1|p'; spellcheck: @$(foreach agent,$(ocf_SCRIPTS), $(do_spellcheck)) clean-local: rm -rf __pycache__ *.pyc diff --git a/heartbeat/powervs-subnet.in b/heartbeat/powervs-subnet.in new file mode 100755 index 000000000..9e3b50887 --- /dev/null +++ b/heartbeat/powervs-subnet.in @@ -0,0 +1,1058 @@ +#!@PYTHON@ -tt +# ------------------------------------------------------------------------ +# Description: Resource Agent to move a Power Virtual Server subnet +# and its IP address from one virtual server instance +# to another. +# +# Authors: Edmund Haefele +# Walter Orb +# +# Copyright (c) 2024 International Business Machines, Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# ------------------------------------------------------------------------ + +import ipaddress +import json +import os +import re +import socket +import subprocess +import sys +import textwrap +import time + +import requests +import requests.adapters +import urllib3.util + +OCF_FUNCTIONS_DIR = os.environ.get( + "OCF_FUNCTIONS_DIR", "%s/lib/heartbeat" % os.environ.get("OCF_ROOT") +) + +sys.path.append(OCF_FUNCTIONS_DIR) + +try: + import ocf +except ImportError: + sys.stderr.write("ImportError: ocf module import failed.") + sys.exit(5) + + +class PowerCloudAPIError(Exception): + def __init__(self, message, exit_code): + ocf.ocf_exit_reason(message) + sys.exit(exit_code) + + +class nmcli: + """A wrapper class to run nmcli system commands.""" + + NMCLI_SYSTEM_CMD = ["nmcli", "-t"] + CONN_PREFIX = "VIP_" + DEV_PREFIX = "env" + ROUTING_PRIO = 50 + ROUTING_TABLE = 500 + _WAIT_FOR_NIC_SLEEP = 3 + + def __init__(self): + """Class implements only classmethods or staticmethods, instantiation is not used.""" + pass + + @classmethod + def _nmcli_os_cmd(cls, nmcli_args): + """run os nmcli command with the specified arguments. + + Returns the output as a dictionary. + """ + + ocf.logger.debug("_nmcli_os_cmd: args: {}".format(nmcli_args)) + output = None + try: + result = subprocess.run( + cls.NMCLI_SYSTEM_CMD + nmcli_args, + capture_output=True, + text=True, + check=True, + env={"LANG": "C"}, + ) + if len(nmcli_args) == 1 or nmcli_args[0] == "-g" or nmcli_args[1] == "show": + # return output as dict + output = dict( + item.split(":", 1) + for item in result.stdout.rstrip().splitlines() + if ":" in item + ) + except subprocess.CalledProcessError as e: + raise PowerCloudAPIError( + f"_nmcli_os_cmd: error executing nmcli: {e.stderr}", + ocf.OCF_ERR_GENERIC, + ) + + return output + + @classmethod + def _nmcli_cmd(cls, command, subcommand=None, name=None, **kwargs): + """Prepare arguments to call nmcli command.""" + + ocf.logger.debug( + f"_nmcli_cmd: args: command: {command}, subcommand: {subcommand}, name: {name}" + ) + if command in ["connection", "device"]: + nmcli_args = [command] + else: + raise PowerCloudAPIError( + f"_nmcli_cmd: nmcli {command} not implemented", + ocf.OCF_ERR_GENERIC, + ) + if name: + if subcommand in ("show", "delete", "down", "up"): + nmcli_args += [subcommand, name] + elif subcommand == "add": + nmcli_args += [subcommand, "type", "ethernet", "con-name", name] + else: + raise PowerCloudAPIError( + f"_nmcli_cmd: nmcli {command} {subcommand} not implemented", + ocf.OCF_ERR_GENERIC, + ) + elif subcommand in ("add", "delete", "down", "up"): + raise PowerCloudAPIError( + f"_nmcli_cmd: name argument required for nmcli {command} {subcommand}", + ocf.OCF_ERR_GENERIC, + ) + + options = kwargs.get("options", {}) + for k, v in options.items(): + nmcli_args += [k, v] + + return cls._nmcli_os_cmd(nmcli_args) + + @classmethod + def _nmcli_find(cls, command, match_key, match_value): + """Find the network object whose attribute with the specified key matches the specified value.""" + + ocf.logger.debug( + f"_nmcli_find: args: command: {command}, key: {match_key}, value: {match_value}" + ) + + nm_object = None + for name in cls._nmcli_cmd(command=command, subcommand="show"): + if not re.search(f"({cls.CONN_PREFIX})?{cls.DEV_PREFIX}", name): + # check only connections or devices with device prefix in name + continue + obj_attrs = cls._nmcli_cmd(command=command, subcommand="show", name=name) + if re.search(match_value, obj_attrs.get(match_key, "")): + ocf.logger.debug(f"_nmcli_find: found match: name: {name}") + nm_object = obj_attrs + break + + return nm_object + + @classmethod + def cleanup(cls): + """Cleanup orphaned Network Manager connections.""" + + connections = cls._nmcli_os_cmd(["-g", "UUID,NAME,ACTIVE", "connection"]) + for uuid in connections: + name, active = connections[uuid].split(":") + if active == "no" and name.startswith(f"{cls.CONN_PREFIX}{cls.DEV_PREFIX}"): + ocf.logger.debug(f"nmcli.cleanup: delete orphaned connection {name}") + nmcli.connection.delete(uuid) + + @classmethod + def wait_for_nic(cls, mac, timeout=720): + """Wait for a NIC with a given MAC address to become available.""" + + ocf.logger.debug(f"wait_for_nic: args: mac: {mac}, timeout: {timeout} s") + mac_address = mac.upper() + retries = (timeout // cls._WAIT_FOR_NIC_SLEEP) - 1 + + for attempt in range(1, retries + 1): + try: + ocf.logger.debug( + f"wait_for_nic: waiting for nic with mac address {mac_address} ..." + ) + nm_object = cls._nmcli_find("device", "GENERAL.HWADDR", mac_address) + if nm_object: + break + finally: + time.sleep(cls._WAIT_FOR_NIC_SLEEP) + else: # no break + raise PowerCloudAPIError( + f"wait_for_nic: timeout while waiting for nic with MAC address {mac_address}", + ocf.OCF_ERR_GENERIC, + ) + + nic = nm_object.get("GENERAL.DEVICE") + wait_time = (attempt - 1) * cls._WAIT_FOR_NIC_SLEEP + + ocf.logger.info( + f"wait_for_nic: found network device {nic} with MAC address {mac_address} after waiting {wait_time} seconds" + ) + + return nic + + @classmethod + def find_gateway(cls, ip): + """Find the gateway address for a given IP.""" + + ocf.logger.debug(f"find_gateway: args: ip: {ip}") + + gateway = None + ip_address = ip.split("/")[0] + dev = cls._nmcli_find("device", "IP4.ADDRESS[1]", ip_address) + if dev: + # Sample IP4.ROUTE[2]: dst = 0.0.0.0/0, nh = 10.10.10.101, mt = 102, table=200 + # extract next hop (nh) value + ip4_route2 = dict( + item.split("=") + for item in dev["IP4.ROUTE[2]"].replace(" ", "").split(",") + ) + gateway = ip4_route2.get("nh", None) + + return gateway + + class connection: + """Provides methods to run nmcli connection commands.""" + + @staticmethod + def show(name=None, **kwargs): + return nmcli._nmcli_cmd("connection", "show", name, **kwargs) + + @staticmethod + def add(name, **kwargs): + return nmcli._nmcli_cmd("connection", "add", name, **kwargs) + + @staticmethod + def delete(name, **kwargs): + return nmcli._nmcli_cmd("connection", "delete", name, **kwargs) + + @staticmethod + def down(name, **kwargs): + return nmcli._nmcli_cmd("connection", "down", name, **kwargs) + + @staticmethod + def up(name, **kwargs): + return nmcli._nmcli_cmd("connection", "up", name, **kwargs) + + @staticmethod + def find(match_key, match_value): + return nmcli._nmcli_find("connection", match_key, match_value) + + class device: + """Provides methods to run nmcli device commands.""" + + @staticmethod + def show(name=None, **kwargs): + return nmcli._nmcli_cmd("device", "show", name, **kwargs) + + @staticmethod + def find(match_key, match_value): + return nmcli._nmcli_find("device", match_key, match_value) + + +class PowerCloudAPI: + """Provides methods to manage Power Virtual Server resources through its REST API.""" + + _URL_IAM_GLOBAL = "https://iam.cloud.ibm.com/identity/token" + _URL_API_PUBLIC = "https://{}.power-iaas.cloud.ibm.com" + _URL_API_PRIVATE = "https://private.{}.power-iaas.cloud.ibm.com" + _URL_API_BASE = "/pcloud/v1/cloud-instances/{}" + + _HTTP_MAX_RETRIES = 10 + _HTTP_BACKOFF_FACTOR = 0.4 + _HTTP_STATUS_FORCE_RETRIES = (403, 500, 502, 503, 504) + _HTTP_RETRY_ALLOWED_METHODS = frozenset({"GET", "POST", "DELETE"}) + + _START_TIME = time.time() + _RESOURCE_ACTION_TIMEOUT = int( + int(os.environ.get("OCF_RESKEY_CRM_meta_timeout", 7200000)) / 1000 + ) + + def __init__( + self, + ip="", + cidr="", + subnet_name="", + api_key="", + api_type="", + region="", + crn_host_map="", + vsi_host_map="", + proxy="", + jumbo="", + use_remote_workspace=False, + ): + """Initialize class variables, including the API token, Cloud Resource Name (CRN), IBM Power Cloud API endpoint URL, and HTTP header.""" + + self._res_options = locals() + + self._validate_and_set_options() + self._set_api_key() + self._set_token() + self._set_header() + + self._instance_check_status() + self.network_id = self._subnet_search_by_cidr() + + def _rest_create_session(self): + """Create a request session with a retry strategy.""" + + # Define the retry strategy + retry_strategy = urllib3.util.Retry( + total=self._HTTP_MAX_RETRIES, # Maximum number of retries + status_forcelist=self._HTTP_STATUS_FORCE_RETRIES, # HTTP status codes to retry on + allowed_methods=self._HTTP_RETRY_ALLOWED_METHODS, # Allowed methods for retry operation + backoff_factor=self._HTTP_BACKOFF_FACTOR, # Sleep for {backoff factor} * (2 ** ({number of previous retries})) + ) + + # Create an HTTP adapter with the retry strategy and mount it to session + adapter = requests.adapters.HTTPAdapter(max_retries=retry_strategy) + + # Create a new session object + session = requests.Session() + session.mount("https://", adapter) + + self._session = session + + return session + + def _rest_api_call(self, method, resource, **kwargs): + """Perform a REST call to the specified URL.""" + + url = self._url + self._base + resource + method = method.upper() + ocf.logger.debug(f"_rest_api_call: {method} {resource}") + + session = self._session or self._rest_create_session() + + r = session.request( + method, url, headers=self._header, proxies=self._proxy, **kwargs + ) + if not r.ok: + raise PowerCloudAPIError( + f"_rest_api_call: {method} call {resource} to {url} failed with reason: {r.reason}, status code: {r.status_code}", + ocf.OCF_ERR_GENERIC, + ) + + return r.json() + + def _set_api_key(self): + """Store an API key in a class variable. + + api_key is a string. If the first character of the string is @, + the rest of the string is assumed to be the name of a file containing the API key. + """ + + api_key = self._res_options["api_key"] + if api_key[0] == "@": + api_key_file = api_key[1:] + try: + with open(api_key_file, "r") as f: + # read the API key from a file + try: + keys = json.loads(f.read()) + # data seems to be in json format + # return the value of the item with the key 'apikey' + api_key = keys.get("apikey", keys) + except ValueError: + # data is text, return as is + api_key = f.read().strip() + except FileNotFoundError: + raise PowerCloudAPIError( + f"_set_api_key: API key file '{api_key_file}' not found", + ocf.OCF_ERR_ARGS, + ) + + self._api_key = api_key + + def _set_token(self): + """Use the stored API key to obtain an IBM Cloud IAM access token.""" + + url = self._URL_IAM_GLOBAL + + headers = { + "content-type": "application/x-www-form-urlencoded", + "accept": "application/json", + } + data = { + "grant_type": "urn:ibm:params:oauth:grant-type:apikey", + "apikey": f"{self._api_key}", + } + token_response = requests.post( + url, headers=headers, data=data, proxies=self._proxy + ) + if token_response.status_code != 200: + raise PowerCloudAPIError( + f"_set_token: failed to obtain token from IBM Cloud IAM: {token_response.status_code}", + ocf.OCF_ERR_GENERIC, + ) + + self._token = json.loads(token_response.text)["access_token"] + + def _set_header(self): + """Set the Cloud Resource Name (CRN), IBM Power Cloud API endpoint URL, and HTTP header.""" + + self._header = { + "Authorization": f"Bearer {self._token}", + "CRN": f"{self._crn}", + "Content-Type": "application/json", + } + + def _instance_check_status(self): + """Check if instance exists in workspace and log the current status.""" + + resource = f"/pvm-instances/{self.instance_id}" + instance = self._rest_api_call("GET", resource) + + server_name = instance["serverName"] + status = instance["status"] + health = instance["health"]["status"] + + if status == "SHUTOFF" or (status == "ACTIVE" and health == "OK"): + ocf.logger.debug( + f"_instance_check_status: OK server_name: {server_name}, status: {status}, health: {health}" + ) + else: + raise PowerCloudAPIError( + f"_instance_check_status: FAIL server_name: {server_name}, status: {status}, health: {health}", + ocf.OCF_ERR_GENERIC, + ) + + def _instance_subnet_is_attached(self): + """Check if a virtual server instance is connected to a specific subnet.""" + + for net in self._instance_subnet_list(): + if self.network_id == net["networkID"]: + return True + return False + + def _instance_subnet_get(self): + """Obtain information about a particular subnet connected to a virtual server instance.""" + + resource = f"/pvm-instances/{self.instance_id}/networks/{self.network_id}" + response = self._rest_api_call("GET", resource) + return response["networks"][0] + + def _instance_subnet_list(self): + """List all subnets connected to a virtual server instance.""" + + resource = f"/pvm-instances/{self.instance_id}/networks" + response = self._rest_api_call("GET", resource) + return response["networks"] + + def _instance_subnet_attach(self): + """Attach a subnet to a virtual server instance.""" + + data = ( + f'{{"networkID":"{self.network_id}","ipAddress":"{self.ip}"}}' + if self.ip + else f'{{"networkID":"{self.network_id}"}}' + ) + + resource = f"/pvm-instances/{self.instance_id}/networks/" + _ = self._rest_api_call("POST", resource, data=data) + + def _instance_subnet_detach(self): + """Detach a subnet from a virtual server instance.""" + + resource = f"/pvm-instances/{self.instance_id}/networks/{self.network_id}" + _ = self._rest_api_call("DELETE", resource) + + def _subnet_create(self): + """Create a subnet in the workspace.""" + + data = ( + f'{{"type":"vlan","cidr":"{self.cidr}","mtu":9000,"name":"{self.subnet_name}"}}' + if self.jumbo + else f'{{"type":"vlan","cidr":"{self.cidr}","name":"{self.subnet_name}"}}' + ) + resource = "/networks" + response = self._rest_api_call("POST", resource, data=data) + self.network_id = response["networkID"] + + def _subnet_delete(self): + """Delete a subnet in the workspace.""" + + resource = f"/networks/{self.network_id}" + _ = self._rest_api_call("DELETE", resource) + + def _subnet_get(self, network_id): + """Get information about a specific subnet in the workspace.""" + + resource = f"/networks/{network_id}" + response = self._rest_api_call("GET", resource) + return response + + def _subnet_list(self): + """List all subnets in the workspace.""" + + resource = "/networks/" + response = self._rest_api_call("GET", resource) + return response + + def _subnet_search_by_cidr(self): + """Find the subnet for a given CIDR.""" + + for network in self._subnet_list()["networks"]: + network_id = network["networkID"] + if self.cidr == self._subnet_get(network_id)["cidr"]: + return network_id + + return None + + def _validate_and_set_options(self): + """Validate the options of the resource agent and derive class variables from the options.""" + + ip = self._res_options["ip"] + try: + validated_ip = ipaddress.ip_address(ip) + except ValueError: + raise PowerCloudAPIError( + f"_validate_and_set_options: {ip} is not a valid IP address.", + ocf.OCF_ERR_CONFIGURED, + ) + self.ip = ip + + cidr = self._res_options["cidr"] + try: + validated_cidr = ipaddress.ip_network(cidr) + except ValueError: + raise PowerCloudAPIError( + f"_validate_and_set_options: {cidr} is not a valid CIDR notation.", + ocf.OCF_ERR_CONFIGURED, + ) + self.cidr = cidr + + if validated_ip not in validated_cidr: + raise PowerCloudAPIError( + f"_validate_and_set_options: {ip} is not in {cidr} range.", + ocf.OCF_ERR_CONFIGURED, + ) + + subnet_name = self._res_options["subnet_name"] + self.subnet_name = subnet_name if subnet_name else self.cidr + + crn_host_map = self._res_options["crn_host_map"] + try: + self._crn_host_map = dict( + item.split(":", 1) for item in crn_host_map.split(";") + ) + except ValueError: + raise PowerCloudAPIError( + f"_validate_and_set_options: crn_host_map: {crn_host_map} has an invalid format.", + ocf.OCF_ERR_CONFIGURED, + ) + + self._hostname = os.uname().nodename + if self._res_options["use_remote_workspace"]: + self._nodename = [k for k in self._crn_host_map if k != self._hostname][0] + else: + self._nodename = self._hostname + + if self._nodename not in self._crn_host_map: + raise PowerCloudAPIError( + f"_validate_and_set_options: {self._nodename} not found in crn_host_map: {crn_host_map}.", + ocf.OCF_ERR_ARGS, + ) + self._crn = self._crn_host_map[self._nodename] + + try: + self._cloud_instance_id = self._crn.split(":")[7] + except IndexError: + raise PowerCloudAPIError( + f"_validate_and_set_options: {self._crn} is not a valid CRN.", + ocf.OCF_ERR_CONFIGURED, + ) + + vsi_host_map = self._res_options["vsi_host_map"] + try: + self._vsi_host_map = dict( + item.split(":") for item in vsi_host_map.split(";") + ) + except ValueError: + raise PowerCloudAPIError( + f"_validate_and_set_options: Option vsi_host_map: {vsi_host_map} has an invalid format.", + ocf.OCF_ERR_CONFIGURED, + ) + + if self._nodename not in self._vsi_host_map: + raise PowerCloudAPIError( + f"_validate_and_set_options: {self._nodename} not found in vsi_host_map: {vsi_host_map}.", + ocf.OCF_ERR_ARGS, + ) + self.instance_id = self._vsi_host_map[self._nodename] + + jumbo = self._res_options["jumbo"].lower() + if ocf.is_true(jumbo): + self.jumbo = True + else: + if jumbo not in ("no", "false", "0", 0, "nein", "off", False): + raise PowerCloudAPIError( + f"_validate_and_set_options: option jumbo: {jumbo} does not match True or False.", + ocf.OCF_ERR_CONFIGURED, + ) + self.jumbo = False + + # Check connect to proxy server + self._proxy = "" + proxy = self._res_options["proxy"] + + if proxy: + # extract ip address and port + match = re.search(r"^https?://([^:]+):(\d+)$", proxy) + if match: + proxy_ip, proxy_port = match.group(1), match.group(2) + + try: + with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s: + s.settimeout(30) + s.connect((proxy_ip, int(proxy_port))) + except socket.error: + raise PowerCloudAPIError( + f"_validate_and_set_options: cannot connect to port {proxy_port} at {proxy_ip}, check option proxy: {proxy}.", + ocf.OCF_ERR_CONFIGURED, + ) + self._proxy = {"https": f"{proxy}"} + else: + raise PowerCloudAPIError( + f"_validate_and_set_options: the option proxy: {proxy} has an invalid format.", + ocf.OCF_ERR_CONFIGURED, + ) + + api_type = self._res_options["api_type"] + if api_type not in ("public", "private"): + raise PowerCloudAPIError( + f"_validate_and_set_options: option api_type: {api_type} does not match public or private.", + ocf.OCF_ERR_CONFIGURED, + ) + # Set API endpoint url + url_api_fmt = ( + self._URL_API_PRIVATE if api_type == "private" else self._URL_API_PUBLIC + ) + self._url = url_api_fmt.format(self._res_options["region"]) + self._base = self._URL_API_BASE.format(self._cloud_instance_id) + self._session = None + + def subnet_add(self): + """Create and attach subnet in local workspace""" + + ocf.logger.debug( + f"subnet_add: options: ip: {self.ip}, cidr: {self.cidr}, name: {self.subnet_name}" + ) + + if self.network_id: + ocf.logger.debug( + f"subnet_add: subnet cidr: {self.cidr} already exists with network id: {self.network_id}" + ) + else: + ocf.logger.debug( + f"subnet_add: create subnet name: {self.subnet_name} with cidr: {self.cidr} and jumbo: {self.jumbo}" + ) + self._subnet_create() + + if self._instance_subnet_is_attached(): + ocf.logger.debug( + f"subnet_add: subnet id {self.network_id} is already attached to instance id {self.instance_id}" + ) + else: + ocf.logger.debug( + f"subnet_add: attach subnet id: {self.network_id} to instance id: {self.instance_id} (IP address {self.ip})" + ) + self._instance_subnet_attach() + + subnet = self._subnet_get(self.network_id) + gateway = subnet["gateway"] + port = self._instance_subnet_get() + mac = port["macAddress"] + ip_address = port["ipAddress"] + self.jumbo = subnet.get("mtu", "") == 9000 + + timeout = self._RESOURCE_ACTION_TIMEOUT - int(time.time() - self._START_TIME) + nic = nmcli.wait_for_nic(mac, timeout) + + return nic, ip_address, mac, gateway + + def subnet_remove(self): + """Detach and delete subnet in local or remote workspace""" + + ocf.logger.debug( + f"subnet_remove: options: cidr: {self.cidr}, network id: {self.network_id}, instance id: {self.instance_id}" + ) + + if self.network_id: + ocf.logger.debug( + f"subnet_remove: subnet id: {self.network_id} with cidr: {self.cidr} exists" + ) + if self._instance_subnet_is_attached(): + ocf.logger.debug( + f"subnet_remove: subnet id: {self.network_id} is attached to instance id {self.instance_id}" + ) + port = self._instance_subnet_get() + mac = port["macAddress"] + dev = nmcli.device.find("GENERAL.HWADDR", mac.upper()) + + if dev: + nm_object = nmcli.connection.find( + "GENERAL.IP-IFACE", dev["GENERAL.DEVICE"] + ) + if nm_object: + conn_name = nm_object["connection.id"] + ocf.logger.debug( + f"stop_action: unconfigure network connection conn_name: {conn_name} with mac address {mac}" + ) + nmcli.connection.down(conn_name) + nmcli.connection.delete(conn_name) + ocf.logger.debug( + f"subnet_remove: detach network id: {self.network_id} from instance id: {self.instance_id}" + ) + self._instance_subnet_detach() + ocf.logger.debug(f"subnet_remove: delete network id: {self.network_id}") + self._subnet_delete() + + +def os_ping(ip): + """Ping an IP address.""" + + command = ["ping", "-c", "1", ip] + response = subprocess.call(command) + return response == 0 + + +def start_action( + ip="", + cidr="", + subnet_name="", + api_key="", + api_type="", + region="", + crn_host_map="", + vsi_host_map="", + proxy="", + jumbo="", +): + """start_action: assign the service ip. + + Create a subnet in the workspace, connect it to the virtual server instance, and configure the NIC. + """ + + res_options = locals() + + ocf.logger.info(f"start_action: options: {res_options}") + + # Detach and remove subnet in remote workspace + remote_ws = PowerCloudAPI(**res_options, use_remote_workspace=True) + ocf.logger.debug( + f"start_action: remove subnet from remote workspace: cidr: {remote_ws.cidr}" + ) + remote_ws.subnet_remove() + + # Delete orphaned Network Manager connections + nmcli.cleanup() + + # Create and attach subnet in local workspace + ws = PowerCloudAPI(**res_options) + + nic, ip_address, mac, gateway = ws.subnet_add() + + ocf.logger.debug( + f"start_action: add nmcli connection: nic: {nic}, ip: {ip_address}, mac: {mac}, gateway: {gateway}, jumbo: {ws.jumbo}" + ) + + conn_name = f"{nmcli.CONN_PREFIX}{nic}" + conn_options = { + "ifname": nic, + "autoconnect": "no", + "ipv4.method": "manual", + "ipv4.addresses": ip_address, + "ipv4.routes": f"0.0.0.0/0 {gateway} table={nmcli.ROUTING_TABLE}", + "ipv4.routing-rules": f"priority {nmcli.ROUTING_PRIO} from {ws.cidr} table {nmcli.ROUTING_TABLE}", + } + if ws.jumbo: + conn_options.update({"802-3-ethernet.mtu": "9000", "ethtool.feature-tso": "on"}) + + nmcli.connection.add(conn_name, options=conn_options) + nmcli.connection.up(conn_name) + + if monitor_action(**res_options) != ocf.OCF_SUCCESS: + raise PowerCloudAPIError(f"start_action: start subnet: {ws.subnet_name} failed") + + ocf.logger.info( + f"start_action: finished, added connection {conn_name} for subnet {ws.subnet_name}" + ) + + return ocf.OCF_SUCCESS + + +def stop_action( + ip="", + cidr="", + subnet_name="", + api_key="", + api_type="", + region="", + crn_host_map="", + vsi_host_map="", + proxy="", + jumbo="", +): + """stop_action: unassign the service ip. + + Delete NIC, detach subnet from virtual server instance, and delete subnet. + """ + + res_options = locals() + + ocf.logger.info(f"stop_action: options: {res_options}") + + ws = PowerCloudAPI(**res_options) + + ws.subnet_remove() + + if monitor_action(**res_options) != ocf.OCF_NOT_RUNNING: + raise PowerCloudAPIError(f"stop_action: stop subnet {ws.subnet_name} failed") + + ocf.logger.info( + f"stop_action: finished, deleted connection for subnet {ws.subnet_name}" + ) + + return ocf.OCF_SUCCESS + + +def monitor_action( + ip="", + cidr="", + subnet_name="", + api_key="", + api_type="", + region="", + crn_host_map="", + vsi_host_map="", + proxy="", + jumbo="", +): + """monitor_action: check if service ip and gateway are responding.""" + + res_options = locals() + is_probe = ocf.is_probe() + + ocf.logger.info(f"monitor_action: options: {res_options}, is_probe: {is_probe}") + + gateway = nmcli.find_gateway(ip) + if gateway and os_ping(gateway): + if os_ping(ip): + ocf.logger.debug( + f"monitor_action: ping to gateway: {gateway} and ip: {ip} successful" + ) + return ocf.OCF_SUCCESS + else: + raise PowerCloudAPIError( + f"monitor_action: ping to ip: {ip} failed", ocf.OCF_ERR_GENERIC + ) + + if not is_probe: + ocf.logger.error(f"monitor_action: ping to gateway: {gateway} failed") + + ws = PowerCloudAPI(**res_options) + + ocf.logger.debug(f"monitor_action: instance id: {ws.instance_id}") + + if not ws.network_id or is_probe: + return ocf.OCF_NOT_RUNNING + + # monitor should never reach this code, exit with raise + raise PowerCloudAPIError( + f"monitor_action: unknown problem with subnet id: {ws.network_id}", + ocf.OCF_ERR_GENERIC, + ) + + +def validate_all_action( + ip="", + cidr="", + subnet_name="", + api_key="", + api_type="", + region="", + crn_host_map="", + vsi_host_map="", + proxy="", + jumbo="", +): + """validate_all_action: Validate the resource agent parameters.""" + + res_options = locals() + + # The class instantation validates the resource agent options and that the instance exists + try: + # Check instance in local workspace + _ = PowerCloudAPI(**res_options, use_remote_workspace=False) + except Exception: + ocf.logger.error( + "validate_all_action: failed to instantiate class in local workspace." + ) + raise + + try: + # Check instance in remote workspace + _ = PowerCloudAPI(**res_options, use_remote_workspace=True) + except Exception: + ocf.logger.error( + "validate_all_action: failed to instantiate class in remote workspace." + ) + raise + + return ocf.OCF_SUCCESS + + +def main(): + """Instantiate the resource agent.""" + + agent_description = textwrap.dedent("""\ + Resource Agent to move a Power Virtual Server subnet and its IP address + from one virtual server instance to another. + The prerequisites for the use of this resource agent are as follows: + + 1. Red Hat Enterprise Linux 9.2 or higher: + Install with @server group to ensure that NetworkManager settings are correct. + Verify that the NetworkManager-config-server package is installed. + + 2. IBM Cloud API Key: + Create a service API key that is privileged for both Power Virtual Server + workspaces. Save the service API key in a file and copy the file to both + cluster nodes. Use same filename and directory location on both cluster nodes. + Reference the path to the key file in the resource definition. + + 3. The hostname of the virtual server instances must be same as the name + of the virtual server instances in the Power Virtual Server workspaces. + + For comprehensive documentation on implementing high availability for + SAP applications on IBM Power Virtual Server, visit https://cloud.ibm.com/docs/sap?topic=sap-ha-overview. + """) + + agent = ocf.Agent( + "powervs-subnet", + shortdesc="Manages moving a Power Virtual Server subnet", + longdesc=agent_description, + version=1.0, + ) + + agent.add_parameter( + "ip", + shortdesc="IP address", + longdesc=( + "IP address within the subnet. The IP address moves together with the subnet." + ), + content_type="string", + required=True, + ) + + agent.add_parameter( + "cidr", + shortdesc="CIDR", + longdesc="Classless Inter-Domain Routing (CIDR) of the subnet.", + content_type="string", + required=True, + ) + + agent.add_parameter( + "subnet_name", + shortdesc="Name of the subnet", + longdesc="Name of the subnet. If not specified, CIDR is used as name.", + content_type="string", + required=False, + ) + + agent.add_parameter( + "api_type", + shortdesc="API type", + longdesc="Connect to Power Virtual Server regional endpoints over a public or private network (public|private).", + content_type="string", + required=True, + default="private", + ) + + agent.add_parameter( + "region", + shortdesc="Power Virtual Server region", + longdesc=( + "Region that represents the geographic area where the instance is located. " + "The region is used to identify the Cloud API endpoint." + ), + content_type="string", + required=True, + ) + + agent.add_parameter( + "api_key", + shortdesc="API Key or @API_KEY_FILE_PATH", + longdesc=( + "API Key or @API_KEY_FILE_PATH for IBM Cloud access. " + "The API key content or the path of an API key file that is indicated by the @ symbol." + ), + content_type="string", + required=True, + ) + + agent.add_parameter( + "crn_host_map", + shortdesc="Mapping of hostnames to IBM Cloud CRN", + longdesc=( + "Map the hostname of the Power Virtual Server instance to the CRN of the Power Virtual Server workspaces hosting the instance. " + "Separate hostname and CRN with a colon ':', separate different hostname and CRN pairs with a semicolon ';'. " + "Example: hostname01:CRN-of-Instance01;hostname02:CRN-of-Instance02" + ), + content_type="string", + required=True, + ) + + agent.add_parameter( + "vsi_host_map", + shortdesc="Mapping of hostnames to PowerVS instance ids", + longdesc=( + "Map the hostname of the Power Virtual Server instance to its instance id. " + "Separate hostname and instance id with a colon ':', separate different hostname and instance id pairs with a semicolon ';'. " + "Example: hostname01:instance-id-01;hostname02:instance-id-02" + ), + content_type="string", + required=True, + ) + + agent.add_parameter( + "proxy", + shortdesc="Proxy", + longdesc="Proxy server to access IBM Cloud API endpoints.", + content_type="string", + required=False, + ) + + agent.add_parameter( + "jumbo", + shortdesc="Use Jumbo frames", + longdesc="Create a Power Virtual Server subnet with an MTU size of 9000 (true|false).", + content_type="string", + required=False, + default="false", + ) + + agent.add_action("start", timeout=900, handler=start_action) + agent.add_action("stop", timeout=450, handler=stop_action) + agent.add_action( + "monitor", depth=0, timeout=60, interval=60, handler=monitor_action + ) + agent.add_action("validate-all", timeout=300, handler=validate_all_action) + agent.run() + + +if __name__ == "__main__": + main()