diff --git a/cts/scheduler/summary/bug-1572-1.summary b/cts/scheduler/summary/bug-1572-1.summary index 6abedea530..c572db21d5 100644 --- a/cts/scheduler/summary/bug-1572-1.summary +++ b/cts/scheduler/summary/bug-1572-1.summary @@ -1,85 +1,85 @@ Current cluster status: * Node List: * Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ] * Full List of Resources: * Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable): * Promoted: [ arc-tkincaidlx.wsicorp.com ] * Unpromoted: [ arc-dknightlx ] * Resource Group: grp_pgsql_mirror: * fs_mirror (ocf:heartbeat:Filesystem): Started arc-tkincaidlx.wsicorp.com * pgsql_5555 (ocf:heartbeat:pgsql): Started arc-tkincaidlx.wsicorp.com * IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Started arc-tkincaidlx.wsicorp.com Transition Summary: - * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability + * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability * Restart rsc_drbd_7788:1 ( Promoted arc-tkincaidlx.wsicorp.com ) due to resource definition change * Restart fs_mirror ( arc-tkincaidlx.wsicorp.com ) due to required ms_drbd_7788 notified * Restart pgsql_5555 ( arc-tkincaidlx.wsicorp.com ) due to required fs_mirror start * Restart IPaddr_147_81_84_133 ( arc-tkincaidlx.wsicorp.com ) due to required pgsql_5555 start Executing Cluster Transition: * Pseudo action: ms_drbd_7788_pre_notify_demote_0 * Pseudo action: grp_pgsql_mirror_stop_0 * Resource action: IPaddr_147_81_84_133 stop on arc-tkincaidlx.wsicorp.com * Resource action: rsc_drbd_7788:0 notify on arc-dknightlx * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-pre_notify_demote_0 * Resource action: pgsql_5555 stop on arc-tkincaidlx.wsicorp.com * Resource action: fs_mirror stop on arc-tkincaidlx.wsicorp.com * Pseudo action: grp_pgsql_mirror_stopped_0 * Pseudo action: ms_drbd_7788_demote_0 * Resource action: rsc_drbd_7788:1 demote on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_demoted_0 * Pseudo action: ms_drbd_7788_post_notify_demoted_0 * Resource action: rsc_drbd_7788:0 notify on arc-dknightlx * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-post_notify_demoted_0 * Pseudo action: ms_drbd_7788_pre_notify_stop_0 * Resource action: rsc_drbd_7788:0 notify on arc-dknightlx * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-pre_notify_stop_0 * Pseudo action: ms_drbd_7788_stop_0 * Resource action: rsc_drbd_7788:0 stop on arc-dknightlx * Resource action: rsc_drbd_7788:1 stop on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_stopped_0 * Cluster action: do_shutdown on arc-dknightlx * Pseudo action: ms_drbd_7788_post_notify_stopped_0 * Pseudo action: ms_drbd_7788_confirmed-post_notify_stopped_0 * Pseudo action: ms_drbd_7788_pre_notify_start_0 * Pseudo action: ms_drbd_7788_confirmed-pre_notify_start_0 * Pseudo action: ms_drbd_7788_start_0 * Resource action: rsc_drbd_7788:1 start on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_running_0 * Pseudo action: ms_drbd_7788_post_notify_running_0 * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-post_notify_running_0 * Pseudo action: ms_drbd_7788_pre_notify_promote_0 * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-pre_notify_promote_0 * Pseudo action: ms_drbd_7788_promote_0 * Resource action: rsc_drbd_7788:1 promote on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_promoted_0 * Pseudo action: ms_drbd_7788_post_notify_promoted_0 * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-post_notify_promoted_0 * Pseudo action: grp_pgsql_mirror_start_0 * Resource action: fs_mirror start on arc-tkincaidlx.wsicorp.com * Resource action: pgsql_5555 start on arc-tkincaidlx.wsicorp.com * Resource action: pgsql_5555 monitor=30000 on arc-tkincaidlx.wsicorp.com * Resource action: IPaddr_147_81_84_133 start on arc-tkincaidlx.wsicorp.com * Resource action: IPaddr_147_81_84_133 monitor=25000 on arc-tkincaidlx.wsicorp.com * Pseudo action: grp_pgsql_mirror_running_0 Revised Cluster Status: * Node List: * Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ] * Full List of Resources: * Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable): * Promoted: [ arc-tkincaidlx.wsicorp.com ] * Stopped: [ arc-dknightlx ] * Resource Group: grp_pgsql_mirror: * fs_mirror (ocf:heartbeat:Filesystem): Started arc-tkincaidlx.wsicorp.com * pgsql_5555 (ocf:heartbeat:pgsql): Started arc-tkincaidlx.wsicorp.com * IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Started arc-tkincaidlx.wsicorp.com diff --git a/cts/scheduler/summary/bug-1572-2.summary b/cts/scheduler/summary/bug-1572-2.summary index 7d4921dc36..012ca78dd6 100644 --- a/cts/scheduler/summary/bug-1572-2.summary +++ b/cts/scheduler/summary/bug-1572-2.summary @@ -1,61 +1,61 @@ Current cluster status: * Node List: * Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ] * Full List of Resources: * Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable): * Promoted: [ arc-tkincaidlx.wsicorp.com ] * Unpromoted: [ arc-dknightlx ] * Resource Group: grp_pgsql_mirror: * fs_mirror (ocf:heartbeat:Filesystem): Started arc-tkincaidlx.wsicorp.com * pgsql_5555 (ocf:heartbeat:pgsql): Started arc-tkincaidlx.wsicorp.com * IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Started arc-tkincaidlx.wsicorp.com Transition Summary: - * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability + * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability * Demote rsc_drbd_7788:1 ( Promoted -> Unpromoted arc-tkincaidlx.wsicorp.com ) * Stop fs_mirror ( arc-tkincaidlx.wsicorp.com ) due to node availability * Stop pgsql_5555 ( arc-tkincaidlx.wsicorp.com ) due to node availability * Stop IPaddr_147_81_84_133 ( arc-tkincaidlx.wsicorp.com ) due to node availability Executing Cluster Transition: * Pseudo action: ms_drbd_7788_pre_notify_demote_0 * Pseudo action: grp_pgsql_mirror_stop_0 * Resource action: IPaddr_147_81_84_133 stop on arc-tkincaidlx.wsicorp.com * Resource action: rsc_drbd_7788:0 notify on arc-dknightlx * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-pre_notify_demote_0 * Resource action: pgsql_5555 stop on arc-tkincaidlx.wsicorp.com * Resource action: fs_mirror stop on arc-tkincaidlx.wsicorp.com * Pseudo action: grp_pgsql_mirror_stopped_0 * Pseudo action: ms_drbd_7788_demote_0 * Resource action: rsc_drbd_7788:1 demote on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_demoted_0 * Pseudo action: ms_drbd_7788_post_notify_demoted_0 * Resource action: rsc_drbd_7788:0 notify on arc-dknightlx * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-post_notify_demoted_0 * Pseudo action: ms_drbd_7788_pre_notify_stop_0 * Resource action: rsc_drbd_7788:0 notify on arc-dknightlx * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-pre_notify_stop_0 * Pseudo action: ms_drbd_7788_stop_0 * Resource action: rsc_drbd_7788:0 stop on arc-dknightlx * Pseudo action: ms_drbd_7788_stopped_0 * Cluster action: do_shutdown on arc-dknightlx * Pseudo action: ms_drbd_7788_post_notify_stopped_0 * Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com * Pseudo action: ms_drbd_7788_confirmed-post_notify_stopped_0 Revised Cluster Status: * Node List: * Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ] * Full List of Resources: * Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable): * Unpromoted: [ arc-tkincaidlx.wsicorp.com ] * Stopped: [ arc-dknightlx ] * Resource Group: grp_pgsql_mirror: * fs_mirror (ocf:heartbeat:Filesystem): Stopped * pgsql_5555 (ocf:heartbeat:pgsql): Stopped * IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Stopped diff --git a/cts/scheduler/summary/bug-5059.summary b/cts/scheduler/summary/bug-5059.summary index a33a2f60a2..c555d1dfb5 100644 --- a/cts/scheduler/summary/bug-5059.summary +++ b/cts/scheduler/summary/bug-5059.summary @@ -1,77 +1,77 @@ Current cluster status: * Node List: * Node gluster03.h: standby * Online: [ gluster01.h gluster02.h ] * OFFLINE: [ gluster04.h ] * Full List of Resources: * Clone Set: ms_stateful [g_stateful] (promotable): * Resource Group: g_stateful:0: * p_stateful1 (ocf:pacemaker:Stateful): Unpromoted gluster01.h * p_stateful2 (ocf:pacemaker:Stateful): Stopped * Resource Group: g_stateful:1: * p_stateful1 (ocf:pacemaker:Stateful): Unpromoted gluster02.h * p_stateful2 (ocf:pacemaker:Stateful): Stopped * Stopped: [ gluster03.h gluster04.h ] * Clone Set: c_dummy [p_dummy1]: * Started: [ gluster01.h gluster02.h ] Transition Summary: - * Promote p_stateful1:0 ( Unpromoted -> Promoted gluster01.h ) - * Promote p_stateful2:0 ( Stopped -> Promoted gluster01.h ) + * Promote p_stateful1:0 ( Unpromoted -> Promoted gluster01.h ) + * Promote p_stateful2:0 ( Stopped -> Promoted gluster01.h ) * Start p_stateful2:1 ( gluster02.h ) Executing Cluster Transition: * Pseudo action: ms_stateful_pre_notify_start_0 * Resource action: iptest delete on gluster02.h * Resource action: ipsrc2 delete on gluster02.h * Resource action: p_stateful1:0 notify on gluster01.h * Resource action: p_stateful1:1 notify on gluster02.h * Pseudo action: ms_stateful_confirmed-pre_notify_start_0 * Pseudo action: ms_stateful_start_0 * Pseudo action: g_stateful:0_start_0 * Resource action: p_stateful2:0 start on gluster01.h * Pseudo action: g_stateful:1_start_0 * Resource action: p_stateful2:1 start on gluster02.h * Pseudo action: g_stateful:0_running_0 * Pseudo action: g_stateful:1_running_0 * Pseudo action: ms_stateful_running_0 * Pseudo action: ms_stateful_post_notify_running_0 * Resource action: p_stateful1:0 notify on gluster01.h * Resource action: p_stateful2:0 notify on gluster01.h * Resource action: p_stateful1:1 notify on gluster02.h * Resource action: p_stateful2:1 notify on gluster02.h * Pseudo action: ms_stateful_confirmed-post_notify_running_0 * Pseudo action: ms_stateful_pre_notify_promote_0 * Resource action: p_stateful1:0 notify on gluster01.h * Resource action: p_stateful2:0 notify on gluster01.h * Resource action: p_stateful1:1 notify on gluster02.h * Resource action: p_stateful2:1 notify on gluster02.h * Pseudo action: ms_stateful_confirmed-pre_notify_promote_0 * Pseudo action: ms_stateful_promote_0 * Pseudo action: g_stateful:0_promote_0 * Resource action: p_stateful1:0 promote on gluster01.h * Resource action: p_stateful2:0 promote on gluster01.h * Pseudo action: g_stateful:0_promoted_0 * Pseudo action: ms_stateful_promoted_0 * Pseudo action: ms_stateful_post_notify_promoted_0 * Resource action: p_stateful1:0 notify on gluster01.h * Resource action: p_stateful2:0 notify on gluster01.h * Resource action: p_stateful1:1 notify on gluster02.h * Resource action: p_stateful2:1 notify on gluster02.h * Pseudo action: ms_stateful_confirmed-post_notify_promoted_0 * Resource action: p_stateful1:1 monitor=10000 on gluster02.h * Resource action: p_stateful2:1 monitor=10000 on gluster02.h Revised Cluster Status: * Node List: * Node gluster03.h: standby * Online: [ gluster01.h gluster02.h ] * OFFLINE: [ gluster04.h ] * Full List of Resources: * Clone Set: ms_stateful [g_stateful] (promotable): * Promoted: [ gluster01.h ] * Unpromoted: [ gluster02.h ] * Clone Set: c_dummy [p_dummy1]: * Started: [ gluster01.h gluster02.h ] diff --git a/cts/scheduler/summary/bug-cl-5212.summary b/cts/scheduler/summary/bug-cl-5212.summary index 48cb54bedc..e7a6e26833 100644 --- a/cts/scheduler/summary/bug-cl-5212.summary +++ b/cts/scheduler/summary/bug-cl-5212.summary @@ -1,69 +1,69 @@ Current cluster status: * Node List: * Node srv01: UNCLEAN (offline) * Node srv02: UNCLEAN (offline) * Online: [ srv03 ] * Full List of Resources: * Resource Group: grpStonith1: * prmStonith1-1 (stonith:external/ssh): Started srv02 (UNCLEAN) * Resource Group: grpStonith2: * prmStonith2-1 (stonith:external/ssh): Started srv01 (UNCLEAN) * Resource Group: grpStonith3: * prmStonith3-1 (stonith:external/ssh): Started srv01 (UNCLEAN) * Clone Set: msPostgresql [pgsql] (promotable): * pgsql (ocf:pacemaker:Stateful): Unpromoted srv02 (UNCLEAN) * pgsql (ocf:pacemaker:Stateful): Promoted srv01 (UNCLEAN) * Unpromoted: [ srv03 ] * Clone Set: clnPingd [prmPingd]: * prmPingd (ocf:pacemaker:ping): Started srv02 (UNCLEAN) * prmPingd (ocf:pacemaker:ping): Started srv01 (UNCLEAN) * Started: [ srv03 ] Transition Summary: * Stop prmStonith1-1 ( srv02 ) blocked * Stop prmStonith2-1 ( srv01 ) blocked * Stop prmStonith3-1 ( srv01 ) due to node availability (blocked) - * Stop pgsql:0 ( Unpromoted srv02 ) due to node availability (blocked) - * Stop pgsql:1 ( Promoted srv01 ) due to node availability (blocked) + * Stop pgsql:0 ( Unpromoted srv02 ) due to node availability (blocked) + * Stop pgsql:1 ( Promoted srv01 ) due to node availability (blocked) * Stop prmPingd:0 ( srv02 ) due to node availability (blocked) * Stop prmPingd:1 ( srv01 ) due to node availability (blocked) Executing Cluster Transition: * Pseudo action: grpStonith1_stop_0 * Pseudo action: grpStonith1_start_0 * Pseudo action: grpStonith2_stop_0 * Pseudo action: grpStonith2_start_0 * Pseudo action: grpStonith3_stop_0 * Pseudo action: msPostgresql_pre_notify_stop_0 * Pseudo action: clnPingd_stop_0 * Resource action: pgsql notify on srv03 * Pseudo action: msPostgresql_confirmed-pre_notify_stop_0 * Pseudo action: msPostgresql_stop_0 * Pseudo action: clnPingd_stopped_0 * Pseudo action: msPostgresql_stopped_0 * Pseudo action: msPostgresql_post_notify_stopped_0 * Resource action: pgsql notify on srv03 * Pseudo action: msPostgresql_confirmed-post_notify_stopped_0 Revised Cluster Status: * Node List: * Node srv01: UNCLEAN (offline) * Node srv02: UNCLEAN (offline) * Online: [ srv03 ] * Full List of Resources: * Resource Group: grpStonith1: * prmStonith1-1 (stonith:external/ssh): Started srv02 (UNCLEAN) * Resource Group: grpStonith2: * prmStonith2-1 (stonith:external/ssh): Started srv01 (UNCLEAN) * Resource Group: grpStonith3: * prmStonith3-1 (stonith:external/ssh): Started srv01 (UNCLEAN) * Clone Set: msPostgresql [pgsql] (promotable): * pgsql (ocf:pacemaker:Stateful): Unpromoted srv02 (UNCLEAN) * pgsql (ocf:pacemaker:Stateful): Promoted srv01 (UNCLEAN) * Unpromoted: [ srv03 ] * Clone Set: clnPingd [prmPingd]: * prmPingd (ocf:pacemaker:ping): Started srv02 (UNCLEAN) * prmPingd (ocf:pacemaker:ping): Started srv01 (UNCLEAN) * Started: [ srv03 ] diff --git a/cts/scheduler/summary/bug-cl-5247.summary b/cts/scheduler/summary/bug-cl-5247.summary index 056e526490..67ad0c3ded 100644 --- a/cts/scheduler/summary/bug-cl-5247.summary +++ b/cts/scheduler/summary/bug-cl-5247.summary @@ -1,87 +1,87 @@ Using the original execution date of: 2015-08-12 02:53:40Z Current cluster status: * Node List: * Online: [ bl460g8n3 bl460g8n4 ] * GuestOnline: [ pgsr01@bl460g8n3 ] * Full List of Resources: * prmDB1 (ocf:heartbeat:VirtualDomain): Started bl460g8n3 * prmDB2 (ocf:heartbeat:VirtualDomain): FAILED bl460g8n4 * Resource Group: grpStonith1: * prmStonith1-2 (stonith:external/ipmi): Started bl460g8n4 * Resource Group: grpStonith2: * prmStonith2-2 (stonith:external/ipmi): Started bl460g8n3 * Resource Group: master-group: * vip-master (ocf:heartbeat:Dummy): FAILED pgsr02 * vip-rep (ocf:heartbeat:Dummy): FAILED pgsr02 * Clone Set: msPostgresql [pgsql] (promotable): * Promoted: [ pgsr01 ] * Stopped: [ bl460g8n3 bl460g8n4 ] Transition Summary: * Fence (off) pgsr02 (resource: prmDB2) 'guest is unclean' * Stop prmDB2 ( bl460g8n4 ) due to node availability * Recover vip-master ( pgsr02 -> pgsr01 ) * Recover vip-rep ( pgsr02 -> pgsr01 ) - * Stop pgsql:0 ( Promoted pgsr02 ) due to node availability + * Stop pgsql:0 ( Promoted pgsr02 ) due to node availability * Stop pgsr02 ( bl460g8n4 ) due to node availability Executing Cluster Transition: * Resource action: vip-master monitor on pgsr01 * Resource action: vip-rep monitor on pgsr01 * Pseudo action: msPostgresql_pre_notify_demote_0 * Resource action: pgsr01 monitor on bl460g8n4 * Resource action: pgsr02 stop on bl460g8n4 * Resource action: pgsr02 monitor on bl460g8n3 * Resource action: prmDB2 stop on bl460g8n4 * Resource action: pgsql notify on pgsr01 * Pseudo action: msPostgresql_confirmed-pre_notify_demote_0 * Pseudo action: msPostgresql_demote_0 * Pseudo action: stonith-pgsr02-off on pgsr02 * Pseudo action: pgsql_post_notify_stop_0 * Pseudo action: pgsql_demote_0 * Pseudo action: msPostgresql_demoted_0 * Pseudo action: msPostgresql_post_notify_demoted_0 * Resource action: pgsql notify on pgsr01 * Pseudo action: msPostgresql_confirmed-post_notify_demoted_0 * Pseudo action: msPostgresql_pre_notify_stop_0 * Pseudo action: master-group_stop_0 * Pseudo action: vip-rep_stop_0 * Resource action: pgsql notify on pgsr01 * Pseudo action: msPostgresql_confirmed-pre_notify_stop_0 * Pseudo action: msPostgresql_stop_0 * Pseudo action: vip-master_stop_0 * Pseudo action: pgsql_stop_0 * Pseudo action: msPostgresql_stopped_0 * Pseudo action: master-group_stopped_0 * Pseudo action: master-group_start_0 * Resource action: vip-master start on pgsr01 * Resource action: vip-rep start on pgsr01 * Pseudo action: msPostgresql_post_notify_stopped_0 * Pseudo action: master-group_running_0 * Resource action: vip-master monitor=10000 on pgsr01 * Resource action: vip-rep monitor=10000 on pgsr01 * Resource action: pgsql notify on pgsr01 * Pseudo action: msPostgresql_confirmed-post_notify_stopped_0 * Pseudo action: pgsql_notified_0 * Resource action: pgsql monitor=9000 on pgsr01 Using the original execution date of: 2015-08-12 02:53:40Z Revised Cluster Status: * Node List: * Online: [ bl460g8n3 bl460g8n4 ] * GuestOnline: [ pgsr01@bl460g8n3 ] * Full List of Resources: * prmDB1 (ocf:heartbeat:VirtualDomain): Started bl460g8n3 * prmDB2 (ocf:heartbeat:VirtualDomain): FAILED * Resource Group: grpStonith1: * prmStonith1-2 (stonith:external/ipmi): Started bl460g8n4 * Resource Group: grpStonith2: * prmStonith2-2 (stonith:external/ipmi): Started bl460g8n3 * Resource Group: master-group: * vip-master (ocf:heartbeat:Dummy): FAILED [ pgsr01 pgsr02 ] * vip-rep (ocf:heartbeat:Dummy): FAILED [ pgsr01 pgsr02 ] * Clone Set: msPostgresql [pgsql] (promotable): * Promoted: [ pgsr01 ] * Stopped: [ bl460g8n3 bl460g8n4 ] diff --git a/cts/scheduler/summary/bug-lf-2606.summary b/cts/scheduler/summary/bug-lf-2606.summary index e0b7ebf0e6..004788e80b 100644 --- a/cts/scheduler/summary/bug-lf-2606.summary +++ b/cts/scheduler/summary/bug-lf-2606.summary @@ -1,46 +1,46 @@ 1 of 5 resource instances DISABLED and 0 BLOCKED from further action due to failure Current cluster status: * Node List: * Node node2: UNCLEAN (online) * Online: [ node1 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * rsc1 (ocf:pacemaker:Dummy): FAILED node2 (disabled) * rsc2 (ocf:pacemaker:Dummy): Started node2 * Clone Set: ms3 [rsc3] (promotable): * Promoted: [ node2 ] * Unpromoted: [ node1 ] Transition Summary: * Fence (reboot) node2 'rsc1 failed there' * Stop rsc1 ( node2 ) due to node availability * Move rsc2 ( node2 -> node1 ) - * Stop rsc3:1 ( Promoted node2 ) due to node availability + * Stop rsc3:1 ( Promoted node2 ) due to node availability Executing Cluster Transition: * Pseudo action: ms3_demote_0 * Fencing node2 (reboot) * Pseudo action: rsc1_stop_0 * Pseudo action: rsc2_stop_0 * Pseudo action: rsc3:1_demote_0 * Pseudo action: ms3_demoted_0 * Pseudo action: ms3_stop_0 * Resource action: rsc2 start on node1 * Pseudo action: rsc3:1_stop_0 * Pseudo action: ms3_stopped_0 * Resource action: rsc2 monitor=10000 on node1 Revised Cluster Status: * Node List: * Online: [ node1 ] * OFFLINE: [ node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * rsc1 (ocf:pacemaker:Dummy): Stopped (disabled) * rsc2 (ocf:pacemaker:Dummy): Started node1 * Clone Set: ms3 [rsc3] (promotable): * Unpromoted: [ node1 ] * Stopped: [ node2 ] diff --git a/cts/scheduler/summary/bug-pm-12.summary b/cts/scheduler/summary/bug-pm-12.summary index 7b811d1a02..c4f3adb908 100644 --- a/cts/scheduler/summary/bug-pm-12.summary +++ b/cts/scheduler/summary/bug-pm-12.summary @@ -1,57 +1,57 @@ Current cluster status: * Node List: * Online: [ node-a node-b ] * Full List of Resources: * Clone Set: ms-sf [group] (promotable) (unique): * Resource Group: group:0: * stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b * stateful-2:0 (ocf:heartbeat:Stateful): Unpromoted node-b * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a * stateful-2:1 (ocf:heartbeat:Stateful): Promoted node-a Transition Summary: - * Restart stateful-2:0 ( Unpromoted node-b ) due to resource definition change - * Restart stateful-2:1 ( Promoted node-a ) due to resource definition change + * Restart stateful-2:0 ( Unpromoted node-b ) due to resource definition change + * Restart stateful-2:1 ( Promoted node-a ) due to resource definition change Executing Cluster Transition: * Pseudo action: ms-sf_demote_0 * Pseudo action: group:1_demote_0 * Resource action: stateful-2:1 demote on node-a * Pseudo action: group:1_demoted_0 * Pseudo action: ms-sf_demoted_0 * Pseudo action: ms-sf_stop_0 * Pseudo action: group:0_stop_0 * Resource action: stateful-2:0 stop on node-b * Pseudo action: group:1_stop_0 * Resource action: stateful-2:1 stop on node-a * Pseudo action: group:0_stopped_0 * Pseudo action: group:1_stopped_0 * Pseudo action: ms-sf_stopped_0 * Pseudo action: ms-sf_start_0 * Pseudo action: group:0_start_0 * Resource action: stateful-2:0 start on node-b * Pseudo action: group:1_start_0 * Resource action: stateful-2:1 start on node-a * Pseudo action: group:0_running_0 * Pseudo action: group:1_running_0 * Pseudo action: ms-sf_running_0 * Pseudo action: ms-sf_promote_0 * Pseudo action: group:1_promote_0 * Resource action: stateful-2:1 promote on node-a * Pseudo action: group:1_promoted_0 * Pseudo action: ms-sf_promoted_0 Revised Cluster Status: * Node List: * Online: [ node-a node-b ] * Full List of Resources: * Clone Set: ms-sf [group] (promotable) (unique): * Resource Group: group:0: * stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b * stateful-2:0 (ocf:heartbeat:Stateful): Unpromoted node-b * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a * stateful-2:1 (ocf:heartbeat:Stateful): Promoted node-a diff --git a/cts/scheduler/summary/bundle-order-fencing.summary b/cts/scheduler/summary/bundle-order-fencing.summary index 387c05532a..8cb40718db 100644 --- a/cts/scheduler/summary/bundle-order-fencing.summary +++ b/cts/scheduler/summary/bundle-order-fencing.summary @@ -1,220 +1,220 @@ Using the original execution date of: 2017-09-12 10:51:59Z Current cluster status: * Node List: * Node controller-0: UNCLEAN (offline) * Online: [ controller-1 controller-2 ] * GuestOnline: [ galera-bundle-1@controller-1 galera-bundle-2@controller-2 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ] * Full List of Resources: * Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]: * rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): FAILED controller-0 (UNCLEAN) * rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1 * rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2 * Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]: * galera-bundle-0 (ocf:heartbeat:galera): FAILED Promoted controller-0 (UNCLEAN) * galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1 * galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2 * Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]: * redis-bundle-0 (ocf:heartbeat:redis): FAILED Promoted controller-0 (UNCLEAN) * redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1 * redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2 * ip-192.168.24.7 (ocf:heartbeat:IPaddr2): Started controller-0 (UNCLEAN) * ip-10.0.0.109 (ocf:heartbeat:IPaddr2): Started controller-0 (UNCLEAN) * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.3.19 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-0 (UNCLEAN) * Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]: * haproxy-bundle-docker-0 (ocf:heartbeat:docker): Started controller-0 (UNCLEAN) * haproxy-bundle-docker-1 (ocf:heartbeat:docker): Started controller-2 * haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-1 * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-2 * stonith-fence_ipmilan-525400efba5c (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-5254003e8e97 (stonith:fence_ipmilan): Started controller-0 (UNCLEAN) * stonith-fence_ipmilan-5254000dcb3f (stonith:fence_ipmilan): Started controller-0 (UNCLEAN) Transition Summary: * Fence (off) redis-bundle-0 (resource: redis-bundle-docker-0) 'guest is unclean' * Fence (off) rabbitmq-bundle-0 (resource: rabbitmq-bundle-docker-0) 'guest is unclean' * Fence (off) galera-bundle-0 (resource: galera-bundle-docker-0) 'guest is unclean' * Fence (reboot) controller-0 'peer is no longer part of the cluster' * Stop rabbitmq-bundle-docker-0 ( controller-0 ) due to node availability * Stop rabbitmq-bundle-0 ( controller-0 ) due to unrunnable rabbitmq-bundle-docker-0 start * Stop rabbitmq:0 ( rabbitmq-bundle-0 ) due to unrunnable rabbitmq-bundle-docker-0 start * Stop galera-bundle-docker-0 ( controller-0 ) due to node availability * Stop galera-bundle-0 ( controller-0 ) due to unrunnable galera-bundle-docker-0 start - * Stop galera:0 ( Promoted galera-bundle-0 ) due to unrunnable galera-bundle-docker-0 start + * Stop galera:0 ( Promoted galera-bundle-0 ) due to unrunnable galera-bundle-docker-0 start * Stop redis-bundle-docker-0 ( controller-0 ) due to node availability * Stop redis-bundle-0 ( controller-0 ) due to unrunnable redis-bundle-docker-0 start - * Stop redis:0 ( Promoted redis-bundle-0 ) due to unrunnable redis-bundle-docker-0 start + * Stop redis:0 ( Promoted redis-bundle-0 ) due to unrunnable redis-bundle-docker-0 start * Promote redis:1 ( Unpromoted -> Promoted redis-bundle-1 ) * Move ip-192.168.24.7 ( controller-0 -> controller-2 ) * Move ip-10.0.0.109 ( controller-0 -> controller-1 ) * Move ip-172.17.4.11 ( controller-0 -> controller-1 ) * Stop haproxy-bundle-docker-0 ( controller-0 ) due to node availability * Move stonith-fence_ipmilan-5254003e8e97 ( controller-0 -> controller-1 ) * Move stonith-fence_ipmilan-5254000dcb3f ( controller-0 -> controller-2 ) Executing Cluster Transition: * Pseudo action: rabbitmq-bundle-clone_pre_notify_stop_0 * Pseudo action: rabbitmq-bundle-0_stop_0 * Resource action: rabbitmq-bundle-0 monitor on controller-2 * Resource action: rabbitmq-bundle-0 monitor on controller-1 * Resource action: rabbitmq-bundle-1 monitor on controller-2 * Resource action: rabbitmq-bundle-2 monitor on controller-1 * Pseudo action: galera-bundle-0_stop_0 * Resource action: galera-bundle-0 monitor on controller-2 * Resource action: galera-bundle-0 monitor on controller-1 * Resource action: galera-bundle-1 monitor on controller-2 * Resource action: galera-bundle-2 monitor on controller-1 * Resource action: redis cancel=45000 on redis-bundle-1 * Resource action: redis cancel=60000 on redis-bundle-1 * Pseudo action: redis-bundle-master_pre_notify_demote_0 * Pseudo action: redis-bundle-0_stop_0 * Resource action: redis-bundle-0 monitor on controller-2 * Resource action: redis-bundle-0 monitor on controller-1 * Resource action: redis-bundle-1 monitor on controller-2 * Resource action: redis-bundle-2 monitor on controller-1 * Pseudo action: stonith-fence_ipmilan-5254003e8e97_stop_0 * Pseudo action: stonith-fence_ipmilan-5254000dcb3f_stop_0 * Pseudo action: haproxy-bundle_stop_0 * Pseudo action: redis-bundle_demote_0 * Pseudo action: galera-bundle_demote_0 * Pseudo action: rabbitmq-bundle_stop_0 * Pseudo action: rabbitmq-bundle_start_0 * Fencing controller-0 (reboot) * Resource action: rabbitmq notify on rabbitmq-bundle-1 * Resource action: rabbitmq notify on rabbitmq-bundle-2 * Pseudo action: rabbitmq-bundle-clone_confirmed-pre_notify_stop_0 * Pseudo action: rabbitmq-bundle-docker-0_stop_0 * Pseudo action: galera-bundle-master_demote_0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-pre_notify_demote_0 * Pseudo action: redis-bundle-master_demote_0 * Pseudo action: haproxy-bundle-docker-0_stop_0 * Resource action: stonith-fence_ipmilan-5254003e8e97 start on controller-1 * Resource action: stonith-fence_ipmilan-5254000dcb3f start on controller-2 * Pseudo action: stonith-redis-bundle-0-off on redis-bundle-0 * Pseudo action: stonith-rabbitmq-bundle-0-off on rabbitmq-bundle-0 * Pseudo action: stonith-galera-bundle-0-off on galera-bundle-0 * Pseudo action: haproxy-bundle_stopped_0 * Pseudo action: rabbitmq_post_notify_stop_0 * Pseudo action: rabbitmq-bundle-clone_stop_0 * Pseudo action: galera_demote_0 * Pseudo action: galera-bundle-master_demoted_0 * Pseudo action: redis_post_notify_stop_0 * Pseudo action: redis_demote_0 * Pseudo action: redis-bundle-master_demoted_0 * Pseudo action: ip-192.168.24.7_stop_0 * Pseudo action: ip-10.0.0.109_stop_0 * Pseudo action: ip-172.17.4.11_stop_0 * Resource action: stonith-fence_ipmilan-5254003e8e97 monitor=60000 on controller-1 * Resource action: stonith-fence_ipmilan-5254000dcb3f monitor=60000 on controller-2 * Pseudo action: galera-bundle_demoted_0 * Pseudo action: galera-bundle_stop_0 * Pseudo action: rabbitmq_stop_0 * Pseudo action: rabbitmq-bundle-clone_stopped_0 * Pseudo action: galera-bundle-master_stop_0 * Pseudo action: galera-bundle-docker-0_stop_0 * Pseudo action: redis-bundle-master_post_notify_demoted_0 * Resource action: ip-192.168.24.7 start on controller-2 * Resource action: ip-10.0.0.109 start on controller-1 * Resource action: ip-172.17.4.11 start on controller-1 * Pseudo action: rabbitmq-bundle-clone_post_notify_stopped_0 * Pseudo action: galera_stop_0 * Pseudo action: galera-bundle-master_stopped_0 * Pseudo action: galera-bundle-master_start_0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-post_notify_demoted_0 * Pseudo action: redis-bundle-master_pre_notify_stop_0 * Resource action: ip-192.168.24.7 monitor=10000 on controller-2 * Resource action: ip-10.0.0.109 monitor=10000 on controller-1 * Resource action: ip-172.17.4.11 monitor=10000 on controller-1 * Pseudo action: redis-bundle_demoted_0 * Pseudo action: redis-bundle_stop_0 * Pseudo action: galera-bundle_stopped_0 * Resource action: rabbitmq notify on rabbitmq-bundle-1 * Resource action: rabbitmq notify on rabbitmq-bundle-2 * Pseudo action: rabbitmq-bundle-clone_confirmed-post_notify_stopped_0 * Pseudo action: rabbitmq-bundle-clone_pre_notify_start_0 * Pseudo action: galera-bundle-master_running_0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-pre_notify_stop_0 * Pseudo action: redis-bundle-master_stop_0 * Pseudo action: redis-bundle-docker-0_stop_0 * Pseudo action: galera-bundle_running_0 * Pseudo action: rabbitmq-bundle_stopped_0 * Pseudo action: rabbitmq_notified_0 * Pseudo action: rabbitmq-bundle-clone_confirmed-pre_notify_start_0 * Pseudo action: rabbitmq-bundle-clone_start_0 * Pseudo action: redis_stop_0 * Pseudo action: redis-bundle-master_stopped_0 * Pseudo action: rabbitmq-bundle-clone_running_0 * Pseudo action: redis-bundle-master_post_notify_stopped_0 * Pseudo action: rabbitmq-bundle-clone_post_notify_running_0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-post_notify_stopped_0 * Pseudo action: redis-bundle-master_pre_notify_start_0 * Pseudo action: redis-bundle_stopped_0 * Pseudo action: rabbitmq-bundle-clone_confirmed-post_notify_running_0 * Pseudo action: redis_notified_0 * Pseudo action: redis-bundle-master_confirmed-pre_notify_start_0 * Pseudo action: redis-bundle-master_start_0 * Pseudo action: rabbitmq-bundle_running_0 * Pseudo action: redis-bundle-master_running_0 * Pseudo action: redis-bundle-master_post_notify_running_0 * Pseudo action: redis-bundle-master_confirmed-post_notify_running_0 * Pseudo action: redis-bundle_running_0 * Pseudo action: redis-bundle-master_pre_notify_promote_0 * Pseudo action: redis-bundle_promote_0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-pre_notify_promote_0 * Pseudo action: redis-bundle-master_promote_0 * Resource action: redis promote on redis-bundle-1 * Pseudo action: redis-bundle-master_promoted_0 * Pseudo action: redis-bundle-master_post_notify_promoted_0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-post_notify_promoted_0 * Pseudo action: redis-bundle_promoted_0 * Resource action: redis monitor=20000 on redis-bundle-1 Using the original execution date of: 2017-09-12 10:51:59Z Revised Cluster Status: * Node List: * Online: [ controller-1 controller-2 ] * OFFLINE: [ controller-0 ] * GuestOnline: [ galera-bundle-1@controller-1 galera-bundle-2@controller-2 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ] * Full List of Resources: * Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]: * rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): FAILED * rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1 * rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2 * Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]: * galera-bundle-0 (ocf:heartbeat:galera): FAILED Promoted * galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1 * galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2 * Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]: * redis-bundle-0 (ocf:heartbeat:redis): FAILED Promoted * redis-bundle-1 (ocf:heartbeat:redis): Promoted controller-1 * redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2 * ip-192.168.24.7 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-10.0.0.109 (ocf:heartbeat:IPaddr2): Started controller-1 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.3.19 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 * Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]: * haproxy-bundle-docker-0 (ocf:heartbeat:docker): Stopped * haproxy-bundle-docker-1 (ocf:heartbeat:docker): Started controller-2 * haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-1 * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-2 * stonith-fence_ipmilan-525400efba5c (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-5254003e8e97 (stonith:fence_ipmilan): Started controller-1 * stonith-fence_ipmilan-5254000dcb3f (stonith:fence_ipmilan): Started controller-2 diff --git a/cts/scheduler/summary/bundle-order-stop-on-remote.summary b/cts/scheduler/summary/bundle-order-stop-on-remote.summary index bf94ce3c72..8cd17eef61 100644 --- a/cts/scheduler/summary/bundle-order-stop-on-remote.summary +++ b/cts/scheduler/summary/bundle-order-stop-on-remote.summary @@ -1,224 +1,224 @@ Current cluster status: * Node List: * RemoteNode database-0: UNCLEAN (offline) * RemoteNode database-2: UNCLEAN (offline) * Online: [ controller-0 controller-1 controller-2 ] * RemoteOnline: [ database-1 messaging-0 messaging-1 messaging-2 ] * GuestOnline: [ galera-bundle-1@controller-2 rabbitmq-bundle-0@controller-2 rabbitmq-bundle-1@controller-2 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-2@controller-2 ] * Full List of Resources: * database-0 (ocf:pacemaker:remote): Stopped * database-1 (ocf:pacemaker:remote): Started controller-2 * database-2 (ocf:pacemaker:remote): Stopped * messaging-0 (ocf:pacemaker:remote): Started controller-2 * messaging-1 (ocf:pacemaker:remote): Started controller-2 * messaging-2 (ocf:pacemaker:remote): Started controller-2 * Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]: * rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started messaging-0 * rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started messaging-1 * rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started messaging-2 * Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]: * galera-bundle-0 (ocf:heartbeat:galera): FAILED Promoted database-0 (UNCLEAN) * galera-bundle-1 (ocf:heartbeat:galera): Promoted database-1 * galera-bundle-2 (ocf:heartbeat:galera): FAILED Promoted database-2 (UNCLEAN) * Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]: * redis-bundle-0 (ocf:heartbeat:redis): Unpromoted controller-0 * redis-bundle-1 (ocf:heartbeat:redis): Stopped * redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2 * ip-192.168.24.11 (ocf:heartbeat:IPaddr2): Stopped * ip-10.0.0.104 (ocf:heartbeat:IPaddr2): Stopped * ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.11 (ocf:heartbeat:IPaddr2): Stopped * ip-172.17.3.13 (ocf:heartbeat:IPaddr2): Stopped * ip-172.17.4.19 (ocf:heartbeat:IPaddr2): Started controller-2 * Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]: * haproxy-bundle-docker-0 (ocf:heartbeat:docker): Started controller-0 * haproxy-bundle-docker-1 (ocf:heartbeat:docker): Stopped * haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-2 * openstack-cinder-volume (systemd:openstack-cinder-volume): Stopped * stonith-fence_ipmilan-525400244e09 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400cdec10 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400c709f7 (stonith:fence_ipmilan): Stopped * stonith-fence_ipmilan-525400a7f9e0 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400a25787 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-5254005ea387 (stonith:fence_ipmilan): Stopped * stonith-fence_ipmilan-525400542c06 (stonith:fence_ipmilan): Stopped * stonith-fence_ipmilan-525400aac413 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400498d34 (stonith:fence_ipmilan): Stopped Transition Summary: * Fence (reboot) galera-bundle-2 (resource: galera-bundle-docker-2) 'guest is unclean' * Fence (reboot) galera-bundle-0 (resource: galera-bundle-docker-0) 'guest is unclean' * Start database-0 ( controller-0 ) * Start database-2 ( controller-1 ) * Recover galera-bundle-docker-0 ( database-0 ) * Start galera-bundle-0 ( controller-0 ) - * Recover galera:0 ( Promoted galera-bundle-0 ) + * Recover galera:0 ( Promoted galera-bundle-0 ) * Recover galera-bundle-docker-2 ( database-2 ) * Start galera-bundle-2 ( controller-1 ) - * Recover galera:2 ( Promoted galera-bundle-2 ) + * Recover galera:2 ( Promoted galera-bundle-2 ) * Promote redis:0 ( Unpromoted -> Promoted redis-bundle-0 ) * Start redis-bundle-docker-1 ( controller-1 ) * Start redis-bundle-1 ( controller-1 ) * Start redis:1 ( redis-bundle-1 ) * Start ip-192.168.24.11 ( controller-0 ) * Start ip-10.0.0.104 ( controller-1 ) * Start ip-172.17.1.11 ( controller-0 ) * Start ip-172.17.3.13 ( controller-1 ) * Start haproxy-bundle-docker-1 ( controller-1 ) * Start openstack-cinder-volume ( controller-0 ) * Start stonith-fence_ipmilan-525400c709f7 ( controller-1 ) * Start stonith-fence_ipmilan-5254005ea387 ( controller-1 ) * Start stonith-fence_ipmilan-525400542c06 ( controller-0 ) * Start stonith-fence_ipmilan-525400498d34 ( controller-1 ) Executing Cluster Transition: * Resource action: database-0 start on controller-0 * Resource action: database-2 start on controller-1 * Pseudo action: redis-bundle-master_pre_notify_start_0 * Resource action: stonith-fence_ipmilan-525400c709f7 start on controller-1 * Resource action: stonith-fence_ipmilan-5254005ea387 start on controller-1 * Resource action: stonith-fence_ipmilan-525400542c06 start on controller-0 * Resource action: stonith-fence_ipmilan-525400498d34 start on controller-1 * Pseudo action: redis-bundle_start_0 * Pseudo action: galera-bundle_demote_0 * Resource action: database-0 monitor=20000 on controller-0 * Resource action: database-2 monitor=20000 on controller-1 * Pseudo action: galera-bundle-master_demote_0 * Resource action: redis notify on redis-bundle-0 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-pre_notify_start_0 * Pseudo action: redis-bundle-master_start_0 * Resource action: stonith-fence_ipmilan-525400c709f7 monitor=60000 on controller-1 * Resource action: stonith-fence_ipmilan-5254005ea387 monitor=60000 on controller-1 * Resource action: stonith-fence_ipmilan-525400542c06 monitor=60000 on controller-0 * Resource action: stonith-fence_ipmilan-525400498d34 monitor=60000 on controller-1 * Pseudo action: galera_demote_0 * Pseudo action: galera_demote_0 * Pseudo action: galera-bundle-master_demoted_0 * Pseudo action: galera-bundle_demoted_0 * Pseudo action: galera-bundle_stop_0 * Resource action: galera-bundle-docker-0 stop on database-0 * Resource action: galera-bundle-docker-2 stop on database-2 * Pseudo action: stonith-galera-bundle-2-reboot on galera-bundle-2 * Pseudo action: stonith-galera-bundle-0-reboot on galera-bundle-0 * Pseudo action: galera-bundle-master_stop_0 * Resource action: redis-bundle-docker-1 start on controller-1 * Resource action: redis-bundle-1 monitor on controller-1 * Resource action: ip-192.168.24.11 start on controller-0 * Resource action: ip-10.0.0.104 start on controller-1 * Resource action: ip-172.17.1.11 start on controller-0 * Resource action: ip-172.17.3.13 start on controller-1 * Resource action: openstack-cinder-volume start on controller-0 * Pseudo action: haproxy-bundle_start_0 * Pseudo action: galera_stop_0 * Resource action: redis-bundle-docker-1 monitor=60000 on controller-1 * Resource action: redis-bundle-1 start on controller-1 * Resource action: ip-192.168.24.11 monitor=10000 on controller-0 * Resource action: ip-10.0.0.104 monitor=10000 on controller-1 * Resource action: ip-172.17.1.11 monitor=10000 on controller-0 * Resource action: ip-172.17.3.13 monitor=10000 on controller-1 * Resource action: haproxy-bundle-docker-1 start on controller-1 * Resource action: openstack-cinder-volume monitor=60000 on controller-0 * Pseudo action: haproxy-bundle_running_0 * Pseudo action: galera_stop_0 * Pseudo action: galera-bundle-master_stopped_0 * Resource action: redis start on redis-bundle-1 * Pseudo action: redis-bundle-master_running_0 * Resource action: redis-bundle-1 monitor=30000 on controller-1 * Resource action: haproxy-bundle-docker-1 monitor=60000 on controller-1 * Pseudo action: galera-bundle_stopped_0 * Pseudo action: galera-bundle_start_0 * Pseudo action: galera-bundle-master_start_0 * Resource action: galera-bundle-docker-0 start on database-0 * Resource action: galera-bundle-0 monitor on controller-1 * Resource action: galera-bundle-docker-2 start on database-2 * Resource action: galera-bundle-2 monitor on controller-1 * Pseudo action: redis-bundle-master_post_notify_running_0 * Resource action: galera-bundle-docker-0 monitor=60000 on database-0 * Resource action: galera-bundle-0 start on controller-0 * Resource action: galera-bundle-docker-2 monitor=60000 on database-2 * Resource action: galera-bundle-2 start on controller-1 * Resource action: redis notify on redis-bundle-0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-post_notify_running_0 * Pseudo action: redis-bundle_running_0 * Resource action: galera start on galera-bundle-0 * Resource action: galera start on galera-bundle-2 * Pseudo action: galera-bundle-master_running_0 * Resource action: galera-bundle-0 monitor=30000 on controller-0 * Resource action: galera-bundle-2 monitor=30000 on controller-1 * Pseudo action: redis-bundle-master_pre_notify_promote_0 * Pseudo action: redis-bundle_promote_0 * Pseudo action: galera-bundle_running_0 * Resource action: redis notify on redis-bundle-0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-pre_notify_promote_0 * Pseudo action: redis-bundle-master_promote_0 * Pseudo action: galera-bundle_promote_0 * Pseudo action: galera-bundle-master_promote_0 * Resource action: redis promote on redis-bundle-0 * Pseudo action: redis-bundle-master_promoted_0 * Resource action: galera promote on galera-bundle-0 * Resource action: galera promote on galera-bundle-2 * Pseudo action: galera-bundle-master_promoted_0 * Pseudo action: redis-bundle-master_post_notify_promoted_0 * Pseudo action: galera-bundle_promoted_0 * Resource action: galera monitor=10000 on galera-bundle-0 * Resource action: galera monitor=10000 on galera-bundle-2 * Resource action: redis notify on redis-bundle-0 * Resource action: redis notify on redis-bundle-1 * Resource action: redis notify on redis-bundle-2 * Pseudo action: redis-bundle-master_confirmed-post_notify_promoted_0 * Pseudo action: redis-bundle_promoted_0 * Resource action: redis monitor=20000 on redis-bundle-0 * Resource action: redis monitor=60000 on redis-bundle-1 * Resource action: redis monitor=45000 on redis-bundle-1 Revised Cluster Status: * Node List: * Online: [ controller-0 controller-1 controller-2 ] * RemoteOnline: [ database-0 database-1 database-2 messaging-0 messaging-1 messaging-2 ] * GuestOnline: [ galera-bundle-0@controller-0 galera-bundle-1@controller-2 galera-bundle-2@controller-1 rabbitmq-bundle-0@controller-2 rabbitmq-bundle-1@controller-2 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ] * Full List of Resources: * database-0 (ocf:pacemaker:remote): Started controller-0 * database-1 (ocf:pacemaker:remote): Started controller-2 * database-2 (ocf:pacemaker:remote): Started controller-1 * messaging-0 (ocf:pacemaker:remote): Started controller-2 * messaging-1 (ocf:pacemaker:remote): Started controller-2 * messaging-2 (ocf:pacemaker:remote): Started controller-2 * Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]: * rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started messaging-0 * rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started messaging-1 * rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started messaging-2 * Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]: * galera-bundle-0 (ocf:heartbeat:galera): Promoted database-0 * galera-bundle-1 (ocf:heartbeat:galera): Promoted database-1 * galera-bundle-2 (ocf:heartbeat:galera): Promoted database-2 * Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]: * redis-bundle-0 (ocf:heartbeat:redis): Promoted controller-0 * redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1 * redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2 * ip-192.168.24.11 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.104 (ocf:heartbeat:IPaddr2): Started controller-1 * ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.11 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.3.13 (ocf:heartbeat:IPaddr2): Started controller-1 * ip-172.17.4.19 (ocf:heartbeat:IPaddr2): Started controller-2 * Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]: * haproxy-bundle-docker-0 (ocf:heartbeat:docker): Started controller-0 * haproxy-bundle-docker-1 (ocf:heartbeat:docker): Started controller-1 * haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-2 * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400244e09 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400cdec10 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400c709f7 (stonith:fence_ipmilan): Started controller-1 * stonith-fence_ipmilan-525400a7f9e0 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400a25787 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-5254005ea387 (stonith:fence_ipmilan): Started controller-1 * stonith-fence_ipmilan-525400542c06 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400aac413 (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400498d34 (stonith:fence_ipmilan): Started controller-1 diff --git a/cts/scheduler/summary/colocation-influence.summary b/cts/scheduler/summary/colocation-influence.summary index 3ea8b3f545..7fa4fcf0c2 100644 --- a/cts/scheduler/summary/colocation-influence.summary +++ b/cts/scheduler/summary/colocation-influence.summary @@ -1,170 +1,170 @@ Current cluster status: * Node List: * Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * GuestOnline: [ bundle10-0@rhel7-2 bundle10-1@rhel7-3 bundle11-0@rhel7-1 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-1 * rsc1a (ocf:pacemaker:Dummy): Started rhel7-2 * rsc1b (ocf:pacemaker:Dummy): Started rhel7-2 * rsc2a (ocf:pacemaker:Dummy): Started rhel7-4 * rsc2b (ocf:pacemaker:Dummy): Started rhel7-4 * rsc3a (ocf:pacemaker:Dummy): Stopped * rsc3b (ocf:pacemaker:Dummy): Stopped * rsc4a (ocf:pacemaker:Dummy): Started rhel7-3 * rsc4b (ocf:pacemaker:Dummy): Started rhel7-3 * rsc5a (ocf:pacemaker:Dummy): Started rhel7-1 * Resource Group: group5a: * rsc5a1 (ocf:pacemaker:Dummy): Started rhel7-1 * rsc5a2 (ocf:pacemaker:Dummy): Started rhel7-1 * Resource Group: group6a: * rsc6a1 (ocf:pacemaker:Dummy): Started rhel7-2 * rsc6a2 (ocf:pacemaker:Dummy): Started rhel7-2 * rsc6a (ocf:pacemaker:Dummy): Started rhel7-2 * Resource Group: group7a: * rsc7a1 (ocf:pacemaker:Dummy): Started rhel7-3 * rsc7a2 (ocf:pacemaker:Dummy): Started rhel7-3 * Clone Set: rsc8a-clone [rsc8a]: * Started: [ rhel7-1 rhel7-3 rhel7-4 ] * Clone Set: rsc8b-clone [rsc8b]: * Started: [ rhel7-1 rhel7-3 rhel7-4 ] * rsc9a (ocf:pacemaker:Dummy): Started rhel7-4 * rsc9b (ocf:pacemaker:Dummy): Started rhel7-4 * rsc9c (ocf:pacemaker:Dummy): Started rhel7-4 * rsc10a (ocf:pacemaker:Dummy): Started rhel7-2 * rsc11a (ocf:pacemaker:Dummy): Started rhel7-1 * rsc12a (ocf:pacemaker:Dummy): Started rhel7-1 * rsc12b (ocf:pacemaker:Dummy): Started rhel7-1 * rsc12c (ocf:pacemaker:Dummy): Started rhel7-1 * Container bundle set: bundle10 [pcmktest:http]: * bundle10-0 (192.168.122.131) (ocf:heartbeat:apache): Started rhel7-2 * bundle10-1 (192.168.122.132) (ocf:heartbeat:apache): Started rhel7-3 * Container bundle set: bundle11 [pcmktest:http]: * bundle11-0 (192.168.122.134) (ocf:pacemaker:Dummy): Started rhel7-1 * bundle11-1 (192.168.122.135) (ocf:pacemaker:Dummy): Stopped * rsc13a (ocf:pacemaker:Dummy): Started rhel7-3 * Clone Set: rsc13b-clone [rsc13b] (promotable): * Promoted: [ rhel7-3 ] * Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 ] * Stopped: [ rhel7-5 ] * rsc14b (ocf:pacemaker:Dummy): Started rhel7-4 * Clone Set: rsc14a-clone [rsc14a] (promotable): * Promoted: [ rhel7-4 ] * Unpromoted: [ rhel7-1 rhel7-2 rhel7-3 ] * Stopped: [ rhel7-5 ] Transition Summary: * Move rsc1a ( rhel7-2 -> rhel7-3 ) * Move rsc1b ( rhel7-2 -> rhel7-3 ) * Stop rsc2a ( rhel7-4 ) due to node availability * Start rsc3a ( rhel7-2 ) * Start rsc3b ( rhel7-2 ) * Stop rsc4a ( rhel7-3 ) due to node availability * Stop rsc5a ( rhel7-1 ) due to node availability * Stop rsc6a1 ( rhel7-2 ) due to node availability * Stop rsc6a2 ( rhel7-2 ) due to node availability * Stop rsc7a2 ( rhel7-3 ) due to node availability * Stop rsc8a:1 ( rhel7-4 ) due to node availability * Stop rsc9c ( rhel7-4 ) due to node availability * Move rsc10a ( rhel7-2 -> rhel7-3 ) * Stop rsc12b ( rhel7-1 ) due to node availability * Start bundle11-1 ( rhel7-5 ) due to unrunnable bundle11-docker-1 start (blocked) * Start bundle11a:1 ( bundle11-1 ) due to unrunnable bundle11-docker-1 start (blocked) * Stop rsc13a ( rhel7-3 ) due to node availability - * Stop rsc14a:1 ( Promoted rhel7-4 ) due to node availability + * Stop rsc14a:1 ( Promoted rhel7-4 ) due to node availability Executing Cluster Transition: * Resource action: rsc1a stop on rhel7-2 * Resource action: rsc1b stop on rhel7-2 * Resource action: rsc2a stop on rhel7-4 * Resource action: rsc3a start on rhel7-2 * Resource action: rsc3b start on rhel7-2 * Resource action: rsc4a stop on rhel7-3 * Resource action: rsc5a stop on rhel7-1 * Pseudo action: group6a_stop_0 * Resource action: rsc6a2 stop on rhel7-2 * Pseudo action: group7a_stop_0 * Resource action: rsc7a2 stop on rhel7-3 * Pseudo action: rsc8a-clone_stop_0 * Resource action: rsc9c stop on rhel7-4 * Resource action: rsc10a stop on rhel7-2 * Resource action: rsc12b stop on rhel7-1 * Resource action: rsc13a stop on rhel7-3 * Pseudo action: rsc14a-clone_demote_0 * Pseudo action: bundle11_start_0 * Resource action: rsc1a start on rhel7-3 * Resource action: rsc1b start on rhel7-3 * Resource action: rsc3a monitor=10000 on rhel7-2 * Resource action: rsc3b monitor=10000 on rhel7-2 * Resource action: rsc6a1 stop on rhel7-2 * Pseudo action: group7a_stopped_0 * Resource action: rsc8a stop on rhel7-4 * Pseudo action: rsc8a-clone_stopped_0 * Resource action: rsc10a start on rhel7-3 * Pseudo action: bundle11-clone_start_0 * Resource action: rsc14a demote on rhel7-4 * Pseudo action: rsc14a-clone_demoted_0 * Pseudo action: rsc14a-clone_stop_0 * Resource action: rsc1a monitor=10000 on rhel7-3 * Resource action: rsc1b monitor=10000 on rhel7-3 * Pseudo action: group6a_stopped_0 * Resource action: rsc10a monitor=10000 on rhel7-3 * Pseudo action: bundle11-clone_running_0 * Resource action: rsc14a stop on rhel7-4 * Pseudo action: rsc14a-clone_stopped_0 * Pseudo action: bundle11_running_0 Revised Cluster Status: * Node List: * Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * GuestOnline: [ bundle10-0@rhel7-2 bundle10-1@rhel7-3 bundle11-0@rhel7-1 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-1 * rsc1a (ocf:pacemaker:Dummy): Started rhel7-3 * rsc1b (ocf:pacemaker:Dummy): Started rhel7-3 * rsc2a (ocf:pacemaker:Dummy): Stopped * rsc2b (ocf:pacemaker:Dummy): Started rhel7-4 * rsc3a (ocf:pacemaker:Dummy): Started rhel7-2 * rsc3b (ocf:pacemaker:Dummy): Started rhel7-2 * rsc4a (ocf:pacemaker:Dummy): Stopped * rsc4b (ocf:pacemaker:Dummy): Started rhel7-3 * rsc5a (ocf:pacemaker:Dummy): Stopped * Resource Group: group5a: * rsc5a1 (ocf:pacemaker:Dummy): Started rhel7-1 * rsc5a2 (ocf:pacemaker:Dummy): Started rhel7-1 * Resource Group: group6a: * rsc6a1 (ocf:pacemaker:Dummy): Stopped * rsc6a2 (ocf:pacemaker:Dummy): Stopped * rsc6a (ocf:pacemaker:Dummy): Started rhel7-2 * Resource Group: group7a: * rsc7a1 (ocf:pacemaker:Dummy): Started rhel7-3 * rsc7a2 (ocf:pacemaker:Dummy): Stopped * Clone Set: rsc8a-clone [rsc8a]: * Started: [ rhel7-1 rhel7-3 ] * Stopped: [ rhel7-2 rhel7-4 rhel7-5 ] * Clone Set: rsc8b-clone [rsc8b]: * Started: [ rhel7-1 rhel7-3 rhel7-4 ] * rsc9a (ocf:pacemaker:Dummy): Started rhel7-4 * rsc9b (ocf:pacemaker:Dummy): Started rhel7-4 * rsc9c (ocf:pacemaker:Dummy): Stopped * rsc10a (ocf:pacemaker:Dummy): Started rhel7-3 * rsc11a (ocf:pacemaker:Dummy): Started rhel7-1 * rsc12a (ocf:pacemaker:Dummy): Started rhel7-1 * rsc12b (ocf:pacemaker:Dummy): Stopped * rsc12c (ocf:pacemaker:Dummy): Started rhel7-1 * Container bundle set: bundle10 [pcmktest:http]: * bundle10-0 (192.168.122.131) (ocf:heartbeat:apache): Started rhel7-2 * bundle10-1 (192.168.122.132) (ocf:heartbeat:apache): Started rhel7-3 * Container bundle set: bundle11 [pcmktest:http]: * bundle11-0 (192.168.122.134) (ocf:pacemaker:Dummy): Started rhel7-1 * bundle11-1 (192.168.122.135) (ocf:pacemaker:Dummy): Stopped * rsc13a (ocf:pacemaker:Dummy): Stopped * Clone Set: rsc13b-clone [rsc13b] (promotable): * Promoted: [ rhel7-3 ] * Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 ] * Stopped: [ rhel7-5 ] * rsc14b (ocf:pacemaker:Dummy): Started rhel7-4 * Clone Set: rsc14a-clone [rsc14a] (promotable): * Unpromoted: [ rhel7-1 rhel7-2 rhel7-3 ] * Stopped: [ rhel7-4 rhel7-5 ] diff --git a/cts/scheduler/summary/dc-fence-ordering.summary b/cts/scheduler/summary/dc-fence-ordering.summary index ac46031f07..305ebd5c19 100644 --- a/cts/scheduler/summary/dc-fence-ordering.summary +++ b/cts/scheduler/summary/dc-fence-ordering.summary @@ -1,82 +1,82 @@ Using the original execution date of: 2018-11-28 18:37:16Z Current cluster status: * Node List: * Node rhel7-1: UNCLEAN (online) * Online: [ rhel7-2 rhel7-4 rhel7-5 ] * OFFLINE: [ rhel7-3 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Stopped * FencingPass (stonith:fence_dummy): Stopped * FencingFail (stonith:fence_dummy): Stopped * rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Stopped * migrator (ocf:pacemaker:Dummy): Stopped * Clone Set: Connectivity [ping-1]: * Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * Clone Set: promotable-1 [stateful-1] (promotable): * Promoted: [ rhel7-1 ] * Unpromoted: [ rhel7-2 rhel7-4 rhel7-5 ] * Stopped: [ rhel7-3 ] * Resource Group: group-1: * r192.168.122.207 (ocf:heartbeat:IPaddr2): Started rhel7-1 * petulant (service:pacemaker-cts-dummyd@10): FAILED rhel7-1 * r192.168.122.208 (ocf:heartbeat:IPaddr2): Stopped * lsb-dummy (lsb:LSBDummy): Stopped Transition Summary: * Fence (reboot) rhel7-1 'petulant failed there' - * Stop stateful-1:0 ( Unpromoted rhel7-5 ) due to node availability - * Stop stateful-1:1 ( Promoted rhel7-1 ) due to node availability - * Stop stateful-1:2 ( Unpromoted rhel7-2 ) due to node availability - * Stop stateful-1:3 ( Unpromoted rhel7-4 ) due to node availability + * Stop stateful-1:0 ( Unpromoted rhel7-5 ) due to node availability + * Stop stateful-1:1 ( Promoted rhel7-1 ) due to node availability + * Stop stateful-1:2 ( Unpromoted rhel7-2 ) due to node availability + * Stop stateful-1:3 ( Unpromoted rhel7-4 ) due to node availability * Stop r192.168.122.207 ( rhel7-1 ) due to node availability * Stop petulant ( rhel7-1 ) due to node availability Executing Cluster Transition: * Fencing rhel7-1 (reboot) * Pseudo action: group-1_stop_0 * Pseudo action: petulant_stop_0 * Pseudo action: r192.168.122.207_stop_0 * Pseudo action: group-1_stopped_0 * Pseudo action: promotable-1_demote_0 * Pseudo action: stateful-1_demote_0 * Pseudo action: promotable-1_demoted_0 * Pseudo action: promotable-1_stop_0 * Resource action: stateful-1 stop on rhel7-5 * Pseudo action: stateful-1_stop_0 * Resource action: stateful-1 stop on rhel7-2 * Resource action: stateful-1 stop on rhel7-4 * Pseudo action: promotable-1_stopped_0 * Cluster action: do_shutdown on rhel7-5 * Cluster action: do_shutdown on rhel7-4 * Cluster action: do_shutdown on rhel7-2 Using the original execution date of: 2018-11-28 18:37:16Z Revised Cluster Status: * Node List: * Online: [ rhel7-2 rhel7-4 rhel7-5 ] * OFFLINE: [ rhel7-1 rhel7-3 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Stopped * FencingPass (stonith:fence_dummy): Stopped * FencingFail (stonith:fence_dummy): Stopped * rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Stopped * rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Stopped * migrator (ocf:pacemaker:Dummy): Stopped * Clone Set: Connectivity [ping-1]: * Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * Clone Set: promotable-1 [stateful-1] (promotable): * Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * Resource Group: group-1: * r192.168.122.207 (ocf:heartbeat:IPaddr2): Stopped * petulant (service:pacemaker-cts-dummyd@10): Stopped * r192.168.122.208 (ocf:heartbeat:IPaddr2): Stopped * lsb-dummy (lsb:LSBDummy): Stopped diff --git a/cts/scheduler/summary/guest-node-host-dies.summary b/cts/scheduler/summary/guest-node-host-dies.summary index b0286b2846..f4509b9029 100644 --- a/cts/scheduler/summary/guest-node-host-dies.summary +++ b/cts/scheduler/summary/guest-node-host-dies.summary @@ -1,82 +1,82 @@ Current cluster status: * Node List: * Node rhel7-1: UNCLEAN (offline) * Online: [ rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-4 * rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-1 (UNCLEAN) * container1 (ocf:heartbeat:VirtualDomain): FAILED rhel7-1 (UNCLEAN) * container2 (ocf:heartbeat:VirtualDomain): FAILED rhel7-1 (UNCLEAN) * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] Transition Summary: * Fence (reboot) lxc2 (resource: container2) 'guest is unclean' * Fence (reboot) lxc1 (resource: container1) 'guest is unclean' * Fence (reboot) rhel7-1 'rsc_rhel7-1 is thought to be active there' * Restart Fencing ( rhel7-4 ) due to resource definition change * Move rsc_rhel7-1 ( rhel7-1 -> rhel7-5 ) * Recover container1 ( rhel7-1 -> rhel7-2 ) * Recover container2 ( rhel7-1 -> rhel7-3 ) - * Recover lxc-ms:0 ( Promoted lxc1 ) - * Recover lxc-ms:1 ( Unpromoted lxc2 ) + * Recover lxc-ms:0 ( Promoted lxc1 ) + * Recover lxc-ms:1 ( Unpromoted lxc2 ) * Move lxc1 ( rhel7-1 -> rhel7-2 ) * Move lxc2 ( rhel7-1 -> rhel7-3 ) Executing Cluster Transition: * Resource action: Fencing stop on rhel7-4 * Pseudo action: lxc-ms-master_demote_0 * Pseudo action: lxc1_stop_0 * Resource action: lxc1 monitor on rhel7-5 * Resource action: lxc1 monitor on rhel7-4 * Resource action: lxc1 monitor on rhel7-3 * Pseudo action: lxc2_stop_0 * Resource action: lxc2 monitor on rhel7-5 * Resource action: lxc2 monitor on rhel7-4 * Resource action: lxc2 monitor on rhel7-2 * Fencing rhel7-1 (reboot) * Pseudo action: rsc_rhel7-1_stop_0 * Pseudo action: container1_stop_0 * Pseudo action: container2_stop_0 * Pseudo action: stonith-lxc2-reboot on lxc2 * Pseudo action: stonith-lxc1-reboot on lxc1 * Resource action: Fencing start on rhel7-4 * Resource action: Fencing monitor=120000 on rhel7-4 * Resource action: rsc_rhel7-1 start on rhel7-5 * Resource action: container1 start on rhel7-2 * Resource action: container2 start on rhel7-3 * Pseudo action: lxc-ms_demote_0 * Pseudo action: lxc-ms-master_demoted_0 * Pseudo action: lxc-ms-master_stop_0 * Resource action: lxc1 start on rhel7-2 * Resource action: lxc2 start on rhel7-3 * Resource action: rsc_rhel7-1 monitor=5000 on rhel7-5 * Pseudo action: lxc-ms_stop_0 * Pseudo action: lxc-ms_stop_0 * Pseudo action: lxc-ms-master_stopped_0 * Pseudo action: lxc-ms-master_start_0 * Resource action: lxc1 monitor=30000 on rhel7-2 * Resource action: lxc2 monitor=30000 on rhel7-3 * Resource action: lxc-ms start on lxc1 * Resource action: lxc-ms start on lxc2 * Pseudo action: lxc-ms-master_running_0 * Resource action: lxc-ms monitor=10000 on lxc2 * Pseudo action: lxc-ms-master_promote_0 * Resource action: lxc-ms promote on lxc1 * Pseudo action: lxc-ms-master_promoted_0 Revised Cluster Status: * Node List: * Online: [ rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * OFFLINE: [ rhel7-1 ] * GuestOnline: [ lxc1@rhel7-2 lxc2@rhel7-3 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-4 * rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-5 * container1 (ocf:heartbeat:VirtualDomain): Started rhel7-2 * container2 (ocf:heartbeat:VirtualDomain): Started rhel7-3 * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Promoted: [ lxc1 ] * Unpromoted: [ lxc2 ] diff --git a/cts/scheduler/summary/migrate-fencing.summary b/cts/scheduler/summary/migrate-fencing.summary index fd4fffa1d3..955bb0f434 100644 --- a/cts/scheduler/summary/migrate-fencing.summary +++ b/cts/scheduler/summary/migrate-fencing.summary @@ -1,108 +1,108 @@ Current cluster status: * Node List: * Node pcmk-4: UNCLEAN (online) * Online: [ pcmk-1 pcmk-2 pcmk-3 ] * Full List of Resources: * Clone Set: Fencing [FencingChild]: * Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Resource Group: group-1: * r192.168.101.181 (ocf:heartbeat:IPaddr): Started pcmk-4 * r192.168.101.182 (ocf:heartbeat:IPaddr): Started pcmk-4 * r192.168.101.183 (ocf:heartbeat:IPaddr): Started pcmk-4 * rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 * rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 * rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 * rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-4 * migrator (ocf:pacemaker:Dummy): Started pcmk-1 * Clone Set: Connectivity [ping-1]: * Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ pcmk-4 ] * Unpromoted: [ pcmk-1 pcmk-2 pcmk-3 ] Transition Summary: * Fence (reboot) pcmk-4 'termination was requested' * Stop FencingChild:0 ( pcmk-4 ) due to node availability * Move r192.168.101.181 ( pcmk-4 -> pcmk-1 ) * Move r192.168.101.182 ( pcmk-4 -> pcmk-1 ) * Move r192.168.101.183 ( pcmk-4 -> pcmk-1 ) * Move rsc_pcmk-4 ( pcmk-4 -> pcmk-2 ) * Move lsb-dummy ( pcmk-4 -> pcmk-1 ) * Migrate migrator ( pcmk-1 -> pcmk-3 ) * Stop ping-1:0 ( pcmk-4 ) due to node availability - * Stop stateful-1:0 ( Promoted pcmk-4 ) due to node availability + * Stop stateful-1:0 ( Promoted pcmk-4 ) due to node availability * Promote stateful-1:1 ( Unpromoted -> Promoted pcmk-1 ) Executing Cluster Transition: * Pseudo action: Fencing_stop_0 * Resource action: stateful-1:3 monitor=15000 on pcmk-3 * Resource action: stateful-1:2 monitor=15000 on pcmk-2 * Fencing pcmk-4 (reboot) * Pseudo action: FencingChild:0_stop_0 * Pseudo action: Fencing_stopped_0 * Pseudo action: rsc_pcmk-4_stop_0 * Pseudo action: lsb-dummy_stop_0 * Resource action: migrator migrate_to on pcmk-1 * Pseudo action: Connectivity_stop_0 * Pseudo action: group-1_stop_0 * Pseudo action: r192.168.101.183_stop_0 * Resource action: rsc_pcmk-4 start on pcmk-2 * Resource action: migrator migrate_from on pcmk-3 * Resource action: migrator stop on pcmk-1 * Pseudo action: ping-1:0_stop_0 * Pseudo action: Connectivity_stopped_0 * Pseudo action: r192.168.101.182_stop_0 * Resource action: rsc_pcmk-4 monitor=5000 on pcmk-2 * Pseudo action: migrator_start_0 * Pseudo action: r192.168.101.181_stop_0 * Resource action: migrator monitor=10000 on pcmk-3 * Pseudo action: group-1_stopped_0 * Pseudo action: master-1_demote_0 * Pseudo action: stateful-1:0_demote_0 * Pseudo action: master-1_demoted_0 * Pseudo action: master-1_stop_0 * Pseudo action: stateful-1:0_stop_0 * Pseudo action: master-1_stopped_0 * Pseudo action: master-1_promote_0 * Resource action: stateful-1:1 promote on pcmk-1 * Pseudo action: master-1_promoted_0 * Pseudo action: group-1_start_0 * Resource action: r192.168.101.181 start on pcmk-1 * Resource action: r192.168.101.182 start on pcmk-1 * Resource action: r192.168.101.183 start on pcmk-1 * Resource action: stateful-1:1 monitor=16000 on pcmk-1 * Pseudo action: group-1_running_0 * Resource action: r192.168.101.181 monitor=5000 on pcmk-1 * Resource action: r192.168.101.182 monitor=5000 on pcmk-1 * Resource action: r192.168.101.183 monitor=5000 on pcmk-1 * Resource action: lsb-dummy start on pcmk-1 * Resource action: lsb-dummy monitor=5000 on pcmk-1 Revised Cluster Status: * Node List: * Online: [ pcmk-1 pcmk-2 pcmk-3 ] * OFFLINE: [ pcmk-4 ] * Full List of Resources: * Clone Set: Fencing [FencingChild]: * Started: [ pcmk-1 pcmk-2 pcmk-3 ] * Stopped: [ pcmk-4 ] * Resource Group: group-1: * r192.168.101.181 (ocf:heartbeat:IPaddr): Started pcmk-1 * r192.168.101.182 (ocf:heartbeat:IPaddr): Started pcmk-1 * r192.168.101.183 (ocf:heartbeat:IPaddr): Started pcmk-1 * rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 * rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 * rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 * rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-2 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-1 * migrator (ocf:pacemaker:Dummy): Started pcmk-3 * Clone Set: Connectivity [ping-1]: * Started: [ pcmk-1 pcmk-2 pcmk-3 ] * Stopped: [ pcmk-4 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ pcmk-1 ] * Unpromoted: [ pcmk-2 pcmk-3 ] * Stopped: [ pcmk-4 ] diff --git a/cts/scheduler/summary/migrate-shutdown.summary b/cts/scheduler/summary/migrate-shutdown.summary index 551a41a175..1da9db21e8 100644 --- a/cts/scheduler/summary/migrate-shutdown.summary +++ b/cts/scheduler/summary/migrate-shutdown.summary @@ -1,92 +1,92 @@ Current cluster status: * Node List: * Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started pcmk-1 * Resource Group: group-1: * r192.168.122.105 (ocf:heartbeat:IPaddr): Started pcmk-2 * r192.168.122.106 (ocf:heartbeat:IPaddr): Started pcmk-2 * r192.168.122.107 (ocf:heartbeat:IPaddr): Started pcmk-2 * rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 * rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 * rsc_pcmk-3 (ocf:heartbeat:IPaddr): Stopped * rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-2 * migrator (ocf:pacemaker:Dummy): Started pcmk-1 * Clone Set: Connectivity [ping-1]: * Started: [ pcmk-1 pcmk-2 pcmk-4 ] * Stopped: [ pcmk-3 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ pcmk-2 ] * Unpromoted: [ pcmk-1 pcmk-4 ] * Stopped: [ pcmk-3 ] Transition Summary: * Stop Fencing ( pcmk-1 ) due to node availability * Stop r192.168.122.105 ( pcmk-2 ) due to node availability * Stop r192.168.122.106 ( pcmk-2 ) due to node availability * Stop r192.168.122.107 ( pcmk-2 ) due to node availability * Stop rsc_pcmk-1 ( pcmk-1 ) due to node availability * Stop rsc_pcmk-2 ( pcmk-2 ) due to node availability * Stop rsc_pcmk-4 ( pcmk-4 ) due to node availability * Stop lsb-dummy ( pcmk-2 ) due to node availability * Stop migrator ( pcmk-1 ) due to node availability * Stop ping-1:0 ( pcmk-1 ) due to node availability * Stop ping-1:1 ( pcmk-2 ) due to node availability * Stop ping-1:2 ( pcmk-4 ) due to node availability - * Stop stateful-1:0 ( Unpromoted pcmk-1 ) due to node availability - * Stop stateful-1:1 ( Promoted pcmk-2 ) due to node availability - * Stop stateful-1:2 ( Unpromoted pcmk-4 ) due to node availability + * Stop stateful-1:0 ( Unpromoted pcmk-1 ) due to node availability + * Stop stateful-1:1 ( Promoted pcmk-2 ) due to node availability + * Stop stateful-1:2 ( Unpromoted pcmk-4 ) due to node availability Executing Cluster Transition: * Resource action: Fencing stop on pcmk-1 * Resource action: rsc_pcmk-1 stop on pcmk-1 * Resource action: rsc_pcmk-2 stop on pcmk-2 * Resource action: rsc_pcmk-4 stop on pcmk-4 * Resource action: lsb-dummy stop on pcmk-2 * Resource action: migrator stop on pcmk-1 * Resource action: migrator stop on pcmk-3 * Pseudo action: Connectivity_stop_0 * Cluster action: do_shutdown on pcmk-3 * Pseudo action: group-1_stop_0 * Resource action: r192.168.122.107 stop on pcmk-2 * Resource action: ping-1:0 stop on pcmk-1 * Resource action: ping-1:1 stop on pcmk-2 * Resource action: ping-1:3 stop on pcmk-4 * Pseudo action: Connectivity_stopped_0 * Resource action: r192.168.122.106 stop on pcmk-2 * Resource action: r192.168.122.105 stop on pcmk-2 * Pseudo action: group-1_stopped_0 * Pseudo action: master-1_demote_0 * Resource action: stateful-1:0 demote on pcmk-2 * Pseudo action: master-1_demoted_0 * Pseudo action: master-1_stop_0 * Resource action: stateful-1:2 stop on pcmk-1 * Resource action: stateful-1:0 stop on pcmk-2 * Resource action: stateful-1:3 stop on pcmk-4 * Pseudo action: master-1_stopped_0 * Cluster action: do_shutdown on pcmk-4 * Cluster action: do_shutdown on pcmk-2 * Cluster action: do_shutdown on pcmk-1 Revised Cluster Status: * Node List: * Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Stopped * Resource Group: group-1: * r192.168.122.105 (ocf:heartbeat:IPaddr): Stopped * r192.168.122.106 (ocf:heartbeat:IPaddr): Stopped * r192.168.122.107 (ocf:heartbeat:IPaddr): Stopped * rsc_pcmk-1 (ocf:heartbeat:IPaddr): Stopped * rsc_pcmk-2 (ocf:heartbeat:IPaddr): Stopped * rsc_pcmk-3 (ocf:heartbeat:IPaddr): Stopped * rsc_pcmk-4 (ocf:heartbeat:IPaddr): Stopped * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Stopped * migrator (ocf:pacemaker:Dummy): Stopped * Clone Set: Connectivity [ping-1]: * Stopped: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Clone Set: master-1 [stateful-1] (promotable): * Stopped: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] diff --git a/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary b/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary index 532f731235..8eb68a4cb9 100644 --- a/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary +++ b/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary @@ -1,103 +1,103 @@ Using the original execution date of: 2020-05-14 10:49:31Z Current cluster status: * Node List: * Online: [ controller-0 controller-1 controller-2 ] * GuestOnline: [ galera-bundle-0@controller-0 galera-bundle-1@controller-1 galera-bundle-2@controller-2 ovn-dbs-bundle-0@controller-0 ovn-dbs-bundle-1@controller-1 ovn-dbs-bundle-2@controller-2 rabbitmq-bundle-0@controller-0 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ] * Full List of Resources: * Container bundle set: galera-bundle [cluster.common.tag/rhosp16-openstack-mariadb:pcmklatest]: * galera-bundle-0 (ocf:heartbeat:galera): Promoted controller-0 * galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1 * galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2 * Container bundle set: rabbitmq-bundle [cluster.common.tag/rhosp16-openstack-rabbitmq:pcmklatest]: * rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started controller-0 * rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1 * rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2 * Container bundle set: redis-bundle [cluster.common.tag/rhosp16-openstack-redis:pcmklatest]: * redis-bundle-0 (ocf:heartbeat:redis): Promoted controller-0 * redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1 * redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2 * Container bundle set: ovn-dbs-bundle [cluster.common.tag/rhosp16-openstack-ovn-northd:pcmklatest]: * ovn-dbs-bundle-0 (ocf:ovn:ovndb-servers): Unpromoted controller-0 * ovn-dbs-bundle-1 (ocf:ovn:ovndb-servers): Unpromoted controller-1 * ovn-dbs-bundle-2 (ocf:ovn:ovndb-servers): Unpromoted controller-2 * stonith-fence_ipmilan-5254005e097a (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400afe30e (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400985679 (stonith:fence_ipmilan): Started controller-1 * Container bundle: openstack-cinder-volume [cluster.common.tag/rhosp16-openstack-cinder-volume:pcmklatest]: * openstack-cinder-volume-podman-0 (ocf:heartbeat:podman): Started controller-0 Transition Summary: * Stop ovn-dbs-bundle-podman-0 ( controller-0 ) due to node availability * Stop ovn-dbs-bundle-0 ( controller-0 ) due to unrunnable ovn-dbs-bundle-podman-0 start - * Stop ovndb_servers:0 ( Unpromoted ovn-dbs-bundle-0 ) due to unrunnable ovn-dbs-bundle-podman-0 start + * Stop ovndb_servers:0 ( Unpromoted ovn-dbs-bundle-0 ) due to unrunnable ovn-dbs-bundle-podman-0 start * Promote ovndb_servers:1 ( Unpromoted -> Promoted ovn-dbs-bundle-1 ) Executing Cluster Transition: * Resource action: ovndb_servers cancel=30000 on ovn-dbs-bundle-1 * Pseudo action: ovn-dbs-bundle-master_pre_notify_stop_0 * Pseudo action: ovn-dbs-bundle_stop_0 * Resource action: ovndb_servers notify on ovn-dbs-bundle-0 * Resource action: ovndb_servers notify on ovn-dbs-bundle-1 * Resource action: ovndb_servers notify on ovn-dbs-bundle-2 * Pseudo action: ovn-dbs-bundle-master_confirmed-pre_notify_stop_0 * Pseudo action: ovn-dbs-bundle-master_stop_0 * Resource action: ovndb_servers stop on ovn-dbs-bundle-0 * Pseudo action: ovn-dbs-bundle-master_stopped_0 * Resource action: ovn-dbs-bundle-0 stop on controller-0 * Pseudo action: ovn-dbs-bundle-master_post_notify_stopped_0 * Resource action: ovn-dbs-bundle-podman-0 stop on controller-0 * Resource action: ovndb_servers notify on ovn-dbs-bundle-1 * Resource action: ovndb_servers notify on ovn-dbs-bundle-2 * Pseudo action: ovn-dbs-bundle-master_confirmed-post_notify_stopped_0 * Pseudo action: ovn-dbs-bundle-master_pre_notify_start_0 * Pseudo action: ovn-dbs-bundle_stopped_0 * Pseudo action: ovn-dbs-bundle-master_confirmed-pre_notify_start_0 * Pseudo action: ovn-dbs-bundle-master_start_0 * Pseudo action: ovn-dbs-bundle-master_running_0 * Pseudo action: ovn-dbs-bundle-master_post_notify_running_0 * Pseudo action: ovn-dbs-bundle-master_confirmed-post_notify_running_0 * Pseudo action: ovn-dbs-bundle_running_0 * Pseudo action: ovn-dbs-bundle-master_pre_notify_promote_0 * Pseudo action: ovn-dbs-bundle_promote_0 * Resource action: ovndb_servers notify on ovn-dbs-bundle-1 * Resource action: ovndb_servers notify on ovn-dbs-bundle-2 * Pseudo action: ovn-dbs-bundle-master_confirmed-pre_notify_promote_0 * Pseudo action: ovn-dbs-bundle-master_promote_0 * Resource action: ovndb_servers promote on ovn-dbs-bundle-1 * Pseudo action: ovn-dbs-bundle-master_promoted_0 * Pseudo action: ovn-dbs-bundle-master_post_notify_promoted_0 * Resource action: ovndb_servers notify on ovn-dbs-bundle-1 * Resource action: ovndb_servers notify on ovn-dbs-bundle-2 * Pseudo action: ovn-dbs-bundle-master_confirmed-post_notify_promoted_0 * Pseudo action: ovn-dbs-bundle_promoted_0 * Resource action: ovndb_servers monitor=10000 on ovn-dbs-bundle-1 Using the original execution date of: 2020-05-14 10:49:31Z Revised Cluster Status: * Node List: * Online: [ controller-0 controller-1 controller-2 ] * GuestOnline: [ galera-bundle-0@controller-0 galera-bundle-1@controller-1 galera-bundle-2@controller-2 ovn-dbs-bundle-1@controller-1 ovn-dbs-bundle-2@controller-2 rabbitmq-bundle-0@controller-0 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ] * Full List of Resources: * Container bundle set: galera-bundle [cluster.common.tag/rhosp16-openstack-mariadb:pcmklatest]: * galera-bundle-0 (ocf:heartbeat:galera): Promoted controller-0 * galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1 * galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2 * Container bundle set: rabbitmq-bundle [cluster.common.tag/rhosp16-openstack-rabbitmq:pcmklatest]: * rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started controller-0 * rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1 * rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2 * Container bundle set: redis-bundle [cluster.common.tag/rhosp16-openstack-redis:pcmklatest]: * redis-bundle-0 (ocf:heartbeat:redis): Promoted controller-0 * redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1 * redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2 * Container bundle set: ovn-dbs-bundle [cluster.common.tag/rhosp16-openstack-ovn-northd:pcmklatest]: * ovn-dbs-bundle-0 (ocf:ovn:ovndb-servers): Stopped * ovn-dbs-bundle-1 (ocf:ovn:ovndb-servers): Promoted controller-1 * ovn-dbs-bundle-2 (ocf:ovn:ovndb-servers): Unpromoted controller-2 * stonith-fence_ipmilan-5254005e097a (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400afe30e (stonith:fence_ipmilan): Started controller-2 * stonith-fence_ipmilan-525400985679 (stonith:fence_ipmilan): Started controller-1 * Container bundle: openstack-cinder-volume [cluster.common.tag/rhosp16-openstack-cinder-volume:pcmklatest]: * openstack-cinder-volume-podman-0 (ocf:heartbeat:podman): Started controller-0 diff --git a/cts/scheduler/summary/novell-239082.summary b/cts/scheduler/summary/novell-239082.summary index 431b6ddc63..01af7656e9 100644 --- a/cts/scheduler/summary/novell-239082.summary +++ b/cts/scheduler/summary/novell-239082.summary @@ -1,59 +1,59 @@ Current cluster status: * Node List: * Online: [ xen-1 xen-2 ] * Full List of Resources: * fs_1 (ocf:heartbeat:Filesystem): Started xen-1 * Clone Set: ms-drbd0 [drbd0] (promotable): * Promoted: [ xen-1 ] * Unpromoted: [ xen-2 ] Transition Summary: * Move fs_1 ( xen-1 -> xen-2 ) * Promote drbd0:0 ( Unpromoted -> Promoted xen-2 ) - * Stop drbd0:1 ( Promoted xen-1 ) due to node availability + * Stop drbd0:1 ( Promoted xen-1 ) due to node availability Executing Cluster Transition: * Resource action: fs_1 stop on xen-1 * Pseudo action: ms-drbd0_pre_notify_demote_0 * Resource action: drbd0:0 notify on xen-2 * Resource action: drbd0:1 notify on xen-1 * Pseudo action: ms-drbd0_confirmed-pre_notify_demote_0 * Pseudo action: ms-drbd0_demote_0 * Resource action: drbd0:1 demote on xen-1 * Pseudo action: ms-drbd0_demoted_0 * Pseudo action: ms-drbd0_post_notify_demoted_0 * Resource action: drbd0:0 notify on xen-2 * Resource action: drbd0:1 notify on xen-1 * Pseudo action: ms-drbd0_confirmed-post_notify_demoted_0 * Pseudo action: ms-drbd0_pre_notify_stop_0 * Resource action: drbd0:0 notify on xen-2 * Resource action: drbd0:1 notify on xen-1 * Pseudo action: ms-drbd0_confirmed-pre_notify_stop_0 * Pseudo action: ms-drbd0_stop_0 * Resource action: drbd0:1 stop on xen-1 * Pseudo action: ms-drbd0_stopped_0 * Cluster action: do_shutdown on xen-1 * Pseudo action: ms-drbd0_post_notify_stopped_0 * Resource action: drbd0:0 notify on xen-2 * Pseudo action: ms-drbd0_confirmed-post_notify_stopped_0 * Pseudo action: ms-drbd0_pre_notify_promote_0 * Resource action: drbd0:0 notify on xen-2 * Pseudo action: ms-drbd0_confirmed-pre_notify_promote_0 * Pseudo action: ms-drbd0_promote_0 * Resource action: drbd0:0 promote on xen-2 * Pseudo action: ms-drbd0_promoted_0 * Pseudo action: ms-drbd0_post_notify_promoted_0 * Resource action: drbd0:0 notify on xen-2 * Pseudo action: ms-drbd0_confirmed-post_notify_promoted_0 * Resource action: fs_1 start on xen-2 Revised Cluster Status: * Node List: * Online: [ xen-1 xen-2 ] * Full List of Resources: * fs_1 (ocf:heartbeat:Filesystem): Started xen-2 * Clone Set: ms-drbd0 [drbd0] (promotable): * Promoted: [ xen-2 ] * Stopped: [ xen-1 ] diff --git a/cts/scheduler/summary/on_fail_demote4.summary b/cts/scheduler/summary/on_fail_demote4.summary index 781f5488bb..b7b1388e58 100644 --- a/cts/scheduler/summary/on_fail_demote4.summary +++ b/cts/scheduler/summary/on_fail_demote4.summary @@ -1,189 +1,189 @@ Using the original execution date of: 2020-06-16 19:23:21Z Current cluster status: * Node List: * RemoteNode remote-rhel7-2: UNCLEAN (offline) * Node rhel7-4: UNCLEAN (offline) * Online: [ rhel7-1 rhel7-3 rhel7-5 ] * GuestOnline: [ lxc1@rhel7-3 stateful-bundle-1@rhel7-1 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-4 (UNCLEAN) * Clone Set: rsc1-clone [rsc1] (promotable): * rsc1 (ocf:pacemaker:Stateful): Promoted rhel7-4 (UNCLEAN) * rsc1 (ocf:pacemaker:Stateful): Unpromoted remote-rhel7-2 (UNCLEAN) * Unpromoted: [ lxc1 rhel7-1 rhel7-3 rhel7-5 ] * Clone Set: rsc2-master [rsc2] (promotable): * rsc2 (ocf:pacemaker:Stateful): Unpromoted rhel7-4 (UNCLEAN) * rsc2 (ocf:pacemaker:Stateful): Promoted remote-rhel7-2 (UNCLEAN) * Unpromoted: [ lxc1 rhel7-1 rhel7-3 rhel7-5 ] * remote-rhel7-2 (ocf:pacemaker:remote): FAILED rhel7-1 * container1 (ocf:heartbeat:VirtualDomain): Started rhel7-3 * container2 (ocf:heartbeat:VirtualDomain): FAILED rhel7-3 * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Unpromoted: [ lxc1 ] * Stopped: [ remote-rhel7-2 rhel7-1 rhel7-3 rhel7-4 rhel7-5 ] * Container bundle set: stateful-bundle [pcmktest:http]: * stateful-bundle-0 (192.168.122.131) (ocf:pacemaker:Stateful): FAILED Promoted rhel7-5 * stateful-bundle-1 (192.168.122.132) (ocf:pacemaker:Stateful): Unpromoted rhel7-1 * stateful-bundle-2 (192.168.122.133) (ocf:pacemaker:Stateful): FAILED rhel7-4 (UNCLEAN) Transition Summary: * Fence (reboot) stateful-bundle-2 (resource: stateful-bundle-docker-2) 'guest is unclean' * Fence (reboot) stateful-bundle-0 (resource: stateful-bundle-docker-0) 'guest is unclean' * Fence (reboot) lxc2 (resource: container2) 'guest is unclean' * Fence (reboot) remote-rhel7-2 'remote connection is unrecoverable' * Fence (reboot) rhel7-4 'peer is no longer part of the cluster' * Move Fencing ( rhel7-4 -> rhel7-5 ) - * Stop rsc1:0 ( Promoted rhel7-4 ) due to node availability - * Promote rsc1:1 ( Unpromoted -> Promoted rhel7-3 ) - * Stop rsc1:4 ( Unpromoted remote-rhel7-2 ) due to node availability - * Recover rsc1:5 ( Unpromoted lxc2 ) - * Stop rsc2:0 ( Unpromoted rhel7-4 ) due to node availability - * Promote rsc2:1 ( Unpromoted -> Promoted rhel7-3 ) - * Stop rsc2:4 ( Promoted remote-rhel7-2 ) due to node availability - * Recover rsc2:5 ( Unpromoted lxc2 ) + * Stop rsc1:0 ( Promoted rhel7-4 ) due to node availability + * Promote rsc1:1 ( Unpromoted -> Promoted rhel7-3 ) + * Stop rsc1:4 ( Unpromoted remote-rhel7-2 ) due to node availability + * Recover rsc1:5 ( Unpromoted lxc2 ) + * Stop rsc2:0 ( Unpromoted rhel7-4 ) due to node availability + * Promote rsc2:1 ( Unpromoted -> Promoted rhel7-3 ) + * Stop rsc2:4 ( Promoted remote-rhel7-2 ) due to node availability + * Recover rsc2:5 ( Unpromoted lxc2 ) * Recover remote-rhel7-2 ( rhel7-1 ) * Recover container2 ( rhel7-3 ) - * Recover lxc-ms:0 ( Promoted lxc2 ) + * Recover lxc-ms:0 ( Promoted lxc2 ) * Recover stateful-bundle-docker-0 ( rhel7-5 ) * Restart stateful-bundle-0 ( rhel7-5 ) due to required stateful-bundle-docker-0 start - * Recover bundled:0 ( Promoted stateful-bundle-0 ) + * Recover bundled:0 ( Promoted stateful-bundle-0 ) * Move stateful-bundle-ip-192.168.122.133 ( rhel7-4 -> rhel7-3 ) * Recover stateful-bundle-docker-2 ( rhel7-4 -> rhel7-3 ) * Move stateful-bundle-2 ( rhel7-4 -> rhel7-3 ) - * Recover bundled:2 ( Unpromoted stateful-bundle-2 ) + * Recover bundled:2 ( Unpromoted stateful-bundle-2 ) * Restart lxc2 ( rhel7-3 ) due to required container2 start Executing Cluster Transition: * Pseudo action: Fencing_stop_0 * Resource action: rsc1 cancel=11000 on rhel7-3 * Pseudo action: rsc1-clone_demote_0 * Resource action: rsc2 cancel=11000 on rhel7-3 * Pseudo action: rsc2-master_demote_0 * Pseudo action: lxc-ms-master_demote_0 * Resource action: stateful-bundle-0 stop on rhel7-5 * Pseudo action: stateful-bundle-2_stop_0 * Resource action: lxc2 stop on rhel7-3 * Pseudo action: stateful-bundle_demote_0 * Fencing remote-rhel7-2 (reboot) * Fencing rhel7-4 (reboot) * Pseudo action: rsc1_demote_0 * Pseudo action: rsc1-clone_demoted_0 * Pseudo action: rsc2_demote_0 * Pseudo action: rsc2-master_demoted_0 * Resource action: container2 stop on rhel7-3 * Pseudo action: stateful-bundle-master_demote_0 * Pseudo action: stonith-stateful-bundle-2-reboot on stateful-bundle-2 * Pseudo action: stonith-lxc2-reboot on lxc2 * Resource action: Fencing start on rhel7-5 * Pseudo action: rsc1-clone_stop_0 * Pseudo action: rsc2-master_stop_0 * Pseudo action: lxc-ms_demote_0 * Pseudo action: lxc-ms-master_demoted_0 * Pseudo action: lxc-ms-master_stop_0 * Pseudo action: bundled_demote_0 * Pseudo action: stateful-bundle-master_demoted_0 * Pseudo action: stateful-bundle_demoted_0 * Pseudo action: stateful-bundle_stop_0 * Resource action: Fencing monitor=120000 on rhel7-5 * Pseudo action: rsc1_stop_0 * Pseudo action: rsc1_stop_0 * Pseudo action: rsc1_stop_0 * Pseudo action: rsc1-clone_stopped_0 * Pseudo action: rsc1-clone_start_0 * Pseudo action: rsc2_stop_0 * Pseudo action: rsc2_stop_0 * Pseudo action: rsc2_stop_0 * Pseudo action: rsc2-master_stopped_0 * Pseudo action: rsc2-master_start_0 * Resource action: remote-rhel7-2 stop on rhel7-1 * Pseudo action: lxc-ms_stop_0 * Pseudo action: lxc-ms-master_stopped_0 * Pseudo action: lxc-ms-master_start_0 * Resource action: stateful-bundle-docker-0 stop on rhel7-5 * Pseudo action: stateful-bundle-docker-2_stop_0 * Pseudo action: stonith-stateful-bundle-0-reboot on stateful-bundle-0 * Resource action: remote-rhel7-2 start on rhel7-1 * Resource action: remote-rhel7-2 monitor=60000 on rhel7-1 * Resource action: container2 start on rhel7-3 * Resource action: container2 monitor=20000 on rhel7-3 * Pseudo action: stateful-bundle-master_stop_0 * Pseudo action: stateful-bundle-ip-192.168.122.133_stop_0 * Resource action: lxc2 start on rhel7-3 * Resource action: lxc2 monitor=30000 on rhel7-3 * Resource action: rsc1 start on lxc2 * Pseudo action: rsc1-clone_running_0 * Resource action: rsc2 start on lxc2 * Pseudo action: rsc2-master_running_0 * Resource action: lxc-ms start on lxc2 * Pseudo action: lxc-ms-master_running_0 * Pseudo action: bundled_stop_0 * Resource action: stateful-bundle-ip-192.168.122.133 start on rhel7-3 * Resource action: rsc1 monitor=11000 on lxc2 * Pseudo action: rsc1-clone_promote_0 * Resource action: rsc2 monitor=11000 on lxc2 * Pseudo action: rsc2-master_promote_0 * Pseudo action: lxc-ms-master_promote_0 * Pseudo action: bundled_stop_0 * Pseudo action: stateful-bundle-master_stopped_0 * Resource action: stateful-bundle-ip-192.168.122.133 monitor=60000 on rhel7-3 * Pseudo action: stateful-bundle_stopped_0 * Pseudo action: stateful-bundle_start_0 * Resource action: rsc1 promote on rhel7-3 * Pseudo action: rsc1-clone_promoted_0 * Resource action: rsc2 promote on rhel7-3 * Pseudo action: rsc2-master_promoted_0 * Resource action: lxc-ms promote on lxc2 * Pseudo action: lxc-ms-master_promoted_0 * Pseudo action: stateful-bundle-master_start_0 * Resource action: stateful-bundle-docker-0 start on rhel7-5 * Resource action: stateful-bundle-docker-0 monitor=60000 on rhel7-5 * Resource action: stateful-bundle-0 start on rhel7-5 * Resource action: stateful-bundle-0 monitor=30000 on rhel7-5 * Resource action: stateful-bundle-docker-2 start on rhel7-3 * Resource action: stateful-bundle-2 start on rhel7-3 * Resource action: rsc1 monitor=10000 on rhel7-3 * Resource action: rsc2 monitor=10000 on rhel7-3 * Resource action: lxc-ms monitor=10000 on lxc2 * Resource action: bundled start on stateful-bundle-0 * Resource action: bundled start on stateful-bundle-2 * Pseudo action: stateful-bundle-master_running_0 * Resource action: stateful-bundle-docker-2 monitor=60000 on rhel7-3 * Resource action: stateful-bundle-2 monitor=30000 on rhel7-3 * Pseudo action: stateful-bundle_running_0 * Resource action: bundled monitor=11000 on stateful-bundle-2 * Pseudo action: stateful-bundle_promote_0 * Pseudo action: stateful-bundle-master_promote_0 * Resource action: bundled promote on stateful-bundle-0 * Pseudo action: stateful-bundle-master_promoted_0 * Pseudo action: stateful-bundle_promoted_0 * Resource action: bundled monitor=10000 on stateful-bundle-0 Using the original execution date of: 2020-06-16 19:23:21Z Revised Cluster Status: * Node List: * Online: [ rhel7-1 rhel7-3 rhel7-5 ] * OFFLINE: [ rhel7-4 ] * RemoteOnline: [ remote-rhel7-2 ] * GuestOnline: [ lxc1@rhel7-3 lxc2@rhel7-3 stateful-bundle-0@rhel7-5 stateful-bundle-1@rhel7-1 stateful-bundle-2@rhel7-3 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-5 * Clone Set: rsc1-clone [rsc1] (promotable): * Promoted: [ rhel7-3 ] * Unpromoted: [ lxc1 lxc2 rhel7-1 rhel7-5 ] * Stopped: [ remote-rhel7-2 rhel7-4 ] * Clone Set: rsc2-master [rsc2] (promotable): * Promoted: [ rhel7-3 ] * Unpromoted: [ lxc1 lxc2 rhel7-1 rhel7-5 ] * Stopped: [ remote-rhel7-2 rhel7-4 ] * remote-rhel7-2 (ocf:pacemaker:remote): Started rhel7-1 * container1 (ocf:heartbeat:VirtualDomain): Started rhel7-3 * container2 (ocf:heartbeat:VirtualDomain): Started rhel7-3 * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Promoted: [ lxc2 ] * Unpromoted: [ lxc1 ] * Container bundle set: stateful-bundle [pcmktest:http]: * stateful-bundle-0 (192.168.122.131) (ocf:pacemaker:Stateful): Promoted rhel7-5 * stateful-bundle-1 (192.168.122.132) (ocf:pacemaker:Stateful): Unpromoted rhel7-1 * stateful-bundle-2 (192.168.122.133) (ocf:pacemaker:Stateful): Unpromoted rhel7-3 diff --git a/cts/scheduler/summary/probe-2.summary b/cts/scheduler/summary/probe-2.summary index f2c60821ab..3523891d30 100644 --- a/cts/scheduler/summary/probe-2.summary +++ b/cts/scheduler/summary/probe-2.summary @@ -1,163 +1,163 @@ Current cluster status: * Node List: * Node wc02: standby (with active resources) * Online: [ wc01 ] * Full List of Resources: * Resource Group: group_www_data: * fs_www_data (ocf:heartbeat:Filesystem): Started wc01 * nfs-kernel-server (lsb:nfs-kernel-server): Started wc01 * intip_nfs (ocf:heartbeat:IPaddr2): Started wc01 * Clone Set: ms_drbd_mysql [drbd_mysql] (promotable): * Promoted: [ wc02 ] * Unpromoted: [ wc01 ] * Resource Group: group_mysql: * fs_mysql (ocf:heartbeat:Filesystem): Started wc02 * intip_sql (ocf:heartbeat:IPaddr2): Started wc02 * mysql-server (ocf:heartbeat:mysql): Started wc02 * Clone Set: ms_drbd_www [drbd_www] (promotable): * Promoted: [ wc01 ] * Unpromoted: [ wc02 ] * Clone Set: clone_nfs-common [group_nfs-common]: * Started: [ wc01 wc02 ] * Clone Set: clone_mysql-proxy [group_mysql-proxy]: * Started: [ wc01 wc02 ] * Clone Set: clone_webservice [group_webservice]: * Started: [ wc01 wc02 ] * Resource Group: group_ftpd: * extip_ftp (ocf:heartbeat:IPaddr2): Started wc01 * pure-ftpd (ocf:heartbeat:Pure-FTPd): Started wc01 * Clone Set: DoFencing [stonith_rackpdu] (unique): * stonith_rackpdu:0 (stonith:external/rackpdu): Started wc01 * stonith_rackpdu:1 (stonith:external/rackpdu): Started wc02 Transition Summary: * Promote drbd_mysql:0 ( Unpromoted -> Promoted wc01 ) - * Stop drbd_mysql:1 ( Promoted wc02 ) due to node availability + * Stop drbd_mysql:1 ( Promoted wc02 ) due to node availability * Move fs_mysql ( wc02 -> wc01 ) * Move intip_sql ( wc02 -> wc01 ) * Move mysql-server ( wc02 -> wc01 ) - * Stop drbd_www:1 ( Unpromoted wc02 ) due to node availability + * Stop drbd_www:1 ( Unpromoted wc02 ) due to node availability * Stop nfs-common:1 ( wc02 ) due to node availability * Stop mysql-proxy:1 ( wc02 ) due to node availability * Stop fs_www:1 ( wc02 ) due to node availability * Stop apache2:1 ( wc02 ) due to node availability * Restart stonith_rackpdu:0 ( wc01 ) * Stop stonith_rackpdu:1 ( wc02 ) due to node availability Executing Cluster Transition: * Resource action: drbd_mysql:0 cancel=10000 on wc01 * Pseudo action: ms_drbd_mysql_pre_notify_demote_0 * Pseudo action: group_mysql_stop_0 * Resource action: mysql-server stop on wc02 * Pseudo action: ms_drbd_www_pre_notify_stop_0 * Pseudo action: clone_mysql-proxy_stop_0 * Pseudo action: clone_webservice_stop_0 * Pseudo action: DoFencing_stop_0 * Resource action: drbd_mysql:0 notify on wc01 * Resource action: drbd_mysql:1 notify on wc02 * Pseudo action: ms_drbd_mysql_confirmed-pre_notify_demote_0 * Resource action: intip_sql stop on wc02 * Resource action: drbd_www:0 notify on wc01 * Resource action: drbd_www:1 notify on wc02 * Pseudo action: ms_drbd_www_confirmed-pre_notify_stop_0 * Pseudo action: ms_drbd_www_stop_0 * Pseudo action: group_mysql-proxy:1_stop_0 * Resource action: mysql-proxy:1 stop on wc02 * Pseudo action: group_webservice:1_stop_0 * Resource action: apache2:1 stop on wc02 * Resource action: stonith_rackpdu:0 stop on wc01 * Resource action: stonith_rackpdu:1 stop on wc02 * Pseudo action: DoFencing_stopped_0 * Pseudo action: DoFencing_start_0 * Resource action: fs_mysql stop on wc02 * Resource action: drbd_www:1 stop on wc02 * Pseudo action: ms_drbd_www_stopped_0 * Pseudo action: group_mysql-proxy:1_stopped_0 * Pseudo action: clone_mysql-proxy_stopped_0 * Resource action: fs_www:1 stop on wc02 * Resource action: stonith_rackpdu:0 start on wc01 * Pseudo action: DoFencing_running_0 * Pseudo action: group_mysql_stopped_0 * Pseudo action: ms_drbd_www_post_notify_stopped_0 * Pseudo action: group_webservice:1_stopped_0 * Pseudo action: clone_webservice_stopped_0 * Resource action: stonith_rackpdu:0 monitor=5000 on wc01 * Pseudo action: ms_drbd_mysql_demote_0 * Resource action: drbd_www:0 notify on wc01 * Pseudo action: ms_drbd_www_confirmed-post_notify_stopped_0 * Pseudo action: clone_nfs-common_stop_0 * Resource action: drbd_mysql:1 demote on wc02 * Pseudo action: ms_drbd_mysql_demoted_0 * Pseudo action: group_nfs-common:1_stop_0 * Resource action: nfs-common:1 stop on wc02 * Pseudo action: ms_drbd_mysql_post_notify_demoted_0 * Pseudo action: group_nfs-common:1_stopped_0 * Pseudo action: clone_nfs-common_stopped_0 * Resource action: drbd_mysql:0 notify on wc01 * Resource action: drbd_mysql:1 notify on wc02 * Pseudo action: ms_drbd_mysql_confirmed-post_notify_demoted_0 * Pseudo action: ms_drbd_mysql_pre_notify_stop_0 * Resource action: drbd_mysql:0 notify on wc01 * Resource action: drbd_mysql:1 notify on wc02 * Pseudo action: ms_drbd_mysql_confirmed-pre_notify_stop_0 * Pseudo action: ms_drbd_mysql_stop_0 * Resource action: drbd_mysql:1 stop on wc02 * Pseudo action: ms_drbd_mysql_stopped_0 * Pseudo action: ms_drbd_mysql_post_notify_stopped_0 * Resource action: drbd_mysql:0 notify on wc01 * Pseudo action: ms_drbd_mysql_confirmed-post_notify_stopped_0 * Pseudo action: ms_drbd_mysql_pre_notify_promote_0 * Resource action: drbd_mysql:0 notify on wc01 * Pseudo action: ms_drbd_mysql_confirmed-pre_notify_promote_0 * Pseudo action: ms_drbd_mysql_promote_0 * Resource action: drbd_mysql:0 promote on wc01 * Pseudo action: ms_drbd_mysql_promoted_0 * Pseudo action: ms_drbd_mysql_post_notify_promoted_0 * Resource action: drbd_mysql:0 notify on wc01 * Pseudo action: ms_drbd_mysql_confirmed-post_notify_promoted_0 * Pseudo action: group_mysql_start_0 * Resource action: fs_mysql start on wc01 * Resource action: intip_sql start on wc01 * Resource action: mysql-server start on wc01 * Resource action: drbd_mysql:0 monitor=5000 on wc01 * Pseudo action: group_mysql_running_0 * Resource action: fs_mysql monitor=30000 on wc01 * Resource action: intip_sql monitor=30000 on wc01 * Resource action: mysql-server monitor=30000 on wc01 Revised Cluster Status: * Node List: * Node wc02: standby * Online: [ wc01 ] * Full List of Resources: * Resource Group: group_www_data: * fs_www_data (ocf:heartbeat:Filesystem): Started wc01 * nfs-kernel-server (lsb:nfs-kernel-server): Started wc01 * intip_nfs (ocf:heartbeat:IPaddr2): Started wc01 * Clone Set: ms_drbd_mysql [drbd_mysql] (promotable): * Promoted: [ wc01 ] * Stopped: [ wc02 ] * Resource Group: group_mysql: * fs_mysql (ocf:heartbeat:Filesystem): Started wc01 * intip_sql (ocf:heartbeat:IPaddr2): Started wc01 * mysql-server (ocf:heartbeat:mysql): Started wc01 * Clone Set: ms_drbd_www [drbd_www] (promotable): * Promoted: [ wc01 ] * Stopped: [ wc02 ] * Clone Set: clone_nfs-common [group_nfs-common]: * Started: [ wc01 ] * Stopped: [ wc02 ] * Clone Set: clone_mysql-proxy [group_mysql-proxy]: * Started: [ wc01 ] * Stopped: [ wc02 ] * Clone Set: clone_webservice [group_webservice]: * Started: [ wc01 ] * Stopped: [ wc02 ] * Resource Group: group_ftpd: * extip_ftp (ocf:heartbeat:IPaddr2): Started wc01 * pure-ftpd (ocf:heartbeat:Pure-FTPd): Started wc01 * Clone Set: DoFencing [stonith_rackpdu] (unique): * stonith_rackpdu:0 (stonith:external/rackpdu): Started wc01 * stonith_rackpdu:1 (stonith:external/rackpdu): Stopped diff --git a/cts/scheduler/summary/promoted-7.summary b/cts/scheduler/summary/promoted-7.summary index 4fc3a85e9a..0602f95895 100644 --- a/cts/scheduler/summary/promoted-7.summary +++ b/cts/scheduler/summary/promoted-7.summary @@ -1,121 +1,121 @@ Current cluster status: * Node List: * Node c001n01: UNCLEAN (offline) * Online: [ c001n02 c001n03 c001n08 ] * Full List of Resources: * DcIPaddr (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN) * Resource Group: group-1: * ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n03 * heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n03 * ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n03 * lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n02 * rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN) * rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08 * rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02 * rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03 * Clone Set: DoFencing [child_DoFencing] (unique): * child_DoFencing:0 (stonith:ssh): Started c001n01 (UNCLEAN) * child_DoFencing:1 (stonith:ssh): Started c001n03 * child_DoFencing:2 (stonith:ssh): Started c001n02 * child_DoFencing:3 (stonith:ssh): Started c001n08 * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique): * ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n01 (UNCLEAN) * ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03 * ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02 * ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08 * ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01 (UNCLEAN) * ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03 * ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02 * ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08 Transition Summary: * Fence (reboot) c001n01 'peer is no longer part of the cluster' * Move DcIPaddr ( c001n01 -> c001n03 ) * Move ocf_192.168.100.181 ( c001n03 -> c001n02 ) * Move heartbeat_192.168.100.182 ( c001n03 -> c001n02 ) * Move ocf_192.168.100.183 ( c001n03 -> c001n02 ) * Move lsb_dummy ( c001n02 -> c001n08 ) * Move rsc_c001n01 ( c001n01 -> c001n03 ) * Stop child_DoFencing:0 ( c001n01 ) due to node availability - * Stop ocf_msdummy:0 ( Promoted c001n01 ) due to node availability - * Stop ocf_msdummy:4 ( Unpromoted c001n01 ) due to node availability + * Stop ocf_msdummy:0 ( Promoted c001n01 ) due to node availability + * Stop ocf_msdummy:4 ( Unpromoted c001n01 ) due to node availability Executing Cluster Transition: * Pseudo action: group-1_stop_0 * Resource action: ocf_192.168.100.183 stop on c001n03 * Resource action: lsb_dummy stop on c001n02 * Resource action: child_DoFencing:2 monitor on c001n08 * Resource action: child_DoFencing:2 monitor on c001n03 * Resource action: child_DoFencing:3 monitor on c001n03 * Resource action: child_DoFencing:3 monitor on c001n02 * Pseudo action: DoFencing_stop_0 * Resource action: ocf_msdummy:4 monitor on c001n08 * Resource action: ocf_msdummy:4 monitor on c001n03 * Resource action: ocf_msdummy:4 monitor on c001n02 * Resource action: ocf_msdummy:5 monitor on c001n08 * Resource action: ocf_msdummy:5 monitor on c001n02 * Resource action: ocf_msdummy:6 monitor on c001n08 * Resource action: ocf_msdummy:6 monitor on c001n03 * Resource action: ocf_msdummy:7 monitor on c001n03 * Resource action: ocf_msdummy:7 monitor on c001n02 * Pseudo action: master_rsc_1_demote_0 * Fencing c001n01 (reboot) * Pseudo action: DcIPaddr_stop_0 * Resource action: heartbeat_192.168.100.182 stop on c001n03 * Resource action: lsb_dummy start on c001n08 * Pseudo action: rsc_c001n01_stop_0 * Pseudo action: child_DoFencing:0_stop_0 * Pseudo action: DoFencing_stopped_0 * Pseudo action: ocf_msdummy:0_demote_0 * Pseudo action: master_rsc_1_demoted_0 * Pseudo action: master_rsc_1_stop_0 * Resource action: DcIPaddr start on c001n03 * Resource action: ocf_192.168.100.181 stop on c001n03 * Resource action: lsb_dummy monitor=5000 on c001n08 * Resource action: rsc_c001n01 start on c001n03 * Pseudo action: ocf_msdummy:0_stop_0 * Pseudo action: ocf_msdummy:4_stop_0 * Pseudo action: master_rsc_1_stopped_0 * Resource action: DcIPaddr monitor=5000 on c001n03 * Pseudo action: group-1_stopped_0 * Pseudo action: group-1_start_0 * Resource action: ocf_192.168.100.181 start on c001n02 * Resource action: heartbeat_192.168.100.182 start on c001n02 * Resource action: ocf_192.168.100.183 start on c001n02 * Resource action: rsc_c001n01 monitor=5000 on c001n03 * Pseudo action: group-1_running_0 * Resource action: ocf_192.168.100.181 monitor=5000 on c001n02 * Resource action: heartbeat_192.168.100.182 monitor=5000 on c001n02 * Resource action: ocf_192.168.100.183 monitor=5000 on c001n02 Revised Cluster Status: * Node List: * Online: [ c001n02 c001n03 c001n08 ] * OFFLINE: [ c001n01 ] * Full List of Resources: * DcIPaddr (ocf:heartbeat:IPaddr): Started c001n03 * Resource Group: group-1: * ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02 * heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02 * ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02 * lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n08 * rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n03 * rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08 * rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02 * rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03 * Clone Set: DoFencing [child_DoFencing] (unique): * child_DoFencing:0 (stonith:ssh): Stopped * child_DoFencing:1 (stonith:ssh): Started c001n03 * child_DoFencing:2 (stonith:ssh): Started c001n02 * child_DoFencing:3 (stonith:ssh): Started c001n08 * Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique): * ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped * ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03 * ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02 * ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08 * ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped * ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03 * ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02 * ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08 diff --git a/cts/scheduler/summary/promoted-asymmetrical-order.summary b/cts/scheduler/summary/promoted-asymmetrical-order.summary index df6e00c9c2..e10568e898 100644 --- a/cts/scheduler/summary/promoted-asymmetrical-order.summary +++ b/cts/scheduler/summary/promoted-asymmetrical-order.summary @@ -1,37 +1,37 @@ 2 of 4 resource instances DISABLED and 0 BLOCKED from further action due to failure Current cluster status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * Clone Set: ms1 [rsc1] (promotable) (disabled): * Promoted: [ node1 ] * Unpromoted: [ node2 ] * Clone Set: ms2 [rsc2] (promotable): * Promoted: [ node2 ] * Unpromoted: [ node1 ] Transition Summary: - * Stop rsc1:0 ( Promoted node1 ) due to node availability - * Stop rsc1:1 ( Unpromoted node2 ) due to node availability + * Stop rsc1:0 ( Promoted node1 ) due to node availability + * Stop rsc1:1 ( Unpromoted node2 ) due to node availability Executing Cluster Transition: * Pseudo action: ms1_demote_0 * Resource action: rsc1:0 demote on node1 * Pseudo action: ms1_demoted_0 * Pseudo action: ms1_stop_0 * Resource action: rsc1:0 stop on node1 * Resource action: rsc1:1 stop on node2 * Pseudo action: ms1_stopped_0 Revised Cluster Status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * Clone Set: ms1 [rsc1] (promotable) (disabled): * Stopped (disabled): [ node1 node2 ] * Clone Set: ms2 [rsc2] (promotable): * Promoted: [ node2 ] * Unpromoted: [ node1 ] diff --git a/cts/scheduler/summary/promoted-demote-2.summary b/cts/scheduler/summary/promoted-demote-2.summary index daea66ae8b..115da9aaaf 100644 --- a/cts/scheduler/summary/promoted-demote-2.summary +++ b/cts/scheduler/summary/promoted-demote-2.summary @@ -1,75 +1,75 @@ Current cluster status: * Node List: * Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started pcmk-1 * Resource Group: group-1: * r192.168.122.105 (ocf:heartbeat:IPaddr): Stopped * r192.168.122.106 (ocf:heartbeat:IPaddr): Stopped * r192.168.122.107 (ocf:heartbeat:IPaddr): Stopped * rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 * rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 * rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 * rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Stopped * migrator (ocf:pacemaker:Dummy): Started pcmk-4 * Clone Set: Connectivity [ping-1]: * Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Clone Set: master-1 [stateful-1] (promotable): * stateful-1 (ocf:pacemaker:Stateful): FAILED pcmk-1 * Unpromoted: [ pcmk-2 pcmk-3 pcmk-4 ] Transition Summary: * Start r192.168.122.105 ( pcmk-2 ) * Start r192.168.122.106 ( pcmk-2 ) * Start r192.168.122.107 ( pcmk-2 ) * Start lsb-dummy ( pcmk-2 ) - * Recover stateful-1:0 ( Unpromoted pcmk-1 ) + * Recover stateful-1:0 ( Unpromoted pcmk-1 ) * Promote stateful-1:1 ( Unpromoted -> Promoted pcmk-2 ) Executing Cluster Transition: * Resource action: stateful-1:0 cancel=15000 on pcmk-2 * Pseudo action: master-1_stop_0 * Resource action: stateful-1:1 stop on pcmk-1 * Pseudo action: master-1_stopped_0 * Pseudo action: master-1_start_0 * Resource action: stateful-1:1 start on pcmk-1 * Pseudo action: master-1_running_0 * Resource action: stateful-1:1 monitor=15000 on pcmk-1 * Pseudo action: master-1_promote_0 * Resource action: stateful-1:0 promote on pcmk-2 * Pseudo action: master-1_promoted_0 * Pseudo action: group-1_start_0 * Resource action: r192.168.122.105 start on pcmk-2 * Resource action: r192.168.122.106 start on pcmk-2 * Resource action: r192.168.122.107 start on pcmk-2 * Resource action: stateful-1:0 monitor=16000 on pcmk-2 * Pseudo action: group-1_running_0 * Resource action: r192.168.122.105 monitor=5000 on pcmk-2 * Resource action: r192.168.122.106 monitor=5000 on pcmk-2 * Resource action: r192.168.122.107 monitor=5000 on pcmk-2 * Resource action: lsb-dummy start on pcmk-2 * Resource action: lsb-dummy monitor=5000 on pcmk-2 Revised Cluster Status: * Node List: * Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started pcmk-1 * Resource Group: group-1: * r192.168.122.105 (ocf:heartbeat:IPaddr): Started pcmk-2 * r192.168.122.106 (ocf:heartbeat:IPaddr): Started pcmk-2 * r192.168.122.107 (ocf:heartbeat:IPaddr): Started pcmk-2 * rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1 * rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2 * rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3 * rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-2 * migrator (ocf:pacemaker:Dummy): Started pcmk-4 * Clone Set: Connectivity [ping-1]: * Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ pcmk-2 ] * Unpromoted: [ pcmk-1 pcmk-3 pcmk-4 ] diff --git a/cts/scheduler/summary/promoted-failed-demote-2.summary b/cts/scheduler/summary/promoted-failed-demote-2.summary index 198d9ad3ee..c8504e9e1d 100644 --- a/cts/scheduler/summary/promoted-failed-demote-2.summary +++ b/cts/scheduler/summary/promoted-failed-demote-2.summary @@ -1,47 +1,47 @@ Current cluster status: * Node List: * Online: [ dl380g5a dl380g5b ] * Full List of Resources: * Clone Set: ms-sf [group] (promotable) (unique): * Resource Group: group:0: * stateful-1:0 (ocf:heartbeat:Stateful): FAILED dl380g5b * stateful-2:0 (ocf:heartbeat:Stateful): Stopped * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a * stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a Transition Summary: - * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability + * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability * Promote stateful-1:1 ( Unpromoted -> Promoted dl380g5a ) * Promote stateful-2:1 ( Unpromoted -> Promoted dl380g5a ) Executing Cluster Transition: * Resource action: stateful-1:1 cancel=20000 on dl380g5a * Resource action: stateful-2:1 cancel=20000 on dl380g5a * Pseudo action: ms-sf_stop_0 * Pseudo action: group:0_stop_0 * Resource action: stateful-1:0 stop on dl380g5b * Pseudo action: group:0_stopped_0 * Pseudo action: ms-sf_stopped_0 * Pseudo action: ms-sf_promote_0 * Pseudo action: group:1_promote_0 * Resource action: stateful-1:1 promote on dl380g5a * Resource action: stateful-2:1 promote on dl380g5a * Pseudo action: group:1_promoted_0 * Resource action: stateful-1:1 monitor=10000 on dl380g5a * Resource action: stateful-2:1 monitor=10000 on dl380g5a * Pseudo action: ms-sf_promoted_0 Revised Cluster Status: * Node List: * Online: [ dl380g5a dl380g5b ] * Full List of Resources: * Clone Set: ms-sf [group] (promotable) (unique): * Resource Group: group:0: * stateful-1:0 (ocf:heartbeat:Stateful): Stopped * stateful-2:0 (ocf:heartbeat:Stateful): Stopped * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Promoted dl380g5a * stateful-2:1 (ocf:heartbeat:Stateful): Promoted dl380g5a diff --git a/cts/scheduler/summary/promoted-failed-demote.summary b/cts/scheduler/summary/promoted-failed-demote.summary index 884a380063..f071025528 100644 --- a/cts/scheduler/summary/promoted-failed-demote.summary +++ b/cts/scheduler/summary/promoted-failed-demote.summary @@ -1,64 +1,64 @@ Current cluster status: * Node List: * Online: [ dl380g5a dl380g5b ] * Full List of Resources: * Clone Set: ms-sf [group] (promotable) (unique): * Resource Group: group:0: * stateful-1:0 (ocf:heartbeat:Stateful): FAILED dl380g5b * stateful-2:0 (ocf:heartbeat:Stateful): Stopped * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a * stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a Transition Summary: - * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability + * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability * Promote stateful-1:1 ( Unpromoted -> Promoted dl380g5a ) * Promote stateful-2:1 ( Unpromoted -> Promoted dl380g5a ) Executing Cluster Transition: * Resource action: stateful-1:1 cancel=20000 on dl380g5a * Resource action: stateful-2:1 cancel=20000 on dl380g5a * Pseudo action: ms-sf_pre_notify_stop_0 * Resource action: stateful-1:0 notify on dl380g5b * Resource action: stateful-1:1 notify on dl380g5a * Resource action: stateful-2:1 notify on dl380g5a * Pseudo action: ms-sf_confirmed-pre_notify_stop_0 * Pseudo action: ms-sf_stop_0 * Pseudo action: group:0_stop_0 * Resource action: stateful-1:0 stop on dl380g5b * Pseudo action: group:0_stopped_0 * Pseudo action: ms-sf_stopped_0 * Pseudo action: ms-sf_post_notify_stopped_0 * Resource action: stateful-1:1 notify on dl380g5a * Resource action: stateful-2:1 notify on dl380g5a * Pseudo action: ms-sf_confirmed-post_notify_stopped_0 * Pseudo action: ms-sf_pre_notify_promote_0 * Resource action: stateful-1:1 notify on dl380g5a * Resource action: stateful-2:1 notify on dl380g5a * Pseudo action: ms-sf_confirmed-pre_notify_promote_0 * Pseudo action: ms-sf_promote_0 * Pseudo action: group:1_promote_0 * Resource action: stateful-1:1 promote on dl380g5a * Resource action: stateful-2:1 promote on dl380g5a * Pseudo action: group:1_promoted_0 * Pseudo action: ms-sf_promoted_0 * Pseudo action: ms-sf_post_notify_promoted_0 * Resource action: stateful-1:1 notify on dl380g5a * Resource action: stateful-2:1 notify on dl380g5a * Pseudo action: ms-sf_confirmed-post_notify_promoted_0 * Resource action: stateful-1:1 monitor=10000 on dl380g5a * Resource action: stateful-2:1 monitor=10000 on dl380g5a Revised Cluster Status: * Node List: * Online: [ dl380g5a dl380g5b ] * Full List of Resources: * Clone Set: ms-sf [group] (promotable) (unique): * Resource Group: group:0: * stateful-1:0 (ocf:heartbeat:Stateful): Stopped * stateful-2:0 (ocf:heartbeat:Stateful): Stopped * Resource Group: group:1: * stateful-1:1 (ocf:heartbeat:Stateful): Promoted dl380g5a * stateful-2:1 (ocf:heartbeat:Stateful): Promoted dl380g5a diff --git a/cts/scheduler/summary/remote-connection-unrecoverable.summary b/cts/scheduler/summary/remote-connection-unrecoverable.summary index bd1adfcfa4..3cfb64565a 100644 --- a/cts/scheduler/summary/remote-connection-unrecoverable.summary +++ b/cts/scheduler/summary/remote-connection-unrecoverable.summary @@ -1,54 +1,54 @@ Current cluster status: * Node List: * Node node1: UNCLEAN (offline) * Online: [ node2 ] * RemoteOnline: [ remote1 ] * Full List of Resources: * remote1 (ocf:pacemaker:remote): Started node1 (UNCLEAN) * killer (stonith:fence_xvm): Started node2 * rsc1 (ocf:pacemaker:Dummy): Started remote1 * Clone Set: rsc2-master [rsc2] (promotable): * rsc2 (ocf:pacemaker:Stateful): Promoted node1 (UNCLEAN) * Promoted: [ node2 ] * Stopped: [ remote1 ] Transition Summary: * Fence (reboot) remote1 'resources are active and the connection is unrecoverable' * Fence (reboot) node1 'peer is no longer part of the cluster' * Stop remote1 ( node1 ) due to node availability * Restart killer ( node2 ) due to resource definition change * Move rsc1 ( remote1 -> node2 ) - * Stop rsc2:0 ( Promoted node1 ) due to node availability + * Stop rsc2:0 ( Promoted node1 ) due to node availability Executing Cluster Transition: * Pseudo action: remote1_stop_0 * Resource action: killer stop on node2 * Resource action: rsc1 monitor on node2 * Fencing node1 (reboot) * Fencing remote1 (reboot) * Resource action: killer start on node2 * Resource action: killer monitor=60000 on node2 * Pseudo action: rsc1_stop_0 * Pseudo action: rsc2-master_demote_0 * Resource action: rsc1 start on node2 * Pseudo action: rsc2_demote_0 * Pseudo action: rsc2-master_demoted_0 * Pseudo action: rsc2-master_stop_0 * Resource action: rsc1 monitor=10000 on node2 * Pseudo action: rsc2_stop_0 * Pseudo action: rsc2-master_stopped_0 Revised Cluster Status: * Node List: * Online: [ node2 ] * OFFLINE: [ node1 ] * RemoteOFFLINE: [ remote1 ] * Full List of Resources: * remote1 (ocf:pacemaker:remote): Stopped * killer (stonith:fence_xvm): Started node2 * rsc1 (ocf:pacemaker:Dummy): Started node2 * Clone Set: rsc2-master [rsc2] (promotable): * Promoted: [ node2 ] * Stopped: [ node1 remote1 ] diff --git a/cts/scheduler/summary/remote-recover-all.summary b/cts/scheduler/summary/remote-recover-all.summary index 176c1de8b3..18d10730bf 100644 --- a/cts/scheduler/summary/remote-recover-all.summary +++ b/cts/scheduler/summary/remote-recover-all.summary @@ -1,146 +1,146 @@ Using the original execution date of: 2017-05-03 13:33:24Z Current cluster status: * Node List: * Node controller-1: UNCLEAN (offline) * Online: [ controller-0 controller-2 ] * RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-1 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 galera-2 ] * Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN) * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * Clone Set: haproxy-clone [haproxy]: * haproxy (systemd:haproxy): Started controller-1 (UNCLEAN) * Started: [ controller-0 controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN) Transition Summary: * Fence (reboot) messaging-1 'resources are active and the connection is unrecoverable' * Fence (reboot) galera-2 'resources are active and the connection is unrecoverable' * Fence (reboot) controller-1 'peer is no longer part of the cluster' * Stop messaging-1 ( controller-1 ) due to node availability * Move galera-0 ( controller-1 -> controller-2 ) * Stop galera-2 ( controller-1 ) due to node availability * Stop rabbitmq:2 ( messaging-1 ) due to node availability - * Stop galera:1 ( Promoted galera-2 ) due to node availability - * Stop redis:0 ( Unpromoted controller-1 ) due to node availability + * Stop galera:1 ( Promoted galera-2 ) due to node availability + * Stop redis:0 ( Unpromoted controller-1 ) due to node availability * Move ip-172.17.1.14 ( controller-1 -> controller-2 ) * Move ip-172.17.1.17 ( controller-1 -> controller-2 ) * Move ip-172.17.4.11 ( controller-1 -> controller-2 ) * Stop haproxy:0 ( controller-1 ) due to node availability * Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 ) Executing Cluster Transition: * Pseudo action: messaging-1_stop_0 * Pseudo action: galera-0_stop_0 * Pseudo action: galera-2_stop_0 * Pseudo action: galera-master_demote_0 * Pseudo action: redis-master_pre_notify_stop_0 * Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0 * Fencing controller-1 (reboot) * Pseudo action: redis_post_notify_stop_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-pre_notify_stop_0 * Pseudo action: redis-master_stop_0 * Pseudo action: haproxy-clone_stop_0 * Fencing galera-2 (reboot) * Pseudo action: galera_demote_0 * Pseudo action: galera-master_demoted_0 * Pseudo action: galera-master_stop_0 * Pseudo action: redis_stop_0 * Pseudo action: redis-master_stopped_0 * Pseudo action: haproxy_stop_0 * Pseudo action: haproxy-clone_stopped_0 * Fencing messaging-1 (reboot) * Resource action: galera-0 start on controller-2 * Pseudo action: rabbitmq_post_notify_stop_0 * Pseudo action: rabbitmq-clone_stop_0 * Pseudo action: galera_stop_0 * Resource action: galera monitor=10000 on galera-0 * Pseudo action: galera-master_stopped_0 * Pseudo action: redis-master_post_notify_stopped_0 * Pseudo action: ip-172.17.1.14_stop_0 * Pseudo action: ip-172.17.1.17_stop_0 * Pseudo action: ip-172.17.4.11_stop_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2 * Resource action: galera-0 monitor=20000 on controller-2 * Resource action: rabbitmq notify on messaging-2 * Resource action: rabbitmq notify on messaging-0 * Pseudo action: rabbitmq_notified_0 * Pseudo action: rabbitmq_stop_0 * Pseudo action: rabbitmq-clone_stopped_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-post_notify_stopped_0 * Resource action: ip-172.17.1.14 start on controller-2 * Resource action: ip-172.17.1.17 start on controller-2 * Resource action: ip-172.17.4.11 start on controller-2 * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2 * Pseudo action: redis_notified_0 * Resource action: ip-172.17.1.14 monitor=10000 on controller-2 * Resource action: ip-172.17.1.17 monitor=10000 on controller-2 * Resource action: ip-172.17.4.11 monitor=10000 on controller-2 Using the original execution date of: 2017-05-03 13:33:24Z Revised Cluster Status: * Node List: * Online: [ controller-0 controller-2 ] * OFFLINE: [ controller-1 ] * RemoteOnline: [ galera-0 galera-1 messaging-0 messaging-2 ] * RemoteOFFLINE: [ galera-2 messaging-1 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Stopped * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-2 * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Stopped * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 messaging-1 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 ] * Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2 * Clone Set: haproxy-clone [haproxy]: * Started: [ controller-0 controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2 diff --git a/cts/scheduler/summary/remote-recover-connection.summary b/cts/scheduler/summary/remote-recover-connection.summary index fd6900dd96..a9723bc5e1 100644 --- a/cts/scheduler/summary/remote-recover-connection.summary +++ b/cts/scheduler/summary/remote-recover-connection.summary @@ -1,132 +1,132 @@ Using the original execution date of: 2017-05-03 13:33:24Z Current cluster status: * Node List: * Node controller-1: UNCLEAN (offline) * Online: [ controller-0 controller-2 ] * RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-1 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 galera-2 ] * Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN) * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * Clone Set: haproxy-clone [haproxy]: * haproxy (systemd:haproxy): Started controller-1 (UNCLEAN) * Started: [ controller-0 controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN) Transition Summary: * Fence (reboot) controller-1 'peer is no longer part of the cluster' * Move messaging-1 ( controller-1 -> controller-2 ) * Move galera-0 ( controller-1 -> controller-2 ) * Move galera-2 ( controller-1 -> controller-2 ) - * Stop redis:0 ( Unpromoted controller-1 ) due to node availability + * Stop redis:0 ( Unpromoted controller-1 ) due to node availability * Move ip-172.17.1.14 ( controller-1 -> controller-2 ) * Move ip-172.17.1.17 ( controller-1 -> controller-2 ) * Move ip-172.17.4.11 ( controller-1 -> controller-2 ) * Stop haproxy:0 ( controller-1 ) due to node availability * Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 ) Executing Cluster Transition: * Pseudo action: messaging-1_stop_0 * Pseudo action: galera-0_stop_0 * Pseudo action: galera-2_stop_0 * Pseudo action: redis-master_pre_notify_stop_0 * Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0 * Fencing controller-1 (reboot) * Resource action: messaging-1 start on controller-2 * Resource action: galera-0 start on controller-2 * Resource action: galera-2 start on controller-2 * Resource action: rabbitmq monitor=10000 on messaging-1 * Resource action: galera monitor=10000 on galera-2 * Resource action: galera monitor=10000 on galera-0 * Pseudo action: redis_post_notify_stop_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-pre_notify_stop_0 * Pseudo action: redis-master_stop_0 * Pseudo action: haproxy-clone_stop_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2 * Resource action: messaging-1 monitor=20000 on controller-2 * Resource action: galera-0 monitor=20000 on controller-2 * Resource action: galera-2 monitor=20000 on controller-2 * Pseudo action: redis_stop_0 * Pseudo action: redis-master_stopped_0 * Pseudo action: haproxy_stop_0 * Pseudo action: haproxy-clone_stopped_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2 * Pseudo action: redis-master_post_notify_stopped_0 * Pseudo action: ip-172.17.1.14_stop_0 * Pseudo action: ip-172.17.1.17_stop_0 * Pseudo action: ip-172.17.4.11_stop_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-post_notify_stopped_0 * Resource action: ip-172.17.1.14 start on controller-2 * Resource action: ip-172.17.1.17 start on controller-2 * Resource action: ip-172.17.4.11 start on controller-2 * Pseudo action: redis_notified_0 * Resource action: ip-172.17.1.14 monitor=10000 on controller-2 * Resource action: ip-172.17.1.17 monitor=10000 on controller-2 * Resource action: ip-172.17.4.11 monitor=10000 on controller-2 Using the original execution date of: 2017-05-03 13:33:24Z Revised Cluster Status: * Node List: * Online: [ controller-0 controller-2 ] * OFFLINE: [ controller-1 ] * RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Started controller-2 * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-2 * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Started controller-2 * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-1 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 galera-2 ] * Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2 * Clone Set: haproxy-clone [haproxy]: * Started: [ controller-0 controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2 diff --git a/cts/scheduler/summary/remote-recover-no-resources.summary b/cts/scheduler/summary/remote-recover-no-resources.summary index 332d1c4123..d7d9ef942c 100644 --- a/cts/scheduler/summary/remote-recover-no-resources.summary +++ b/cts/scheduler/summary/remote-recover-no-resources.summary @@ -1,137 +1,137 @@ Using the original execution date of: 2017-05-03 13:33:24Z Current cluster status: * Node List: * Node controller-1: UNCLEAN (offline) * Online: [ controller-0 controller-2 ] * RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-1 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 ] * Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN) * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * Clone Set: haproxy-clone [haproxy]: * haproxy (systemd:haproxy): Started controller-1 (UNCLEAN) * Started: [ controller-0 controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN) Transition Summary: * Fence (reboot) messaging-1 'resources are active and the connection is unrecoverable' * Fence (reboot) controller-1 'peer is no longer part of the cluster' * Stop messaging-1 ( controller-1 ) due to node availability * Move galera-0 ( controller-1 -> controller-2 ) * Stop galera-2 ( controller-1 ) due to node availability * Stop rabbitmq:2 ( messaging-1 ) due to node availability - * Stop redis:0 ( Unpromoted controller-1 ) due to node availability + * Stop redis:0 ( Unpromoted controller-1 ) due to node availability * Move ip-172.17.1.14 ( controller-1 -> controller-2 ) * Move ip-172.17.1.17 ( controller-1 -> controller-2 ) * Move ip-172.17.4.11 ( controller-1 -> controller-2 ) * Stop haproxy:0 ( controller-1 ) due to node availability * Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 ) Executing Cluster Transition: * Pseudo action: messaging-1_stop_0 * Pseudo action: galera-0_stop_0 * Pseudo action: galera-2_stop_0 * Pseudo action: redis-master_pre_notify_stop_0 * Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0 * Fencing controller-1 (reboot) * Pseudo action: redis_post_notify_stop_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-pre_notify_stop_0 * Pseudo action: redis-master_stop_0 * Pseudo action: haproxy-clone_stop_0 * Fencing messaging-1 (reboot) * Resource action: galera-0 start on controller-2 * Pseudo action: rabbitmq_post_notify_stop_0 * Pseudo action: rabbitmq-clone_stop_0 * Resource action: galera monitor=10000 on galera-0 * Pseudo action: redis_stop_0 * Pseudo action: redis-master_stopped_0 * Pseudo action: haproxy_stop_0 * Pseudo action: haproxy-clone_stopped_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2 * Resource action: galera-0 monitor=20000 on controller-2 * Resource action: rabbitmq notify on messaging-2 * Resource action: rabbitmq notify on messaging-0 * Pseudo action: rabbitmq_notified_0 * Pseudo action: rabbitmq_stop_0 * Pseudo action: rabbitmq-clone_stopped_0 * Pseudo action: redis-master_post_notify_stopped_0 * Pseudo action: ip-172.17.1.14_stop_0 * Pseudo action: ip-172.17.1.17_stop_0 * Pseudo action: ip-172.17.4.11_stop_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-post_notify_stopped_0 * Resource action: ip-172.17.1.14 start on controller-2 * Resource action: ip-172.17.1.17 start on controller-2 * Resource action: ip-172.17.4.11 start on controller-2 * Pseudo action: redis_notified_0 * Resource action: ip-172.17.1.14 monitor=10000 on controller-2 * Resource action: ip-172.17.1.17 monitor=10000 on controller-2 * Resource action: ip-172.17.4.11 monitor=10000 on controller-2 Using the original execution date of: 2017-05-03 13:33:24Z Revised Cluster Status: * Node List: * Online: [ controller-0 controller-2 ] * OFFLINE: [ controller-1 ] * RemoteOnline: [ galera-0 galera-1 messaging-0 messaging-2 ] * RemoteOFFLINE: [ galera-2 messaging-1 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Stopped * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-2 * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Stopped * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 messaging-1 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 ] * Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2 * Clone Set: haproxy-clone [haproxy]: * Started: [ controller-0 controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2 diff --git a/cts/scheduler/summary/remote-recover-unknown.summary b/cts/scheduler/summary/remote-recover-unknown.summary index ac5143a16e..4f3d045284 100644 --- a/cts/scheduler/summary/remote-recover-unknown.summary +++ b/cts/scheduler/summary/remote-recover-unknown.summary @@ -1,139 +1,139 @@ Using the original execution date of: 2017-05-03 13:33:24Z Current cluster status: * Node List: * Node controller-1: UNCLEAN (offline) * Online: [ controller-0 controller-2 ] * RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-1 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 ] * Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN) * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * Clone Set: haproxy-clone [haproxy]: * haproxy (systemd:haproxy): Started controller-1 (UNCLEAN) * Started: [ controller-0 controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN) Transition Summary: * Fence (reboot) galera-2 'resources are in an unknown state and the connection is unrecoverable' * Fence (reboot) messaging-1 'resources are active and the connection is unrecoverable' * Fence (reboot) controller-1 'peer is no longer part of the cluster' * Stop messaging-1 ( controller-1 ) due to node availability * Move galera-0 ( controller-1 -> controller-2 ) * Stop galera-2 ( controller-1 ) due to node availability * Stop rabbitmq:2 ( messaging-1 ) due to node availability - * Stop redis:0 ( Unpromoted controller-1 ) due to node availability + * Stop redis:0 ( Unpromoted controller-1 ) due to node availability * Move ip-172.17.1.14 ( controller-1 -> controller-2 ) * Move ip-172.17.1.17 ( controller-1 -> controller-2 ) * Move ip-172.17.4.11 ( controller-1 -> controller-2 ) * Stop haproxy:0 ( controller-1 ) due to node availability * Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 ) Executing Cluster Transition: * Pseudo action: messaging-1_stop_0 * Pseudo action: galera-0_stop_0 * Pseudo action: galera-2_stop_0 * Pseudo action: redis-master_pre_notify_stop_0 * Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0 * Fencing controller-1 (reboot) * Pseudo action: redis_post_notify_stop_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-pre_notify_stop_0 * Pseudo action: redis-master_stop_0 * Pseudo action: haproxy-clone_stop_0 * Fencing galera-2 (reboot) * Fencing messaging-1 (reboot) * Resource action: galera-0 start on controller-2 * Pseudo action: rabbitmq_post_notify_stop_0 * Pseudo action: rabbitmq-clone_stop_0 * Resource action: galera monitor=10000 on galera-0 * Pseudo action: redis_stop_0 * Pseudo action: redis-master_stopped_0 * Pseudo action: haproxy_stop_0 * Pseudo action: haproxy-clone_stopped_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2 * Resource action: galera-0 monitor=20000 on controller-2 * Resource action: rabbitmq notify on messaging-2 * Resource action: rabbitmq notify on messaging-0 * Pseudo action: rabbitmq_notified_0 * Pseudo action: rabbitmq_stop_0 * Pseudo action: rabbitmq-clone_stopped_0 * Pseudo action: redis-master_post_notify_stopped_0 * Pseudo action: ip-172.17.1.14_stop_0 * Pseudo action: ip-172.17.1.17_stop_0 * Pseudo action: ip-172.17.4.11_stop_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-post_notify_stopped_0 * Resource action: ip-172.17.1.14 start on controller-2 * Resource action: ip-172.17.1.17 start on controller-2 * Resource action: ip-172.17.4.11 start on controller-2 * Pseudo action: redis_notified_0 * Resource action: ip-172.17.1.14 monitor=10000 on controller-2 * Resource action: ip-172.17.1.17 monitor=10000 on controller-2 * Resource action: ip-172.17.4.11 monitor=10000 on controller-2 Using the original execution date of: 2017-05-03 13:33:24Z Revised Cluster Status: * Node List: * Online: [ controller-0 controller-2 ] * OFFLINE: [ controller-1 ] * RemoteOnline: [ galera-0 galera-1 messaging-0 messaging-2 ] * RemoteOFFLINE: [ galera-2 messaging-1 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Stopped * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-2 * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Stopped * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 messaging-1 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 ] * Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2 * Clone Set: haproxy-clone [haproxy]: * Started: [ controller-0 controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2 diff --git a/cts/scheduler/summary/remote-recovery.summary b/cts/scheduler/summary/remote-recovery.summary index fd6900dd96..a9723bc5e1 100644 --- a/cts/scheduler/summary/remote-recovery.summary +++ b/cts/scheduler/summary/remote-recovery.summary @@ -1,132 +1,132 @@ Using the original execution date of: 2017-05-03 13:33:24Z Current cluster status: * Node List: * Node controller-1: UNCLEAN (offline) * Online: [ controller-0 controller-2 ] * RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN) * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-1 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 galera-2 ] * Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN) * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN) * Clone Set: haproxy-clone [haproxy]: * haproxy (systemd:haproxy): Started controller-1 (UNCLEAN) * Started: [ controller-0 controller-2 ] * Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN) Transition Summary: * Fence (reboot) controller-1 'peer is no longer part of the cluster' * Move messaging-1 ( controller-1 -> controller-2 ) * Move galera-0 ( controller-1 -> controller-2 ) * Move galera-2 ( controller-1 -> controller-2 ) - * Stop redis:0 ( Unpromoted controller-1 ) due to node availability + * Stop redis:0 ( Unpromoted controller-1 ) due to node availability * Move ip-172.17.1.14 ( controller-1 -> controller-2 ) * Move ip-172.17.1.17 ( controller-1 -> controller-2 ) * Move ip-172.17.4.11 ( controller-1 -> controller-2 ) * Stop haproxy:0 ( controller-1 ) due to node availability * Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 ) Executing Cluster Transition: * Pseudo action: messaging-1_stop_0 * Pseudo action: galera-0_stop_0 * Pseudo action: galera-2_stop_0 * Pseudo action: redis-master_pre_notify_stop_0 * Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0 * Fencing controller-1 (reboot) * Resource action: messaging-1 start on controller-2 * Resource action: galera-0 start on controller-2 * Resource action: galera-2 start on controller-2 * Resource action: rabbitmq monitor=10000 on messaging-1 * Resource action: galera monitor=10000 on galera-2 * Resource action: galera monitor=10000 on galera-0 * Pseudo action: redis_post_notify_stop_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-pre_notify_stop_0 * Pseudo action: redis-master_stop_0 * Pseudo action: haproxy-clone_stop_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2 * Resource action: messaging-1 monitor=20000 on controller-2 * Resource action: galera-0 monitor=20000 on controller-2 * Resource action: galera-2 monitor=20000 on controller-2 * Pseudo action: redis_stop_0 * Pseudo action: redis-master_stopped_0 * Pseudo action: haproxy_stop_0 * Pseudo action: haproxy-clone_stopped_0 * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2 * Pseudo action: redis-master_post_notify_stopped_0 * Pseudo action: ip-172.17.1.14_stop_0 * Pseudo action: ip-172.17.1.17_stop_0 * Pseudo action: ip-172.17.4.11_stop_0 * Resource action: redis notify on controller-0 * Resource action: redis notify on controller-2 * Pseudo action: redis-master_confirmed-post_notify_stopped_0 * Resource action: ip-172.17.1.14 start on controller-2 * Resource action: ip-172.17.1.17 start on controller-2 * Resource action: ip-172.17.4.11 start on controller-2 * Pseudo action: redis_notified_0 * Resource action: ip-172.17.1.14 monitor=10000 on controller-2 * Resource action: ip-172.17.1.17 monitor=10000 on controller-2 * Resource action: ip-172.17.4.11 monitor=10000 on controller-2 Using the original execution date of: 2017-05-03 13:33:24Z Revised Cluster Status: * Node List: * Online: [ controller-0 controller-2 ] * OFFLINE: [ controller-1 ] * RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * Full List of Resources: * messaging-0 (ocf:pacemaker:remote): Started controller-0 * messaging-1 (ocf:pacemaker:remote): Started controller-2 * messaging-2 (ocf:pacemaker:remote): Started controller-0 * galera-0 (ocf:pacemaker:remote): Started controller-2 * galera-1 (ocf:pacemaker:remote): Started controller-0 * galera-2 (ocf:pacemaker:remote): Started controller-2 * Clone Set: rabbitmq-clone [rabbitmq]: * Started: [ messaging-0 messaging-1 messaging-2 ] * Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ] * Clone Set: galera-master [galera] (promotable): * Promoted: [ galera-0 galera-1 galera-2 ] * Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ] * Clone Set: redis-master [redis] (promotable): * Promoted: [ controller-0 ] * Unpromoted: [ controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2 * ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0 * ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2 * Clone Set: haproxy-clone [haproxy]: * Started: [ controller-0 controller-2 ] * Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ] * openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0 * stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0 * stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2 diff --git a/cts/scheduler/summary/rsc-sets-promoted.summary b/cts/scheduler/summary/rsc-sets-promoted.summary index a45e4b16e8..3db15881a0 100644 --- a/cts/scheduler/summary/rsc-sets-promoted.summary +++ b/cts/scheduler/summary/rsc-sets-promoted.summary @@ -1,49 +1,49 @@ Current cluster status: * Node List: * Node node1: standby (with active resources) * Online: [ node2 ] * Full List of Resources: * Clone Set: ms-rsc [rsc] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] * rsc1 (ocf:pacemaker:Dummy): Started node1 * rsc2 (ocf:pacemaker:Dummy): Started node1 * rsc3 (ocf:pacemaker:Dummy): Started node1 Transition Summary: - * Stop rsc:0 ( Promoted node1 ) due to node availability + * Stop rsc:0 ( Promoted node1 ) due to node availability * Promote rsc:1 ( Unpromoted -> Promoted node2 ) * Move rsc1 ( node1 -> node2 ) * Move rsc2 ( node1 -> node2 ) * Move rsc3 ( node1 -> node2 ) Executing Cluster Transition: * Resource action: rsc1 stop on node1 * Resource action: rsc2 stop on node1 * Resource action: rsc3 stop on node1 * Pseudo action: ms-rsc_demote_0 * Resource action: rsc:0 demote on node1 * Pseudo action: ms-rsc_demoted_0 * Pseudo action: ms-rsc_stop_0 * Resource action: rsc:0 stop on node1 * Pseudo action: ms-rsc_stopped_0 * Pseudo action: ms-rsc_promote_0 * Resource action: rsc:1 promote on node2 * Pseudo action: ms-rsc_promoted_0 * Resource action: rsc1 start on node2 * Resource action: rsc2 start on node2 * Resource action: rsc3 start on node2 Revised Cluster Status: * Node List: * Node node1: standby * Online: [ node2 ] * Full List of Resources: * Clone Set: ms-rsc [rsc] (promotable): * Promoted: [ node2 ] * Stopped: [ node1 ] * rsc1 (ocf:pacemaker:Dummy): Started node2 * rsc2 (ocf:pacemaker:Dummy): Started node2 * rsc3 (ocf:pacemaker:Dummy): Started node2 diff --git a/cts/scheduler/summary/ticket-promoted-14.summary b/cts/scheduler/summary/ticket-promoted-14.summary index ee8912b2e9..80ff84346b 100644 --- a/cts/scheduler/summary/ticket-promoted-14.summary +++ b/cts/scheduler/summary/ticket-promoted-14.summary @@ -1,31 +1,31 @@ Current cluster status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] Transition Summary: - * Stop rsc1:0 ( Promoted node1 ) due to node availability - * Stop rsc1:1 ( Unpromoted node2 ) due to node availability + * Stop rsc1:0 ( Promoted node1 ) due to node availability + * Stop rsc1:1 ( Unpromoted node2 ) due to node availability Executing Cluster Transition: * Pseudo action: ms1_demote_0 * Resource action: rsc1:1 demote on node1 * Pseudo action: ms1_demoted_0 * Pseudo action: ms1_stop_0 * Resource action: rsc1:1 stop on node1 * Resource action: rsc1:0 stop on node2 * Pseudo action: ms1_stopped_0 Revised Cluster Status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] diff --git a/cts/scheduler/summary/ticket-promoted-15.summary b/cts/scheduler/summary/ticket-promoted-15.summary index ee8912b2e9..80ff84346b 100644 --- a/cts/scheduler/summary/ticket-promoted-15.summary +++ b/cts/scheduler/summary/ticket-promoted-15.summary @@ -1,31 +1,31 @@ Current cluster status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] Transition Summary: - * Stop rsc1:0 ( Promoted node1 ) due to node availability - * Stop rsc1:1 ( Unpromoted node2 ) due to node availability + * Stop rsc1:0 ( Promoted node1 ) due to node availability + * Stop rsc1:1 ( Unpromoted node2 ) due to node availability Executing Cluster Transition: * Pseudo action: ms1_demote_0 * Resource action: rsc1:1 demote on node1 * Pseudo action: ms1_demoted_0 * Pseudo action: ms1_stop_0 * Resource action: rsc1:1 stop on node1 * Resource action: rsc1:0 stop on node2 * Pseudo action: ms1_stopped_0 Revised Cluster Status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] diff --git a/cts/scheduler/summary/ticket-promoted-21.summary b/cts/scheduler/summary/ticket-promoted-21.summary index f116a2eea0..788573facb 100644 --- a/cts/scheduler/summary/ticket-promoted-21.summary +++ b/cts/scheduler/summary/ticket-promoted-21.summary @@ -1,36 +1,36 @@ Current cluster status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] Transition Summary: * Fence (reboot) node1 'deadman ticket was lost' * Move rsc_stonith ( node1 -> node2 ) - * Stop rsc1:0 ( Promoted node1 ) due to node availability + * Stop rsc1:0 ( Promoted node1 ) due to node availability Executing Cluster Transition: * Pseudo action: rsc_stonith_stop_0 * Pseudo action: ms1_demote_0 * Fencing node1 (reboot) * Resource action: rsc_stonith start on node2 * Pseudo action: rsc1:1_demote_0 * Pseudo action: ms1_demoted_0 * Pseudo action: ms1_stop_0 * Pseudo action: rsc1:1_stop_0 * Pseudo action: ms1_stopped_0 Revised Cluster Status: * Node List: * Online: [ node2 ] * OFFLINE: [ node1 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node2 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node2 ] * Stopped: [ node1 ] diff --git a/cts/scheduler/summary/ticket-promoted-3.summary b/cts/scheduler/summary/ticket-promoted-3.summary index ee8912b2e9..80ff84346b 100644 --- a/cts/scheduler/summary/ticket-promoted-3.summary +++ b/cts/scheduler/summary/ticket-promoted-3.summary @@ -1,31 +1,31 @@ Current cluster status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] Transition Summary: - * Stop rsc1:0 ( Promoted node1 ) due to node availability - * Stop rsc1:1 ( Unpromoted node2 ) due to node availability + * Stop rsc1:0 ( Promoted node1 ) due to node availability + * Stop rsc1:1 ( Unpromoted node2 ) due to node availability Executing Cluster Transition: * Pseudo action: ms1_demote_0 * Resource action: rsc1:1 demote on node1 * Pseudo action: ms1_demoted_0 * Pseudo action: ms1_stop_0 * Resource action: rsc1:1 stop on node1 * Resource action: rsc1:0 stop on node2 * Pseudo action: ms1_stopped_0 Revised Cluster Status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Stopped: [ node1 node2 ] diff --git a/cts/scheduler/summary/ticket-promoted-9.summary b/cts/scheduler/summary/ticket-promoted-9.summary index f116a2eea0..788573facb 100644 --- a/cts/scheduler/summary/ticket-promoted-9.summary +++ b/cts/scheduler/summary/ticket-promoted-9.summary @@ -1,36 +1,36 @@ Current cluster status: * Node List: * Online: [ node1 node2 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node1 * Clone Set: ms1 [rsc1] (promotable): * Promoted: [ node1 ] * Unpromoted: [ node2 ] Transition Summary: * Fence (reboot) node1 'deadman ticket was lost' * Move rsc_stonith ( node1 -> node2 ) - * Stop rsc1:0 ( Promoted node1 ) due to node availability + * Stop rsc1:0 ( Promoted node1 ) due to node availability Executing Cluster Transition: * Pseudo action: rsc_stonith_stop_0 * Pseudo action: ms1_demote_0 * Fencing node1 (reboot) * Resource action: rsc_stonith start on node2 * Pseudo action: rsc1:1_demote_0 * Pseudo action: ms1_demoted_0 * Pseudo action: ms1_stop_0 * Pseudo action: rsc1:1_stop_0 * Pseudo action: ms1_stopped_0 Revised Cluster Status: * Node List: * Online: [ node2 ] * OFFLINE: [ node1 ] * Full List of Resources: * rsc_stonith (stonith:null): Started node2 * Clone Set: ms1 [rsc1] (promotable): * Unpromoted: [ node2 ] * Stopped: [ node1 ] diff --git a/cts/scheduler/summary/whitebox-ms-ordering-move.summary b/cts/scheduler/summary/whitebox-ms-ordering-move.summary index 6a5fb6eaeb..c9b13e032d 100644 --- a/cts/scheduler/summary/whitebox-ms-ordering-move.summary +++ b/cts/scheduler/summary/whitebox-ms-ordering-move.summary @@ -1,107 +1,107 @@ Current cluster status: * Node List: * Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * GuestOnline: [ lxc1@rhel7-1 lxc2@rhel7-1 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-3 * FencingPass (stonith:fence_dummy): Started rhel7-4 * FencingFail (stonith:fence_dummy): Started rhel7-5 * rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-1 * rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Started rhel7-2 * rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Started rhel7-3 * rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Started rhel7-4 * rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Started rhel7-5 * migrator (ocf:pacemaker:Dummy): Started rhel7-4 * Clone Set: Connectivity [ping-1]: * Started: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * Stopped: [ lxc1 lxc2 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ rhel7-3 ] * Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 rhel7-5 ] * Resource Group: group-1: * r192.168.122.207 (ocf:heartbeat:IPaddr2): Started rhel7-3 * petulant (service:DummySD): Started rhel7-3 * r192.168.122.208 (ocf:heartbeat:IPaddr2): Started rhel7-3 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started rhel7-3 * container1 (ocf:heartbeat:VirtualDomain): Started rhel7-1 * container2 (ocf:heartbeat:VirtualDomain): Started rhel7-1 * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Promoted: [ lxc1 ] * Unpromoted: [ lxc2 ] Transition Summary: * Move container1 ( rhel7-1 -> rhel7-2 ) - * Restart lxc-ms:0 ( Promoted lxc1 ) due to required container1 start + * Restart lxc-ms:0 ( Promoted lxc1 ) due to required container1 start * Move lxc1 ( rhel7-1 -> rhel7-2 ) Executing Cluster Transition: * Resource action: rsc_rhel7-1 monitor on lxc2 * Resource action: rsc_rhel7-2 monitor on lxc2 * Resource action: rsc_rhel7-3 monitor on lxc2 * Resource action: rsc_rhel7-4 monitor on lxc2 * Resource action: rsc_rhel7-5 monitor on lxc2 * Resource action: migrator monitor on lxc2 * Resource action: ping-1 monitor on lxc2 * Resource action: stateful-1 monitor on lxc2 * Resource action: r192.168.122.207 monitor on lxc2 * Resource action: petulant monitor on lxc2 * Resource action: r192.168.122.208 monitor on lxc2 * Resource action: lsb-dummy monitor on lxc2 * Pseudo action: lxc-ms-master_demote_0 * Resource action: lxc1 monitor on rhel7-5 * Resource action: lxc1 monitor on rhel7-4 * Resource action: lxc1 monitor on rhel7-3 * Resource action: lxc1 monitor on rhel7-2 * Resource action: lxc2 monitor on rhel7-5 * Resource action: lxc2 monitor on rhel7-4 * Resource action: lxc2 monitor on rhel7-3 * Resource action: lxc2 monitor on rhel7-2 * Resource action: lxc-ms demote on lxc1 * Pseudo action: lxc-ms-master_demoted_0 * Pseudo action: lxc-ms-master_stop_0 * Resource action: lxc-ms stop on lxc1 * Pseudo action: lxc-ms-master_stopped_0 * Pseudo action: lxc-ms-master_start_0 * Resource action: lxc1 stop on rhel7-1 * Resource action: container1 stop on rhel7-1 * Resource action: container1 start on rhel7-2 * Resource action: lxc1 start on rhel7-2 * Resource action: lxc-ms start on lxc1 * Pseudo action: lxc-ms-master_running_0 * Resource action: lxc1 monitor=30000 on rhel7-2 * Pseudo action: lxc-ms-master_promote_0 * Resource action: lxc-ms promote on lxc1 * Pseudo action: lxc-ms-master_promoted_0 Revised Cluster Status: * Node List: * Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * GuestOnline: [ lxc1@rhel7-2 lxc2@rhel7-1 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started rhel7-3 * FencingPass (stonith:fence_dummy): Started rhel7-4 * FencingFail (stonith:fence_dummy): Started rhel7-5 * rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-1 * rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Started rhel7-2 * rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Started rhel7-3 * rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Started rhel7-4 * rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Started rhel7-5 * migrator (ocf:pacemaker:Dummy): Started rhel7-4 * Clone Set: Connectivity [ping-1]: * Started: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ] * Stopped: [ lxc1 lxc2 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ rhel7-3 ] * Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 rhel7-5 ] * Resource Group: group-1: * r192.168.122.207 (ocf:heartbeat:IPaddr2): Started rhel7-3 * petulant (service:DummySD): Started rhel7-3 * r192.168.122.208 (ocf:heartbeat:IPaddr2): Started rhel7-3 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started rhel7-3 * container1 (ocf:heartbeat:VirtualDomain): Started rhel7-2 * container2 (ocf:heartbeat:VirtualDomain): Started rhel7-1 * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Promoted: [ lxc1 ] * Unpromoted: [ lxc2 ] diff --git a/cts/scheduler/summary/whitebox-ms-ordering.summary b/cts/scheduler/summary/whitebox-ms-ordering.summary index 921f6d068d..4d23221fa6 100644 --- a/cts/scheduler/summary/whitebox-ms-ordering.summary +++ b/cts/scheduler/summary/whitebox-ms-ordering.summary @@ -1,73 +1,73 @@ Current cluster status: * Node List: * Online: [ 18node1 18node2 18node3 ] * Full List of Resources: * shooter (stonith:fence_xvm): Started 18node2 * container1 (ocf:heartbeat:VirtualDomain): FAILED * container2 (ocf:heartbeat:VirtualDomain): FAILED * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Stopped: [ 18node1 18node2 18node3 ] Transition Summary: * Fence (reboot) lxc2 (resource: container2) 'guest is unclean' * Fence (reboot) lxc1 (resource: container1) 'guest is unclean' * Start container1 ( 18node1 ) * Start container2 ( 18node1 ) - * Recover lxc-ms:0 ( Promoted lxc1 ) - * Recover lxc-ms:1 ( Unpromoted lxc2 ) + * Recover lxc-ms:0 ( Promoted lxc1 ) + * Recover lxc-ms:1 ( Unpromoted lxc2 ) * Start lxc1 ( 18node1 ) * Start lxc2 ( 18node1 ) Executing Cluster Transition: * Resource action: container1 monitor on 18node3 * Resource action: container1 monitor on 18node2 * Resource action: container1 monitor on 18node1 * Resource action: container2 monitor on 18node3 * Resource action: container2 monitor on 18node2 * Resource action: container2 monitor on 18node1 * Resource action: lxc-ms monitor on 18node3 * Resource action: lxc-ms monitor on 18node2 * Resource action: lxc-ms monitor on 18node1 * Pseudo action: lxc-ms-master_demote_0 * Resource action: lxc1 monitor on 18node3 * Resource action: lxc1 monitor on 18node2 * Resource action: lxc1 monitor on 18node1 * Resource action: lxc2 monitor on 18node3 * Resource action: lxc2 monitor on 18node2 * Resource action: lxc2 monitor on 18node1 * Pseudo action: stonith-lxc2-reboot on lxc2 * Pseudo action: stonith-lxc1-reboot on lxc1 * Resource action: container1 start on 18node1 * Resource action: container2 start on 18node1 * Pseudo action: lxc-ms_demote_0 * Pseudo action: lxc-ms-master_demoted_0 * Pseudo action: lxc-ms-master_stop_0 * Resource action: lxc1 start on 18node1 * Resource action: lxc2 start on 18node1 * Pseudo action: lxc-ms_stop_0 * Pseudo action: lxc-ms_stop_0 * Pseudo action: lxc-ms-master_stopped_0 * Pseudo action: lxc-ms-master_start_0 * Resource action: lxc1 monitor=30000 on 18node1 * Resource action: lxc2 monitor=30000 on 18node1 * Resource action: lxc-ms start on lxc1 * Resource action: lxc-ms start on lxc2 * Pseudo action: lxc-ms-master_running_0 * Resource action: lxc-ms monitor=10000 on lxc2 * Pseudo action: lxc-ms-master_promote_0 * Resource action: lxc-ms promote on lxc1 * Pseudo action: lxc-ms-master_promoted_0 Revised Cluster Status: * Node List: * Online: [ 18node1 18node2 18node3 ] * GuestOnline: [ lxc1@18node1 lxc2@18node1 ] * Full List of Resources: * shooter (stonith:fence_xvm): Started 18node2 * container1 (ocf:heartbeat:VirtualDomain): Started 18node1 * container2 (ocf:heartbeat:VirtualDomain): Started 18node1 * Clone Set: lxc-ms-master [lxc-ms] (promotable): * Promoted: [ lxc1 ] * Unpromoted: [ lxc2 ] diff --git a/cts/scheduler/summary/whitebox-orphan-ms.summary b/cts/scheduler/summary/whitebox-orphan-ms.summary index 0d0007dcc6..7e1b45b272 100644 --- a/cts/scheduler/summary/whitebox-orphan-ms.summary +++ b/cts/scheduler/summary/whitebox-orphan-ms.summary @@ -1,87 +1,87 @@ Current cluster status: * Node List: * Online: [ 18node1 18node2 18node3 ] * GuestOnline: [ lxc1@18node1 lxc2@18node1 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started 18node2 * FencingPass (stonith:fence_dummy): Started 18node3 * FencingFail (stonith:fence_dummy): Started 18node3 * rsc_18node1 (ocf:heartbeat:IPaddr2): Started 18node1 * rsc_18node2 (ocf:heartbeat:IPaddr2): Started 18node2 * rsc_18node3 (ocf:heartbeat:IPaddr2): Started 18node3 * migrator (ocf:pacemaker:Dummy): Started 18node1 * Clone Set: Connectivity [ping-1]: * Started: [ 18node1 18node2 18node3 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ 18node1 ] * Unpromoted: [ 18node2 18node3 ] * Resource Group: group-1: * r192.168.122.87 (ocf:heartbeat:IPaddr2): Started 18node1 * r192.168.122.88 (ocf:heartbeat:IPaddr2): Started 18node1 * r192.168.122.89 (ocf:heartbeat:IPaddr2): Started 18node1 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started 18node1 * container2 (ocf:heartbeat:VirtualDomain): ORPHANED Started 18node1 * lxc1 (ocf:pacemaker:remote): ORPHANED Started 18node1 * lxc-ms (ocf:pacemaker:Stateful): ORPHANED Promoted [ lxc1 lxc2 ] * lxc2 (ocf:pacemaker:remote): ORPHANED Started 18node1 * container1 (ocf:heartbeat:VirtualDomain): ORPHANED Started 18node1 Transition Summary: * Move FencingFail ( 18node3 -> 18node1 ) * Stop container2 ( 18node1 ) due to node availability * Stop lxc1 ( 18node1 ) due to node availability - * Stop lxc-ms ( Promoted lxc1 ) due to node availability - * Stop lxc-ms ( Promoted lxc2 ) due to node availability + * Stop lxc-ms ( Promoted lxc1 ) due to node availability + * Stop lxc-ms ( Promoted lxc2 ) due to node availability * Stop lxc2 ( 18node1 ) due to node availability * Stop container1 ( 18node1 ) due to node availability Executing Cluster Transition: * Resource action: FencingFail stop on 18node3 * Resource action: lxc-ms demote on lxc2 * Resource action: lxc-ms demote on lxc1 * Resource action: FencingFail start on 18node1 * Resource action: lxc-ms stop on lxc2 * Resource action: lxc-ms stop on lxc1 * Resource action: lxc-ms delete on 18node3 * Resource action: lxc-ms delete on 18node2 * Resource action: lxc-ms delete on 18node1 * Resource action: lxc2 stop on 18node1 * Resource action: lxc2 delete on 18node3 * Resource action: lxc2 delete on 18node2 * Resource action: lxc2 delete on 18node1 * Resource action: container2 stop on 18node1 * Resource action: container2 delete on 18node3 * Resource action: container2 delete on 18node2 * Resource action: container2 delete on 18node1 * Resource action: lxc1 stop on 18node1 * Resource action: lxc1 delete on 18node3 * Resource action: lxc1 delete on 18node2 * Resource action: lxc1 delete on 18node1 * Resource action: container1 stop on 18node1 * Resource action: container1 delete on 18node3 * Resource action: container1 delete on 18node2 * Resource action: container1 delete on 18node1 Revised Cluster Status: * Node List: * Online: [ 18node1 18node2 18node3 ] * Full List of Resources: * Fencing (stonith:fence_xvm): Started 18node2 * FencingPass (stonith:fence_dummy): Started 18node3 * FencingFail (stonith:fence_dummy): Started 18node1 * rsc_18node1 (ocf:heartbeat:IPaddr2): Started 18node1 * rsc_18node2 (ocf:heartbeat:IPaddr2): Started 18node2 * rsc_18node3 (ocf:heartbeat:IPaddr2): Started 18node3 * migrator (ocf:pacemaker:Dummy): Started 18node1 * Clone Set: Connectivity [ping-1]: * Started: [ 18node1 18node2 18node3 ] * Clone Set: master-1 [stateful-1] (promotable): * Promoted: [ 18node1 ] * Unpromoted: [ 18node2 18node3 ] * Resource Group: group-1: * r192.168.122.87 (ocf:heartbeat:IPaddr2): Started 18node1 * r192.168.122.88 (ocf:heartbeat:IPaddr2): Started 18node1 * r192.168.122.89 (ocf:heartbeat:IPaddr2): Started 18node1 * lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started 18node1