Page Menu
Home
ClusterLabs Projects
Search
Configure Global Search
Log In
Files
F4512652
No One
Temporary
Actions
View File
Edit File
Delete File
View Transforms
Subscribe
Mute Notifications
Flag For Later
Award Token
Size
200 KB
Referenced Files
None
Subscribers
None
View Options
diff --git a/cts/scheduler/summary/bug-1572-1.summary b/cts/scheduler/summary/bug-1572-1.summary
index 6abedea530..c572db21d5 100644
--- a/cts/scheduler/summary/bug-1572-1.summary
+++ b/cts/scheduler/summary/bug-1572-1.summary
@@ -1,85 +1,85 @@
Current cluster status:
* Node List:
* Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ]
* Full List of Resources:
* Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable):
* Promoted: [ arc-tkincaidlx.wsicorp.com ]
* Unpromoted: [ arc-dknightlx ]
* Resource Group: grp_pgsql_mirror:
* fs_mirror (ocf:heartbeat:Filesystem): Started arc-tkincaidlx.wsicorp.com
* pgsql_5555 (ocf:heartbeat:pgsql): Started arc-tkincaidlx.wsicorp.com
* IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Started arc-tkincaidlx.wsicorp.com
Transition Summary:
- * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability
+ * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability
* Restart rsc_drbd_7788:1 ( Promoted arc-tkincaidlx.wsicorp.com ) due to resource definition change
* Restart fs_mirror ( arc-tkincaidlx.wsicorp.com ) due to required ms_drbd_7788 notified
* Restart pgsql_5555 ( arc-tkincaidlx.wsicorp.com ) due to required fs_mirror start
* Restart IPaddr_147_81_84_133 ( arc-tkincaidlx.wsicorp.com ) due to required pgsql_5555 start
Executing Cluster Transition:
* Pseudo action: ms_drbd_7788_pre_notify_demote_0
* Pseudo action: grp_pgsql_mirror_stop_0
* Resource action: IPaddr_147_81_84_133 stop on arc-tkincaidlx.wsicorp.com
* Resource action: rsc_drbd_7788:0 notify on arc-dknightlx
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-pre_notify_demote_0
* Resource action: pgsql_5555 stop on arc-tkincaidlx.wsicorp.com
* Resource action: fs_mirror stop on arc-tkincaidlx.wsicorp.com
* Pseudo action: grp_pgsql_mirror_stopped_0
* Pseudo action: ms_drbd_7788_demote_0
* Resource action: rsc_drbd_7788:1 demote on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_demoted_0
* Pseudo action: ms_drbd_7788_post_notify_demoted_0
* Resource action: rsc_drbd_7788:0 notify on arc-dknightlx
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-post_notify_demoted_0
* Pseudo action: ms_drbd_7788_pre_notify_stop_0
* Resource action: rsc_drbd_7788:0 notify on arc-dknightlx
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-pre_notify_stop_0
* Pseudo action: ms_drbd_7788_stop_0
* Resource action: rsc_drbd_7788:0 stop on arc-dknightlx
* Resource action: rsc_drbd_7788:1 stop on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_stopped_0
* Cluster action: do_shutdown on arc-dknightlx
* Pseudo action: ms_drbd_7788_post_notify_stopped_0
* Pseudo action: ms_drbd_7788_confirmed-post_notify_stopped_0
* Pseudo action: ms_drbd_7788_pre_notify_start_0
* Pseudo action: ms_drbd_7788_confirmed-pre_notify_start_0
* Pseudo action: ms_drbd_7788_start_0
* Resource action: rsc_drbd_7788:1 start on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_running_0
* Pseudo action: ms_drbd_7788_post_notify_running_0
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-post_notify_running_0
* Pseudo action: ms_drbd_7788_pre_notify_promote_0
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-pre_notify_promote_0
* Pseudo action: ms_drbd_7788_promote_0
* Resource action: rsc_drbd_7788:1 promote on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_promoted_0
* Pseudo action: ms_drbd_7788_post_notify_promoted_0
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-post_notify_promoted_0
* Pseudo action: grp_pgsql_mirror_start_0
* Resource action: fs_mirror start on arc-tkincaidlx.wsicorp.com
* Resource action: pgsql_5555 start on arc-tkincaidlx.wsicorp.com
* Resource action: pgsql_5555 monitor=30000 on arc-tkincaidlx.wsicorp.com
* Resource action: IPaddr_147_81_84_133 start on arc-tkincaidlx.wsicorp.com
* Resource action: IPaddr_147_81_84_133 monitor=25000 on arc-tkincaidlx.wsicorp.com
* Pseudo action: grp_pgsql_mirror_running_0
Revised Cluster Status:
* Node List:
* Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ]
* Full List of Resources:
* Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable):
* Promoted: [ arc-tkincaidlx.wsicorp.com ]
* Stopped: [ arc-dknightlx ]
* Resource Group: grp_pgsql_mirror:
* fs_mirror (ocf:heartbeat:Filesystem): Started arc-tkincaidlx.wsicorp.com
* pgsql_5555 (ocf:heartbeat:pgsql): Started arc-tkincaidlx.wsicorp.com
* IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Started arc-tkincaidlx.wsicorp.com
diff --git a/cts/scheduler/summary/bug-1572-2.summary b/cts/scheduler/summary/bug-1572-2.summary
index 7d4921dc36..012ca78dd6 100644
--- a/cts/scheduler/summary/bug-1572-2.summary
+++ b/cts/scheduler/summary/bug-1572-2.summary
@@ -1,61 +1,61 @@
Current cluster status:
* Node List:
* Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ]
* Full List of Resources:
* Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable):
* Promoted: [ arc-tkincaidlx.wsicorp.com ]
* Unpromoted: [ arc-dknightlx ]
* Resource Group: grp_pgsql_mirror:
* fs_mirror (ocf:heartbeat:Filesystem): Started arc-tkincaidlx.wsicorp.com
* pgsql_5555 (ocf:heartbeat:pgsql): Started arc-tkincaidlx.wsicorp.com
* IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Started arc-tkincaidlx.wsicorp.com
Transition Summary:
- * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability
+ * Stop rsc_drbd_7788:0 ( Unpromoted arc-dknightlx ) due to node availability
* Demote rsc_drbd_7788:1 ( Promoted -> Unpromoted arc-tkincaidlx.wsicorp.com )
* Stop fs_mirror ( arc-tkincaidlx.wsicorp.com ) due to node availability
* Stop pgsql_5555 ( arc-tkincaidlx.wsicorp.com ) due to node availability
* Stop IPaddr_147_81_84_133 ( arc-tkincaidlx.wsicorp.com ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms_drbd_7788_pre_notify_demote_0
* Pseudo action: grp_pgsql_mirror_stop_0
* Resource action: IPaddr_147_81_84_133 stop on arc-tkincaidlx.wsicorp.com
* Resource action: rsc_drbd_7788:0 notify on arc-dknightlx
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-pre_notify_demote_0
* Resource action: pgsql_5555 stop on arc-tkincaidlx.wsicorp.com
* Resource action: fs_mirror stop on arc-tkincaidlx.wsicorp.com
* Pseudo action: grp_pgsql_mirror_stopped_0
* Pseudo action: ms_drbd_7788_demote_0
* Resource action: rsc_drbd_7788:1 demote on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_demoted_0
* Pseudo action: ms_drbd_7788_post_notify_demoted_0
* Resource action: rsc_drbd_7788:0 notify on arc-dknightlx
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-post_notify_demoted_0
* Pseudo action: ms_drbd_7788_pre_notify_stop_0
* Resource action: rsc_drbd_7788:0 notify on arc-dknightlx
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-pre_notify_stop_0
* Pseudo action: ms_drbd_7788_stop_0
* Resource action: rsc_drbd_7788:0 stop on arc-dknightlx
* Pseudo action: ms_drbd_7788_stopped_0
* Cluster action: do_shutdown on arc-dknightlx
* Pseudo action: ms_drbd_7788_post_notify_stopped_0
* Resource action: rsc_drbd_7788:1 notify on arc-tkincaidlx.wsicorp.com
* Pseudo action: ms_drbd_7788_confirmed-post_notify_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ arc-dknightlx arc-tkincaidlx.wsicorp.com ]
* Full List of Resources:
* Clone Set: ms_drbd_7788 [rsc_drbd_7788] (promotable):
* Unpromoted: [ arc-tkincaidlx.wsicorp.com ]
* Stopped: [ arc-dknightlx ]
* Resource Group: grp_pgsql_mirror:
* fs_mirror (ocf:heartbeat:Filesystem): Stopped
* pgsql_5555 (ocf:heartbeat:pgsql): Stopped
* IPaddr_147_81_84_133 (ocf:heartbeat:IPaddr): Stopped
diff --git a/cts/scheduler/summary/bug-5059.summary b/cts/scheduler/summary/bug-5059.summary
index a33a2f60a2..c555d1dfb5 100644
--- a/cts/scheduler/summary/bug-5059.summary
+++ b/cts/scheduler/summary/bug-5059.summary
@@ -1,77 +1,77 @@
Current cluster status:
* Node List:
* Node gluster03.h: standby
* Online: [ gluster01.h gluster02.h ]
* OFFLINE: [ gluster04.h ]
* Full List of Resources:
* Clone Set: ms_stateful [g_stateful] (promotable):
* Resource Group: g_stateful:0:
* p_stateful1 (ocf:pacemaker:Stateful): Unpromoted gluster01.h
* p_stateful2 (ocf:pacemaker:Stateful): Stopped
* Resource Group: g_stateful:1:
* p_stateful1 (ocf:pacemaker:Stateful): Unpromoted gluster02.h
* p_stateful2 (ocf:pacemaker:Stateful): Stopped
* Stopped: [ gluster03.h gluster04.h ]
* Clone Set: c_dummy [p_dummy1]:
* Started: [ gluster01.h gluster02.h ]
Transition Summary:
- * Promote p_stateful1:0 ( Unpromoted -> Promoted gluster01.h )
- * Promote p_stateful2:0 ( Stopped -> Promoted gluster01.h )
+ * Promote p_stateful1:0 ( Unpromoted -> Promoted gluster01.h )
+ * Promote p_stateful2:0 ( Stopped -> Promoted gluster01.h )
* Start p_stateful2:1 ( gluster02.h )
Executing Cluster Transition:
* Pseudo action: ms_stateful_pre_notify_start_0
* Resource action: iptest delete on gluster02.h
* Resource action: ipsrc2 delete on gluster02.h
* Resource action: p_stateful1:0 notify on gluster01.h
* Resource action: p_stateful1:1 notify on gluster02.h
* Pseudo action: ms_stateful_confirmed-pre_notify_start_0
* Pseudo action: ms_stateful_start_0
* Pseudo action: g_stateful:0_start_0
* Resource action: p_stateful2:0 start on gluster01.h
* Pseudo action: g_stateful:1_start_0
* Resource action: p_stateful2:1 start on gluster02.h
* Pseudo action: g_stateful:0_running_0
* Pseudo action: g_stateful:1_running_0
* Pseudo action: ms_stateful_running_0
* Pseudo action: ms_stateful_post_notify_running_0
* Resource action: p_stateful1:0 notify on gluster01.h
* Resource action: p_stateful2:0 notify on gluster01.h
* Resource action: p_stateful1:1 notify on gluster02.h
* Resource action: p_stateful2:1 notify on gluster02.h
* Pseudo action: ms_stateful_confirmed-post_notify_running_0
* Pseudo action: ms_stateful_pre_notify_promote_0
* Resource action: p_stateful1:0 notify on gluster01.h
* Resource action: p_stateful2:0 notify on gluster01.h
* Resource action: p_stateful1:1 notify on gluster02.h
* Resource action: p_stateful2:1 notify on gluster02.h
* Pseudo action: ms_stateful_confirmed-pre_notify_promote_0
* Pseudo action: ms_stateful_promote_0
* Pseudo action: g_stateful:0_promote_0
* Resource action: p_stateful1:0 promote on gluster01.h
* Resource action: p_stateful2:0 promote on gluster01.h
* Pseudo action: g_stateful:0_promoted_0
* Pseudo action: ms_stateful_promoted_0
* Pseudo action: ms_stateful_post_notify_promoted_0
* Resource action: p_stateful1:0 notify on gluster01.h
* Resource action: p_stateful2:0 notify on gluster01.h
* Resource action: p_stateful1:1 notify on gluster02.h
* Resource action: p_stateful2:1 notify on gluster02.h
* Pseudo action: ms_stateful_confirmed-post_notify_promoted_0
* Resource action: p_stateful1:1 monitor=10000 on gluster02.h
* Resource action: p_stateful2:1 monitor=10000 on gluster02.h
Revised Cluster Status:
* Node List:
* Node gluster03.h: standby
* Online: [ gluster01.h gluster02.h ]
* OFFLINE: [ gluster04.h ]
* Full List of Resources:
* Clone Set: ms_stateful [g_stateful] (promotable):
* Promoted: [ gluster01.h ]
* Unpromoted: [ gluster02.h ]
* Clone Set: c_dummy [p_dummy1]:
* Started: [ gluster01.h gluster02.h ]
diff --git a/cts/scheduler/summary/bug-cl-5212.summary b/cts/scheduler/summary/bug-cl-5212.summary
index 48cb54bedc..e7a6e26833 100644
--- a/cts/scheduler/summary/bug-cl-5212.summary
+++ b/cts/scheduler/summary/bug-cl-5212.summary
@@ -1,69 +1,69 @@
Current cluster status:
* Node List:
* Node srv01: UNCLEAN (offline)
* Node srv02: UNCLEAN (offline)
* Online: [ srv03 ]
* Full List of Resources:
* Resource Group: grpStonith1:
* prmStonith1-1 (stonith:external/ssh): Started srv02 (UNCLEAN)
* Resource Group: grpStonith2:
* prmStonith2-1 (stonith:external/ssh): Started srv01 (UNCLEAN)
* Resource Group: grpStonith3:
* prmStonith3-1 (stonith:external/ssh): Started srv01 (UNCLEAN)
* Clone Set: msPostgresql [pgsql] (promotable):
* pgsql (ocf:pacemaker:Stateful): Unpromoted srv02 (UNCLEAN)
* pgsql (ocf:pacemaker:Stateful): Promoted srv01 (UNCLEAN)
* Unpromoted: [ srv03 ]
* Clone Set: clnPingd [prmPingd]:
* prmPingd (ocf:pacemaker:ping): Started srv02 (UNCLEAN)
* prmPingd (ocf:pacemaker:ping): Started srv01 (UNCLEAN)
* Started: [ srv03 ]
Transition Summary:
* Stop prmStonith1-1 ( srv02 ) blocked
* Stop prmStonith2-1 ( srv01 ) blocked
* Stop prmStonith3-1 ( srv01 ) due to node availability (blocked)
- * Stop pgsql:0 ( Unpromoted srv02 ) due to node availability (blocked)
- * Stop pgsql:1 ( Promoted srv01 ) due to node availability (blocked)
+ * Stop pgsql:0 ( Unpromoted srv02 ) due to node availability (blocked)
+ * Stop pgsql:1 ( Promoted srv01 ) due to node availability (blocked)
* Stop prmPingd:0 ( srv02 ) due to node availability (blocked)
* Stop prmPingd:1 ( srv01 ) due to node availability (blocked)
Executing Cluster Transition:
* Pseudo action: grpStonith1_stop_0
* Pseudo action: grpStonith1_start_0
* Pseudo action: grpStonith2_stop_0
* Pseudo action: grpStonith2_start_0
* Pseudo action: grpStonith3_stop_0
* Pseudo action: msPostgresql_pre_notify_stop_0
* Pseudo action: clnPingd_stop_0
* Resource action: pgsql notify on srv03
* Pseudo action: msPostgresql_confirmed-pre_notify_stop_0
* Pseudo action: msPostgresql_stop_0
* Pseudo action: clnPingd_stopped_0
* Pseudo action: msPostgresql_stopped_0
* Pseudo action: msPostgresql_post_notify_stopped_0
* Resource action: pgsql notify on srv03
* Pseudo action: msPostgresql_confirmed-post_notify_stopped_0
Revised Cluster Status:
* Node List:
* Node srv01: UNCLEAN (offline)
* Node srv02: UNCLEAN (offline)
* Online: [ srv03 ]
* Full List of Resources:
* Resource Group: grpStonith1:
* prmStonith1-1 (stonith:external/ssh): Started srv02 (UNCLEAN)
* Resource Group: grpStonith2:
* prmStonith2-1 (stonith:external/ssh): Started srv01 (UNCLEAN)
* Resource Group: grpStonith3:
* prmStonith3-1 (stonith:external/ssh): Started srv01 (UNCLEAN)
* Clone Set: msPostgresql [pgsql] (promotable):
* pgsql (ocf:pacemaker:Stateful): Unpromoted srv02 (UNCLEAN)
* pgsql (ocf:pacemaker:Stateful): Promoted srv01 (UNCLEAN)
* Unpromoted: [ srv03 ]
* Clone Set: clnPingd [prmPingd]:
* prmPingd (ocf:pacemaker:ping): Started srv02 (UNCLEAN)
* prmPingd (ocf:pacemaker:ping): Started srv01 (UNCLEAN)
* Started: [ srv03 ]
diff --git a/cts/scheduler/summary/bug-cl-5247.summary b/cts/scheduler/summary/bug-cl-5247.summary
index 056e526490..67ad0c3ded 100644
--- a/cts/scheduler/summary/bug-cl-5247.summary
+++ b/cts/scheduler/summary/bug-cl-5247.summary
@@ -1,87 +1,87 @@
Using the original execution date of: 2015-08-12 02:53:40Z
Current cluster status:
* Node List:
* Online: [ bl460g8n3 bl460g8n4 ]
* GuestOnline: [ pgsr01@bl460g8n3 ]
* Full List of Resources:
* prmDB1 (ocf:heartbeat:VirtualDomain): Started bl460g8n3
* prmDB2 (ocf:heartbeat:VirtualDomain): FAILED bl460g8n4
* Resource Group: grpStonith1:
* prmStonith1-2 (stonith:external/ipmi): Started bl460g8n4
* Resource Group: grpStonith2:
* prmStonith2-2 (stonith:external/ipmi): Started bl460g8n3
* Resource Group: master-group:
* vip-master (ocf:heartbeat:Dummy): FAILED pgsr02
* vip-rep (ocf:heartbeat:Dummy): FAILED pgsr02
* Clone Set: msPostgresql [pgsql] (promotable):
* Promoted: [ pgsr01 ]
* Stopped: [ bl460g8n3 bl460g8n4 ]
Transition Summary:
* Fence (off) pgsr02 (resource: prmDB2) 'guest is unclean'
* Stop prmDB2 ( bl460g8n4 ) due to node availability
* Recover vip-master ( pgsr02 -> pgsr01 )
* Recover vip-rep ( pgsr02 -> pgsr01 )
- * Stop pgsql:0 ( Promoted pgsr02 ) due to node availability
+ * Stop pgsql:0 ( Promoted pgsr02 ) due to node availability
* Stop pgsr02 ( bl460g8n4 ) due to node availability
Executing Cluster Transition:
* Resource action: vip-master monitor on pgsr01
* Resource action: vip-rep monitor on pgsr01
* Pseudo action: msPostgresql_pre_notify_demote_0
* Resource action: pgsr01 monitor on bl460g8n4
* Resource action: pgsr02 stop on bl460g8n4
* Resource action: pgsr02 monitor on bl460g8n3
* Resource action: prmDB2 stop on bl460g8n4
* Resource action: pgsql notify on pgsr01
* Pseudo action: msPostgresql_confirmed-pre_notify_demote_0
* Pseudo action: msPostgresql_demote_0
* Pseudo action: stonith-pgsr02-off on pgsr02
* Pseudo action: pgsql_post_notify_stop_0
* Pseudo action: pgsql_demote_0
* Pseudo action: msPostgresql_demoted_0
* Pseudo action: msPostgresql_post_notify_demoted_0
* Resource action: pgsql notify on pgsr01
* Pseudo action: msPostgresql_confirmed-post_notify_demoted_0
* Pseudo action: msPostgresql_pre_notify_stop_0
* Pseudo action: master-group_stop_0
* Pseudo action: vip-rep_stop_0
* Resource action: pgsql notify on pgsr01
* Pseudo action: msPostgresql_confirmed-pre_notify_stop_0
* Pseudo action: msPostgresql_stop_0
* Pseudo action: vip-master_stop_0
* Pseudo action: pgsql_stop_0
* Pseudo action: msPostgresql_stopped_0
* Pseudo action: master-group_stopped_0
* Pseudo action: master-group_start_0
* Resource action: vip-master start on pgsr01
* Resource action: vip-rep start on pgsr01
* Pseudo action: msPostgresql_post_notify_stopped_0
* Pseudo action: master-group_running_0
* Resource action: vip-master monitor=10000 on pgsr01
* Resource action: vip-rep monitor=10000 on pgsr01
* Resource action: pgsql notify on pgsr01
* Pseudo action: msPostgresql_confirmed-post_notify_stopped_0
* Pseudo action: pgsql_notified_0
* Resource action: pgsql monitor=9000 on pgsr01
Using the original execution date of: 2015-08-12 02:53:40Z
Revised Cluster Status:
* Node List:
* Online: [ bl460g8n3 bl460g8n4 ]
* GuestOnline: [ pgsr01@bl460g8n3 ]
* Full List of Resources:
* prmDB1 (ocf:heartbeat:VirtualDomain): Started bl460g8n3
* prmDB2 (ocf:heartbeat:VirtualDomain): FAILED
* Resource Group: grpStonith1:
* prmStonith1-2 (stonith:external/ipmi): Started bl460g8n4
* Resource Group: grpStonith2:
* prmStonith2-2 (stonith:external/ipmi): Started bl460g8n3
* Resource Group: master-group:
* vip-master (ocf:heartbeat:Dummy): FAILED [ pgsr01 pgsr02 ]
* vip-rep (ocf:heartbeat:Dummy): FAILED [ pgsr01 pgsr02 ]
* Clone Set: msPostgresql [pgsql] (promotable):
* Promoted: [ pgsr01 ]
* Stopped: [ bl460g8n3 bl460g8n4 ]
diff --git a/cts/scheduler/summary/bug-lf-2606.summary b/cts/scheduler/summary/bug-lf-2606.summary
index e0b7ebf0e6..004788e80b 100644
--- a/cts/scheduler/summary/bug-lf-2606.summary
+++ b/cts/scheduler/summary/bug-lf-2606.summary
@@ -1,46 +1,46 @@
1 of 5 resource instances DISABLED and 0 BLOCKED from further action due to failure
Current cluster status:
* Node List:
* Node node2: UNCLEAN (online)
* Online: [ node1 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* rsc1 (ocf:pacemaker:Dummy): FAILED node2 (disabled)
* rsc2 (ocf:pacemaker:Dummy): Started node2
* Clone Set: ms3 [rsc3] (promotable):
* Promoted: [ node2 ]
* Unpromoted: [ node1 ]
Transition Summary:
* Fence (reboot) node2 'rsc1 failed there'
* Stop rsc1 ( node2 ) due to node availability
* Move rsc2 ( node2 -> node1 )
- * Stop rsc3:1 ( Promoted node2 ) due to node availability
+ * Stop rsc3:1 ( Promoted node2 ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms3_demote_0
* Fencing node2 (reboot)
* Pseudo action: rsc1_stop_0
* Pseudo action: rsc2_stop_0
* Pseudo action: rsc3:1_demote_0
* Pseudo action: ms3_demoted_0
* Pseudo action: ms3_stop_0
* Resource action: rsc2 start on node1
* Pseudo action: rsc3:1_stop_0
* Pseudo action: ms3_stopped_0
* Resource action: rsc2 monitor=10000 on node1
Revised Cluster Status:
* Node List:
* Online: [ node1 ]
* OFFLINE: [ node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* rsc1 (ocf:pacemaker:Dummy): Stopped (disabled)
* rsc2 (ocf:pacemaker:Dummy): Started node1
* Clone Set: ms3 [rsc3] (promotable):
* Unpromoted: [ node1 ]
* Stopped: [ node2 ]
diff --git a/cts/scheduler/summary/bug-pm-12.summary b/cts/scheduler/summary/bug-pm-12.summary
index 7b811d1a02..c4f3adb908 100644
--- a/cts/scheduler/summary/bug-pm-12.summary
+++ b/cts/scheduler/summary/bug-pm-12.summary
@@ -1,57 +1,57 @@
Current cluster status:
* Node List:
* Online: [ node-a node-b ]
* Full List of Resources:
* Clone Set: ms-sf [group] (promotable) (unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* stateful-2:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted node-a
Transition Summary:
- * Restart stateful-2:0 ( Unpromoted node-b ) due to resource definition change
- * Restart stateful-2:1 ( Promoted node-a ) due to resource definition change
+ * Restart stateful-2:0 ( Unpromoted node-b ) due to resource definition change
+ * Restart stateful-2:1 ( Promoted node-a ) due to resource definition change
Executing Cluster Transition:
* Pseudo action: ms-sf_demote_0
* Pseudo action: group:1_demote_0
* Resource action: stateful-2:1 demote on node-a
* Pseudo action: group:1_demoted_0
* Pseudo action: ms-sf_demoted_0
* Pseudo action: ms-sf_stop_0
* Pseudo action: group:0_stop_0
* Resource action: stateful-2:0 stop on node-b
* Pseudo action: group:1_stop_0
* Resource action: stateful-2:1 stop on node-a
* Pseudo action: group:0_stopped_0
* Pseudo action: group:1_stopped_0
* Pseudo action: ms-sf_stopped_0
* Pseudo action: ms-sf_start_0
* Pseudo action: group:0_start_0
* Resource action: stateful-2:0 start on node-b
* Pseudo action: group:1_start_0
* Resource action: stateful-2:1 start on node-a
* Pseudo action: group:0_running_0
* Pseudo action: group:1_running_0
* Pseudo action: ms-sf_running_0
* Pseudo action: ms-sf_promote_0
* Pseudo action: group:1_promote_0
* Resource action: stateful-2:1 promote on node-a
* Pseudo action: group:1_promoted_0
* Pseudo action: ms-sf_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ node-a node-b ]
* Full List of Resources:
* Clone Set: ms-sf [group] (promotable) (unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* stateful-2:0 (ocf:heartbeat:Stateful): Unpromoted node-b
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted node-a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted node-a
diff --git a/cts/scheduler/summary/bundle-order-fencing.summary b/cts/scheduler/summary/bundle-order-fencing.summary
index 387c05532a..8cb40718db 100644
--- a/cts/scheduler/summary/bundle-order-fencing.summary
+++ b/cts/scheduler/summary/bundle-order-fencing.summary
@@ -1,220 +1,220 @@
Using the original execution date of: 2017-09-12 10:51:59Z
Current cluster status:
* Node List:
* Node controller-0: UNCLEAN (offline)
* Online: [ controller-1 controller-2 ]
* GuestOnline: [ galera-bundle-1@controller-1 galera-bundle-2@controller-2 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ]
* Full List of Resources:
* Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]:
* rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): FAILED controller-0 (UNCLEAN)
* rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1
* rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2
* Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]:
* galera-bundle-0 (ocf:heartbeat:galera): FAILED Promoted controller-0 (UNCLEAN)
* galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1
* galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2
* Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]:
* redis-bundle-0 (ocf:heartbeat:redis): FAILED Promoted controller-0 (UNCLEAN)
* redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1
* redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2
* ip-192.168.24.7 (ocf:heartbeat:IPaddr2): Started controller-0 (UNCLEAN)
* ip-10.0.0.109 (ocf:heartbeat:IPaddr2): Started controller-0 (UNCLEAN)
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.3.19 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-0 (UNCLEAN)
* Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]:
* haproxy-bundle-docker-0 (ocf:heartbeat:docker): Started controller-0 (UNCLEAN)
* haproxy-bundle-docker-1 (ocf:heartbeat:docker): Started controller-2
* haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-1
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-2
* stonith-fence_ipmilan-525400efba5c (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-5254003e8e97 (stonith:fence_ipmilan): Started controller-0 (UNCLEAN)
* stonith-fence_ipmilan-5254000dcb3f (stonith:fence_ipmilan): Started controller-0 (UNCLEAN)
Transition Summary:
* Fence (off) redis-bundle-0 (resource: redis-bundle-docker-0) 'guest is unclean'
* Fence (off) rabbitmq-bundle-0 (resource: rabbitmq-bundle-docker-0) 'guest is unclean'
* Fence (off) galera-bundle-0 (resource: galera-bundle-docker-0) 'guest is unclean'
* Fence (reboot) controller-0 'peer is no longer part of the cluster'
* Stop rabbitmq-bundle-docker-0 ( controller-0 ) due to node availability
* Stop rabbitmq-bundle-0 ( controller-0 ) due to unrunnable rabbitmq-bundle-docker-0 start
* Stop rabbitmq:0 ( rabbitmq-bundle-0 ) due to unrunnable rabbitmq-bundle-docker-0 start
* Stop galera-bundle-docker-0 ( controller-0 ) due to node availability
* Stop galera-bundle-0 ( controller-0 ) due to unrunnable galera-bundle-docker-0 start
- * Stop galera:0 ( Promoted galera-bundle-0 ) due to unrunnable galera-bundle-docker-0 start
+ * Stop galera:0 ( Promoted galera-bundle-0 ) due to unrunnable galera-bundle-docker-0 start
* Stop redis-bundle-docker-0 ( controller-0 ) due to node availability
* Stop redis-bundle-0 ( controller-0 ) due to unrunnable redis-bundle-docker-0 start
- * Stop redis:0 ( Promoted redis-bundle-0 ) due to unrunnable redis-bundle-docker-0 start
+ * Stop redis:0 ( Promoted redis-bundle-0 ) due to unrunnable redis-bundle-docker-0 start
* Promote redis:1 ( Unpromoted -> Promoted redis-bundle-1 )
* Move ip-192.168.24.7 ( controller-0 -> controller-2 )
* Move ip-10.0.0.109 ( controller-0 -> controller-1 )
* Move ip-172.17.4.11 ( controller-0 -> controller-1 )
* Stop haproxy-bundle-docker-0 ( controller-0 ) due to node availability
* Move stonith-fence_ipmilan-5254003e8e97 ( controller-0 -> controller-1 )
* Move stonith-fence_ipmilan-5254000dcb3f ( controller-0 -> controller-2 )
Executing Cluster Transition:
* Pseudo action: rabbitmq-bundle-clone_pre_notify_stop_0
* Pseudo action: rabbitmq-bundle-0_stop_0
* Resource action: rabbitmq-bundle-0 monitor on controller-2
* Resource action: rabbitmq-bundle-0 monitor on controller-1
* Resource action: rabbitmq-bundle-1 monitor on controller-2
* Resource action: rabbitmq-bundle-2 monitor on controller-1
* Pseudo action: galera-bundle-0_stop_0
* Resource action: galera-bundle-0 monitor on controller-2
* Resource action: galera-bundle-0 monitor on controller-1
* Resource action: galera-bundle-1 monitor on controller-2
* Resource action: galera-bundle-2 monitor on controller-1
* Resource action: redis cancel=45000 on redis-bundle-1
* Resource action: redis cancel=60000 on redis-bundle-1
* Pseudo action: redis-bundle-master_pre_notify_demote_0
* Pseudo action: redis-bundle-0_stop_0
* Resource action: redis-bundle-0 monitor on controller-2
* Resource action: redis-bundle-0 monitor on controller-1
* Resource action: redis-bundle-1 monitor on controller-2
* Resource action: redis-bundle-2 monitor on controller-1
* Pseudo action: stonith-fence_ipmilan-5254003e8e97_stop_0
* Pseudo action: stonith-fence_ipmilan-5254000dcb3f_stop_0
* Pseudo action: haproxy-bundle_stop_0
* Pseudo action: redis-bundle_demote_0
* Pseudo action: galera-bundle_demote_0
* Pseudo action: rabbitmq-bundle_stop_0
* Pseudo action: rabbitmq-bundle_start_0
* Fencing controller-0 (reboot)
* Resource action: rabbitmq notify on rabbitmq-bundle-1
* Resource action: rabbitmq notify on rabbitmq-bundle-2
* Pseudo action: rabbitmq-bundle-clone_confirmed-pre_notify_stop_0
* Pseudo action: rabbitmq-bundle-docker-0_stop_0
* Pseudo action: galera-bundle-master_demote_0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-pre_notify_demote_0
* Pseudo action: redis-bundle-master_demote_0
* Pseudo action: haproxy-bundle-docker-0_stop_0
* Resource action: stonith-fence_ipmilan-5254003e8e97 start on controller-1
* Resource action: stonith-fence_ipmilan-5254000dcb3f start on controller-2
* Pseudo action: stonith-redis-bundle-0-off on redis-bundle-0
* Pseudo action: stonith-rabbitmq-bundle-0-off on rabbitmq-bundle-0
* Pseudo action: stonith-galera-bundle-0-off on galera-bundle-0
* Pseudo action: haproxy-bundle_stopped_0
* Pseudo action: rabbitmq_post_notify_stop_0
* Pseudo action: rabbitmq-bundle-clone_stop_0
* Pseudo action: galera_demote_0
* Pseudo action: galera-bundle-master_demoted_0
* Pseudo action: redis_post_notify_stop_0
* Pseudo action: redis_demote_0
* Pseudo action: redis-bundle-master_demoted_0
* Pseudo action: ip-192.168.24.7_stop_0
* Pseudo action: ip-10.0.0.109_stop_0
* Pseudo action: ip-172.17.4.11_stop_0
* Resource action: stonith-fence_ipmilan-5254003e8e97 monitor=60000 on controller-1
* Resource action: stonith-fence_ipmilan-5254000dcb3f monitor=60000 on controller-2
* Pseudo action: galera-bundle_demoted_0
* Pseudo action: galera-bundle_stop_0
* Pseudo action: rabbitmq_stop_0
* Pseudo action: rabbitmq-bundle-clone_stopped_0
* Pseudo action: galera-bundle-master_stop_0
* Pseudo action: galera-bundle-docker-0_stop_0
* Pseudo action: redis-bundle-master_post_notify_demoted_0
* Resource action: ip-192.168.24.7 start on controller-2
* Resource action: ip-10.0.0.109 start on controller-1
* Resource action: ip-172.17.4.11 start on controller-1
* Pseudo action: rabbitmq-bundle-clone_post_notify_stopped_0
* Pseudo action: galera_stop_0
* Pseudo action: galera-bundle-master_stopped_0
* Pseudo action: galera-bundle-master_start_0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-post_notify_demoted_0
* Pseudo action: redis-bundle-master_pre_notify_stop_0
* Resource action: ip-192.168.24.7 monitor=10000 on controller-2
* Resource action: ip-10.0.0.109 monitor=10000 on controller-1
* Resource action: ip-172.17.4.11 monitor=10000 on controller-1
* Pseudo action: redis-bundle_demoted_0
* Pseudo action: redis-bundle_stop_0
* Pseudo action: galera-bundle_stopped_0
* Resource action: rabbitmq notify on rabbitmq-bundle-1
* Resource action: rabbitmq notify on rabbitmq-bundle-2
* Pseudo action: rabbitmq-bundle-clone_confirmed-post_notify_stopped_0
* Pseudo action: rabbitmq-bundle-clone_pre_notify_start_0
* Pseudo action: galera-bundle-master_running_0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-pre_notify_stop_0
* Pseudo action: redis-bundle-master_stop_0
* Pseudo action: redis-bundle-docker-0_stop_0
* Pseudo action: galera-bundle_running_0
* Pseudo action: rabbitmq-bundle_stopped_0
* Pseudo action: rabbitmq_notified_0
* Pseudo action: rabbitmq-bundle-clone_confirmed-pre_notify_start_0
* Pseudo action: rabbitmq-bundle-clone_start_0
* Pseudo action: redis_stop_0
* Pseudo action: redis-bundle-master_stopped_0
* Pseudo action: rabbitmq-bundle-clone_running_0
* Pseudo action: redis-bundle-master_post_notify_stopped_0
* Pseudo action: rabbitmq-bundle-clone_post_notify_running_0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-post_notify_stopped_0
* Pseudo action: redis-bundle-master_pre_notify_start_0
* Pseudo action: redis-bundle_stopped_0
* Pseudo action: rabbitmq-bundle-clone_confirmed-post_notify_running_0
* Pseudo action: redis_notified_0
* Pseudo action: redis-bundle-master_confirmed-pre_notify_start_0
* Pseudo action: redis-bundle-master_start_0
* Pseudo action: rabbitmq-bundle_running_0
* Pseudo action: redis-bundle-master_running_0
* Pseudo action: redis-bundle-master_post_notify_running_0
* Pseudo action: redis-bundle-master_confirmed-post_notify_running_0
* Pseudo action: redis-bundle_running_0
* Pseudo action: redis-bundle-master_pre_notify_promote_0
* Pseudo action: redis-bundle_promote_0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-pre_notify_promote_0
* Pseudo action: redis-bundle-master_promote_0
* Resource action: redis promote on redis-bundle-1
* Pseudo action: redis-bundle-master_promoted_0
* Pseudo action: redis-bundle-master_post_notify_promoted_0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-post_notify_promoted_0
* Pseudo action: redis-bundle_promoted_0
* Resource action: redis monitor=20000 on redis-bundle-1
Using the original execution date of: 2017-09-12 10:51:59Z
Revised Cluster Status:
* Node List:
* Online: [ controller-1 controller-2 ]
* OFFLINE: [ controller-0 ]
* GuestOnline: [ galera-bundle-1@controller-1 galera-bundle-2@controller-2 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ]
* Full List of Resources:
* Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]:
* rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): FAILED
* rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1
* rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2
* Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]:
* galera-bundle-0 (ocf:heartbeat:galera): FAILED Promoted
* galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1
* galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2
* Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]:
* redis-bundle-0 (ocf:heartbeat:redis): FAILED Promoted
* redis-bundle-1 (ocf:heartbeat:redis): Promoted controller-1
* redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2
* ip-192.168.24.7 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-10.0.0.109 (ocf:heartbeat:IPaddr2): Started controller-1
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.3.19 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1
* Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]:
* haproxy-bundle-docker-0 (ocf:heartbeat:docker): Stopped
* haproxy-bundle-docker-1 (ocf:heartbeat:docker): Started controller-2
* haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-1
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-2
* stonith-fence_ipmilan-525400efba5c (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-5254003e8e97 (stonith:fence_ipmilan): Started controller-1
* stonith-fence_ipmilan-5254000dcb3f (stonith:fence_ipmilan): Started controller-2
diff --git a/cts/scheduler/summary/bundle-order-stop-on-remote.summary b/cts/scheduler/summary/bundle-order-stop-on-remote.summary
index bf94ce3c72..8cd17eef61 100644
--- a/cts/scheduler/summary/bundle-order-stop-on-remote.summary
+++ b/cts/scheduler/summary/bundle-order-stop-on-remote.summary
@@ -1,224 +1,224 @@
Current cluster status:
* Node List:
* RemoteNode database-0: UNCLEAN (offline)
* RemoteNode database-2: UNCLEAN (offline)
* Online: [ controller-0 controller-1 controller-2 ]
* RemoteOnline: [ database-1 messaging-0 messaging-1 messaging-2 ]
* GuestOnline: [ galera-bundle-1@controller-2 rabbitmq-bundle-0@controller-2 rabbitmq-bundle-1@controller-2 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-2@controller-2 ]
* Full List of Resources:
* database-0 (ocf:pacemaker:remote): Stopped
* database-1 (ocf:pacemaker:remote): Started controller-2
* database-2 (ocf:pacemaker:remote): Stopped
* messaging-0 (ocf:pacemaker:remote): Started controller-2
* messaging-1 (ocf:pacemaker:remote): Started controller-2
* messaging-2 (ocf:pacemaker:remote): Started controller-2
* Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]:
* rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started messaging-0
* rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started messaging-1
* rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started messaging-2
* Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]:
* galera-bundle-0 (ocf:heartbeat:galera): FAILED Promoted database-0 (UNCLEAN)
* galera-bundle-1 (ocf:heartbeat:galera): Promoted database-1
* galera-bundle-2 (ocf:heartbeat:galera): FAILED Promoted database-2 (UNCLEAN)
* Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]:
* redis-bundle-0 (ocf:heartbeat:redis): Unpromoted controller-0
* redis-bundle-1 (ocf:heartbeat:redis): Stopped
* redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2
* ip-192.168.24.11 (ocf:heartbeat:IPaddr2): Stopped
* ip-10.0.0.104 (ocf:heartbeat:IPaddr2): Stopped
* ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.11 (ocf:heartbeat:IPaddr2): Stopped
* ip-172.17.3.13 (ocf:heartbeat:IPaddr2): Stopped
* ip-172.17.4.19 (ocf:heartbeat:IPaddr2): Started controller-2
* Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]:
* haproxy-bundle-docker-0 (ocf:heartbeat:docker): Started controller-0
* haproxy-bundle-docker-1 (ocf:heartbeat:docker): Stopped
* haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-2
* openstack-cinder-volume (systemd:openstack-cinder-volume): Stopped
* stonith-fence_ipmilan-525400244e09 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400cdec10 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400c709f7 (stonith:fence_ipmilan): Stopped
* stonith-fence_ipmilan-525400a7f9e0 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400a25787 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-5254005ea387 (stonith:fence_ipmilan): Stopped
* stonith-fence_ipmilan-525400542c06 (stonith:fence_ipmilan): Stopped
* stonith-fence_ipmilan-525400aac413 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400498d34 (stonith:fence_ipmilan): Stopped
Transition Summary:
* Fence (reboot) galera-bundle-2 (resource: galera-bundle-docker-2) 'guest is unclean'
* Fence (reboot) galera-bundle-0 (resource: galera-bundle-docker-0) 'guest is unclean'
* Start database-0 ( controller-0 )
* Start database-2 ( controller-1 )
* Recover galera-bundle-docker-0 ( database-0 )
* Start galera-bundle-0 ( controller-0 )
- * Recover galera:0 ( Promoted galera-bundle-0 )
+ * Recover galera:0 ( Promoted galera-bundle-0 )
* Recover galera-bundle-docker-2 ( database-2 )
* Start galera-bundle-2 ( controller-1 )
- * Recover galera:2 ( Promoted galera-bundle-2 )
+ * Recover galera:2 ( Promoted galera-bundle-2 )
* Promote redis:0 ( Unpromoted -> Promoted redis-bundle-0 )
* Start redis-bundle-docker-1 ( controller-1 )
* Start redis-bundle-1 ( controller-1 )
* Start redis:1 ( redis-bundle-1 )
* Start ip-192.168.24.11 ( controller-0 )
* Start ip-10.0.0.104 ( controller-1 )
* Start ip-172.17.1.11 ( controller-0 )
* Start ip-172.17.3.13 ( controller-1 )
* Start haproxy-bundle-docker-1 ( controller-1 )
* Start openstack-cinder-volume ( controller-0 )
* Start stonith-fence_ipmilan-525400c709f7 ( controller-1 )
* Start stonith-fence_ipmilan-5254005ea387 ( controller-1 )
* Start stonith-fence_ipmilan-525400542c06 ( controller-0 )
* Start stonith-fence_ipmilan-525400498d34 ( controller-1 )
Executing Cluster Transition:
* Resource action: database-0 start on controller-0
* Resource action: database-2 start on controller-1
* Pseudo action: redis-bundle-master_pre_notify_start_0
* Resource action: stonith-fence_ipmilan-525400c709f7 start on controller-1
* Resource action: stonith-fence_ipmilan-5254005ea387 start on controller-1
* Resource action: stonith-fence_ipmilan-525400542c06 start on controller-0
* Resource action: stonith-fence_ipmilan-525400498d34 start on controller-1
* Pseudo action: redis-bundle_start_0
* Pseudo action: galera-bundle_demote_0
* Resource action: database-0 monitor=20000 on controller-0
* Resource action: database-2 monitor=20000 on controller-1
* Pseudo action: galera-bundle-master_demote_0
* Resource action: redis notify on redis-bundle-0
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-pre_notify_start_0
* Pseudo action: redis-bundle-master_start_0
* Resource action: stonith-fence_ipmilan-525400c709f7 monitor=60000 on controller-1
* Resource action: stonith-fence_ipmilan-5254005ea387 monitor=60000 on controller-1
* Resource action: stonith-fence_ipmilan-525400542c06 monitor=60000 on controller-0
* Resource action: stonith-fence_ipmilan-525400498d34 monitor=60000 on controller-1
* Pseudo action: galera_demote_0
* Pseudo action: galera_demote_0
* Pseudo action: galera-bundle-master_demoted_0
* Pseudo action: galera-bundle_demoted_0
* Pseudo action: galera-bundle_stop_0
* Resource action: galera-bundle-docker-0 stop on database-0
* Resource action: galera-bundle-docker-2 stop on database-2
* Pseudo action: stonith-galera-bundle-2-reboot on galera-bundle-2
* Pseudo action: stonith-galera-bundle-0-reboot on galera-bundle-0
* Pseudo action: galera-bundle-master_stop_0
* Resource action: redis-bundle-docker-1 start on controller-1
* Resource action: redis-bundle-1 monitor on controller-1
* Resource action: ip-192.168.24.11 start on controller-0
* Resource action: ip-10.0.0.104 start on controller-1
* Resource action: ip-172.17.1.11 start on controller-0
* Resource action: ip-172.17.3.13 start on controller-1
* Resource action: openstack-cinder-volume start on controller-0
* Pseudo action: haproxy-bundle_start_0
* Pseudo action: galera_stop_0
* Resource action: redis-bundle-docker-1 monitor=60000 on controller-1
* Resource action: redis-bundle-1 start on controller-1
* Resource action: ip-192.168.24.11 monitor=10000 on controller-0
* Resource action: ip-10.0.0.104 monitor=10000 on controller-1
* Resource action: ip-172.17.1.11 monitor=10000 on controller-0
* Resource action: ip-172.17.3.13 monitor=10000 on controller-1
* Resource action: haproxy-bundle-docker-1 start on controller-1
* Resource action: openstack-cinder-volume monitor=60000 on controller-0
* Pseudo action: haproxy-bundle_running_0
* Pseudo action: galera_stop_0
* Pseudo action: galera-bundle-master_stopped_0
* Resource action: redis start on redis-bundle-1
* Pseudo action: redis-bundle-master_running_0
* Resource action: redis-bundle-1 monitor=30000 on controller-1
* Resource action: haproxy-bundle-docker-1 monitor=60000 on controller-1
* Pseudo action: galera-bundle_stopped_0
* Pseudo action: galera-bundle_start_0
* Pseudo action: galera-bundle-master_start_0
* Resource action: galera-bundle-docker-0 start on database-0
* Resource action: galera-bundle-0 monitor on controller-1
* Resource action: galera-bundle-docker-2 start on database-2
* Resource action: galera-bundle-2 monitor on controller-1
* Pseudo action: redis-bundle-master_post_notify_running_0
* Resource action: galera-bundle-docker-0 monitor=60000 on database-0
* Resource action: galera-bundle-0 start on controller-0
* Resource action: galera-bundle-docker-2 monitor=60000 on database-2
* Resource action: galera-bundle-2 start on controller-1
* Resource action: redis notify on redis-bundle-0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-post_notify_running_0
* Pseudo action: redis-bundle_running_0
* Resource action: galera start on galera-bundle-0
* Resource action: galera start on galera-bundle-2
* Pseudo action: galera-bundle-master_running_0
* Resource action: galera-bundle-0 monitor=30000 on controller-0
* Resource action: galera-bundle-2 monitor=30000 on controller-1
* Pseudo action: redis-bundle-master_pre_notify_promote_0
* Pseudo action: redis-bundle_promote_0
* Pseudo action: galera-bundle_running_0
* Resource action: redis notify on redis-bundle-0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-pre_notify_promote_0
* Pseudo action: redis-bundle-master_promote_0
* Pseudo action: galera-bundle_promote_0
* Pseudo action: galera-bundle-master_promote_0
* Resource action: redis promote on redis-bundle-0
* Pseudo action: redis-bundle-master_promoted_0
* Resource action: galera promote on galera-bundle-0
* Resource action: galera promote on galera-bundle-2
* Pseudo action: galera-bundle-master_promoted_0
* Pseudo action: redis-bundle-master_post_notify_promoted_0
* Pseudo action: galera-bundle_promoted_0
* Resource action: galera monitor=10000 on galera-bundle-0
* Resource action: galera monitor=10000 on galera-bundle-2
* Resource action: redis notify on redis-bundle-0
* Resource action: redis notify on redis-bundle-1
* Resource action: redis notify on redis-bundle-2
* Pseudo action: redis-bundle-master_confirmed-post_notify_promoted_0
* Pseudo action: redis-bundle_promoted_0
* Resource action: redis monitor=20000 on redis-bundle-0
* Resource action: redis monitor=60000 on redis-bundle-1
* Resource action: redis monitor=45000 on redis-bundle-1
Revised Cluster Status:
* Node List:
* Online: [ controller-0 controller-1 controller-2 ]
* RemoteOnline: [ database-0 database-1 database-2 messaging-0 messaging-1 messaging-2 ]
* GuestOnline: [ galera-bundle-0@controller-0 galera-bundle-1@controller-2 galera-bundle-2@controller-1 rabbitmq-bundle-0@controller-2 rabbitmq-bundle-1@controller-2 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ]
* Full List of Resources:
* database-0 (ocf:pacemaker:remote): Started controller-0
* database-1 (ocf:pacemaker:remote): Started controller-2
* database-2 (ocf:pacemaker:remote): Started controller-1
* messaging-0 (ocf:pacemaker:remote): Started controller-2
* messaging-1 (ocf:pacemaker:remote): Started controller-2
* messaging-2 (ocf:pacemaker:remote): Started controller-2
* Container bundle set: rabbitmq-bundle [192.168.24.1:8787/rhosp12/openstack-rabbitmq-docker:pcmklatest]:
* rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started messaging-0
* rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started messaging-1
* rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started messaging-2
* Container bundle set: galera-bundle [192.168.24.1:8787/rhosp12/openstack-mariadb-docker:pcmklatest]:
* galera-bundle-0 (ocf:heartbeat:galera): Promoted database-0
* galera-bundle-1 (ocf:heartbeat:galera): Promoted database-1
* galera-bundle-2 (ocf:heartbeat:galera): Promoted database-2
* Container bundle set: redis-bundle [192.168.24.1:8787/rhosp12/openstack-redis-docker:pcmklatest]:
* redis-bundle-0 (ocf:heartbeat:redis): Promoted controller-0
* redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1
* redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2
* ip-192.168.24.11 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.104 (ocf:heartbeat:IPaddr2): Started controller-1
* ip-172.17.1.19 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.11 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.3.13 (ocf:heartbeat:IPaddr2): Started controller-1
* ip-172.17.4.19 (ocf:heartbeat:IPaddr2): Started controller-2
* Container bundle set: haproxy-bundle [192.168.24.1:8787/rhosp12/openstack-haproxy-docker:pcmklatest]:
* haproxy-bundle-docker-0 (ocf:heartbeat:docker): Started controller-0
* haproxy-bundle-docker-1 (ocf:heartbeat:docker): Started controller-1
* haproxy-bundle-docker-2 (ocf:heartbeat:docker): Started controller-2
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400244e09 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400cdec10 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400c709f7 (stonith:fence_ipmilan): Started controller-1
* stonith-fence_ipmilan-525400a7f9e0 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400a25787 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-5254005ea387 (stonith:fence_ipmilan): Started controller-1
* stonith-fence_ipmilan-525400542c06 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400aac413 (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400498d34 (stonith:fence_ipmilan): Started controller-1
diff --git a/cts/scheduler/summary/colocation-influence.summary b/cts/scheduler/summary/colocation-influence.summary
index 3ea8b3f545..7fa4fcf0c2 100644
--- a/cts/scheduler/summary/colocation-influence.summary
+++ b/cts/scheduler/summary/colocation-influence.summary
@@ -1,170 +1,170 @@
Current cluster status:
* Node List:
* Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* GuestOnline: [ bundle10-0@rhel7-2 bundle10-1@rhel7-3 bundle11-0@rhel7-1 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-1
* rsc1a (ocf:pacemaker:Dummy): Started rhel7-2
* rsc1b (ocf:pacemaker:Dummy): Started rhel7-2
* rsc2a (ocf:pacemaker:Dummy): Started rhel7-4
* rsc2b (ocf:pacemaker:Dummy): Started rhel7-4
* rsc3a (ocf:pacemaker:Dummy): Stopped
* rsc3b (ocf:pacemaker:Dummy): Stopped
* rsc4a (ocf:pacemaker:Dummy): Started rhel7-3
* rsc4b (ocf:pacemaker:Dummy): Started rhel7-3
* rsc5a (ocf:pacemaker:Dummy): Started rhel7-1
* Resource Group: group5a:
* rsc5a1 (ocf:pacemaker:Dummy): Started rhel7-1
* rsc5a2 (ocf:pacemaker:Dummy): Started rhel7-1
* Resource Group: group6a:
* rsc6a1 (ocf:pacemaker:Dummy): Started rhel7-2
* rsc6a2 (ocf:pacemaker:Dummy): Started rhel7-2
* rsc6a (ocf:pacemaker:Dummy): Started rhel7-2
* Resource Group: group7a:
* rsc7a1 (ocf:pacemaker:Dummy): Started rhel7-3
* rsc7a2 (ocf:pacemaker:Dummy): Started rhel7-3
* Clone Set: rsc8a-clone [rsc8a]:
* Started: [ rhel7-1 rhel7-3 rhel7-4 ]
* Clone Set: rsc8b-clone [rsc8b]:
* Started: [ rhel7-1 rhel7-3 rhel7-4 ]
* rsc9a (ocf:pacemaker:Dummy): Started rhel7-4
* rsc9b (ocf:pacemaker:Dummy): Started rhel7-4
* rsc9c (ocf:pacemaker:Dummy): Started rhel7-4
* rsc10a (ocf:pacemaker:Dummy): Started rhel7-2
* rsc11a (ocf:pacemaker:Dummy): Started rhel7-1
* rsc12a (ocf:pacemaker:Dummy): Started rhel7-1
* rsc12b (ocf:pacemaker:Dummy): Started rhel7-1
* rsc12c (ocf:pacemaker:Dummy): Started rhel7-1
* Container bundle set: bundle10 [pcmktest:http]:
* bundle10-0 (192.168.122.131) (ocf:heartbeat:apache): Started rhel7-2
* bundle10-1 (192.168.122.132) (ocf:heartbeat:apache): Started rhel7-3
* Container bundle set: bundle11 [pcmktest:http]:
* bundle11-0 (192.168.122.134) (ocf:pacemaker:Dummy): Started rhel7-1
* bundle11-1 (192.168.122.135) (ocf:pacemaker:Dummy): Stopped
* rsc13a (ocf:pacemaker:Dummy): Started rhel7-3
* Clone Set: rsc13b-clone [rsc13b] (promotable):
* Promoted: [ rhel7-3 ]
* Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 ]
* Stopped: [ rhel7-5 ]
* rsc14b (ocf:pacemaker:Dummy): Started rhel7-4
* Clone Set: rsc14a-clone [rsc14a] (promotable):
* Promoted: [ rhel7-4 ]
* Unpromoted: [ rhel7-1 rhel7-2 rhel7-3 ]
* Stopped: [ rhel7-5 ]
Transition Summary:
* Move rsc1a ( rhel7-2 -> rhel7-3 )
* Move rsc1b ( rhel7-2 -> rhel7-3 )
* Stop rsc2a ( rhel7-4 ) due to node availability
* Start rsc3a ( rhel7-2 )
* Start rsc3b ( rhel7-2 )
* Stop rsc4a ( rhel7-3 ) due to node availability
* Stop rsc5a ( rhel7-1 ) due to node availability
* Stop rsc6a1 ( rhel7-2 ) due to node availability
* Stop rsc6a2 ( rhel7-2 ) due to node availability
* Stop rsc7a2 ( rhel7-3 ) due to node availability
* Stop rsc8a:1 ( rhel7-4 ) due to node availability
* Stop rsc9c ( rhel7-4 ) due to node availability
* Move rsc10a ( rhel7-2 -> rhel7-3 )
* Stop rsc12b ( rhel7-1 ) due to node availability
* Start bundle11-1 ( rhel7-5 ) due to unrunnable bundle11-docker-1 start (blocked)
* Start bundle11a:1 ( bundle11-1 ) due to unrunnable bundle11-docker-1 start (blocked)
* Stop rsc13a ( rhel7-3 ) due to node availability
- * Stop rsc14a:1 ( Promoted rhel7-4 ) due to node availability
+ * Stop rsc14a:1 ( Promoted rhel7-4 ) due to node availability
Executing Cluster Transition:
* Resource action: rsc1a stop on rhel7-2
* Resource action: rsc1b stop on rhel7-2
* Resource action: rsc2a stop on rhel7-4
* Resource action: rsc3a start on rhel7-2
* Resource action: rsc3b start on rhel7-2
* Resource action: rsc4a stop on rhel7-3
* Resource action: rsc5a stop on rhel7-1
* Pseudo action: group6a_stop_0
* Resource action: rsc6a2 stop on rhel7-2
* Pseudo action: group7a_stop_0
* Resource action: rsc7a2 stop on rhel7-3
* Pseudo action: rsc8a-clone_stop_0
* Resource action: rsc9c stop on rhel7-4
* Resource action: rsc10a stop on rhel7-2
* Resource action: rsc12b stop on rhel7-1
* Resource action: rsc13a stop on rhel7-3
* Pseudo action: rsc14a-clone_demote_0
* Pseudo action: bundle11_start_0
* Resource action: rsc1a start on rhel7-3
* Resource action: rsc1b start on rhel7-3
* Resource action: rsc3a monitor=10000 on rhel7-2
* Resource action: rsc3b monitor=10000 on rhel7-2
* Resource action: rsc6a1 stop on rhel7-2
* Pseudo action: group7a_stopped_0
* Resource action: rsc8a stop on rhel7-4
* Pseudo action: rsc8a-clone_stopped_0
* Resource action: rsc10a start on rhel7-3
* Pseudo action: bundle11-clone_start_0
* Resource action: rsc14a demote on rhel7-4
* Pseudo action: rsc14a-clone_demoted_0
* Pseudo action: rsc14a-clone_stop_0
* Resource action: rsc1a monitor=10000 on rhel7-3
* Resource action: rsc1b monitor=10000 on rhel7-3
* Pseudo action: group6a_stopped_0
* Resource action: rsc10a monitor=10000 on rhel7-3
* Pseudo action: bundle11-clone_running_0
* Resource action: rsc14a stop on rhel7-4
* Pseudo action: rsc14a-clone_stopped_0
* Pseudo action: bundle11_running_0
Revised Cluster Status:
* Node List:
* Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* GuestOnline: [ bundle10-0@rhel7-2 bundle10-1@rhel7-3 bundle11-0@rhel7-1 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-1
* rsc1a (ocf:pacemaker:Dummy): Started rhel7-3
* rsc1b (ocf:pacemaker:Dummy): Started rhel7-3
* rsc2a (ocf:pacemaker:Dummy): Stopped
* rsc2b (ocf:pacemaker:Dummy): Started rhel7-4
* rsc3a (ocf:pacemaker:Dummy): Started rhel7-2
* rsc3b (ocf:pacemaker:Dummy): Started rhel7-2
* rsc4a (ocf:pacemaker:Dummy): Stopped
* rsc4b (ocf:pacemaker:Dummy): Started rhel7-3
* rsc5a (ocf:pacemaker:Dummy): Stopped
* Resource Group: group5a:
* rsc5a1 (ocf:pacemaker:Dummy): Started rhel7-1
* rsc5a2 (ocf:pacemaker:Dummy): Started rhel7-1
* Resource Group: group6a:
* rsc6a1 (ocf:pacemaker:Dummy): Stopped
* rsc6a2 (ocf:pacemaker:Dummy): Stopped
* rsc6a (ocf:pacemaker:Dummy): Started rhel7-2
* Resource Group: group7a:
* rsc7a1 (ocf:pacemaker:Dummy): Started rhel7-3
* rsc7a2 (ocf:pacemaker:Dummy): Stopped
* Clone Set: rsc8a-clone [rsc8a]:
* Started: [ rhel7-1 rhel7-3 ]
* Stopped: [ rhel7-2 rhel7-4 rhel7-5 ]
* Clone Set: rsc8b-clone [rsc8b]:
* Started: [ rhel7-1 rhel7-3 rhel7-4 ]
* rsc9a (ocf:pacemaker:Dummy): Started rhel7-4
* rsc9b (ocf:pacemaker:Dummy): Started rhel7-4
* rsc9c (ocf:pacemaker:Dummy): Stopped
* rsc10a (ocf:pacemaker:Dummy): Started rhel7-3
* rsc11a (ocf:pacemaker:Dummy): Started rhel7-1
* rsc12a (ocf:pacemaker:Dummy): Started rhel7-1
* rsc12b (ocf:pacemaker:Dummy): Stopped
* rsc12c (ocf:pacemaker:Dummy): Started rhel7-1
* Container bundle set: bundle10 [pcmktest:http]:
* bundle10-0 (192.168.122.131) (ocf:heartbeat:apache): Started rhel7-2
* bundle10-1 (192.168.122.132) (ocf:heartbeat:apache): Started rhel7-3
* Container bundle set: bundle11 [pcmktest:http]:
* bundle11-0 (192.168.122.134) (ocf:pacemaker:Dummy): Started rhel7-1
* bundle11-1 (192.168.122.135) (ocf:pacemaker:Dummy): Stopped
* rsc13a (ocf:pacemaker:Dummy): Stopped
* Clone Set: rsc13b-clone [rsc13b] (promotable):
* Promoted: [ rhel7-3 ]
* Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 ]
* Stopped: [ rhel7-5 ]
* rsc14b (ocf:pacemaker:Dummy): Started rhel7-4
* Clone Set: rsc14a-clone [rsc14a] (promotable):
* Unpromoted: [ rhel7-1 rhel7-2 rhel7-3 ]
* Stopped: [ rhel7-4 rhel7-5 ]
diff --git a/cts/scheduler/summary/dc-fence-ordering.summary b/cts/scheduler/summary/dc-fence-ordering.summary
index ac46031f07..305ebd5c19 100644
--- a/cts/scheduler/summary/dc-fence-ordering.summary
+++ b/cts/scheduler/summary/dc-fence-ordering.summary
@@ -1,82 +1,82 @@
Using the original execution date of: 2018-11-28 18:37:16Z
Current cluster status:
* Node List:
* Node rhel7-1: UNCLEAN (online)
* Online: [ rhel7-2 rhel7-4 rhel7-5 ]
* OFFLINE: [ rhel7-3 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Stopped
* FencingPass (stonith:fence_dummy): Stopped
* FencingFail (stonith:fence_dummy): Stopped
* rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Stopped
* migrator (ocf:pacemaker:Dummy): Stopped
* Clone Set: Connectivity [ping-1]:
* Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* Clone Set: promotable-1 [stateful-1] (promotable):
* Promoted: [ rhel7-1 ]
* Unpromoted: [ rhel7-2 rhel7-4 rhel7-5 ]
* Stopped: [ rhel7-3 ]
* Resource Group: group-1:
* r192.168.122.207 (ocf:heartbeat:IPaddr2): Started rhel7-1
* petulant (service:pacemaker-cts-dummyd@10): FAILED rhel7-1
* r192.168.122.208 (ocf:heartbeat:IPaddr2): Stopped
* lsb-dummy (lsb:LSBDummy): Stopped
Transition Summary:
* Fence (reboot) rhel7-1 'petulant failed there'
- * Stop stateful-1:0 ( Unpromoted rhel7-5 ) due to node availability
- * Stop stateful-1:1 ( Promoted rhel7-1 ) due to node availability
- * Stop stateful-1:2 ( Unpromoted rhel7-2 ) due to node availability
- * Stop stateful-1:3 ( Unpromoted rhel7-4 ) due to node availability
+ * Stop stateful-1:0 ( Unpromoted rhel7-5 ) due to node availability
+ * Stop stateful-1:1 ( Promoted rhel7-1 ) due to node availability
+ * Stop stateful-1:2 ( Unpromoted rhel7-2 ) due to node availability
+ * Stop stateful-1:3 ( Unpromoted rhel7-4 ) due to node availability
* Stop r192.168.122.207 ( rhel7-1 ) due to node availability
* Stop petulant ( rhel7-1 ) due to node availability
Executing Cluster Transition:
* Fencing rhel7-1 (reboot)
* Pseudo action: group-1_stop_0
* Pseudo action: petulant_stop_0
* Pseudo action: r192.168.122.207_stop_0
* Pseudo action: group-1_stopped_0
* Pseudo action: promotable-1_demote_0
* Pseudo action: stateful-1_demote_0
* Pseudo action: promotable-1_demoted_0
* Pseudo action: promotable-1_stop_0
* Resource action: stateful-1 stop on rhel7-5
* Pseudo action: stateful-1_stop_0
* Resource action: stateful-1 stop on rhel7-2
* Resource action: stateful-1 stop on rhel7-4
* Pseudo action: promotable-1_stopped_0
* Cluster action: do_shutdown on rhel7-5
* Cluster action: do_shutdown on rhel7-4
* Cluster action: do_shutdown on rhel7-2
Using the original execution date of: 2018-11-28 18:37:16Z
Revised Cluster Status:
* Node List:
* Online: [ rhel7-2 rhel7-4 rhel7-5 ]
* OFFLINE: [ rhel7-1 rhel7-3 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Stopped
* FencingPass (stonith:fence_dummy): Stopped
* FencingFail (stonith:fence_dummy): Stopped
* rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Stopped
* rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Stopped
* migrator (ocf:pacemaker:Dummy): Stopped
* Clone Set: Connectivity [ping-1]:
* Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* Clone Set: promotable-1 [stateful-1] (promotable):
* Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* Resource Group: group-1:
* r192.168.122.207 (ocf:heartbeat:IPaddr2): Stopped
* petulant (service:pacemaker-cts-dummyd@10): Stopped
* r192.168.122.208 (ocf:heartbeat:IPaddr2): Stopped
* lsb-dummy (lsb:LSBDummy): Stopped
diff --git a/cts/scheduler/summary/guest-node-host-dies.summary b/cts/scheduler/summary/guest-node-host-dies.summary
index b0286b2846..f4509b9029 100644
--- a/cts/scheduler/summary/guest-node-host-dies.summary
+++ b/cts/scheduler/summary/guest-node-host-dies.summary
@@ -1,82 +1,82 @@
Current cluster status:
* Node List:
* Node rhel7-1: UNCLEAN (offline)
* Online: [ rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-4
* rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-1 (UNCLEAN)
* container1 (ocf:heartbeat:VirtualDomain): FAILED rhel7-1 (UNCLEAN)
* container2 (ocf:heartbeat:VirtualDomain): FAILED rhel7-1 (UNCLEAN)
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Stopped: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
Transition Summary:
* Fence (reboot) lxc2 (resource: container2) 'guest is unclean'
* Fence (reboot) lxc1 (resource: container1) 'guest is unclean'
* Fence (reboot) rhel7-1 'rsc_rhel7-1 is thought to be active there'
* Restart Fencing ( rhel7-4 ) due to resource definition change
* Move rsc_rhel7-1 ( rhel7-1 -> rhel7-5 )
* Recover container1 ( rhel7-1 -> rhel7-2 )
* Recover container2 ( rhel7-1 -> rhel7-3 )
- * Recover lxc-ms:0 ( Promoted lxc1 )
- * Recover lxc-ms:1 ( Unpromoted lxc2 )
+ * Recover lxc-ms:0 ( Promoted lxc1 )
+ * Recover lxc-ms:1 ( Unpromoted lxc2 )
* Move lxc1 ( rhel7-1 -> rhel7-2 )
* Move lxc2 ( rhel7-1 -> rhel7-3 )
Executing Cluster Transition:
* Resource action: Fencing stop on rhel7-4
* Pseudo action: lxc-ms-master_demote_0
* Pseudo action: lxc1_stop_0
* Resource action: lxc1 monitor on rhel7-5
* Resource action: lxc1 monitor on rhel7-4
* Resource action: lxc1 monitor on rhel7-3
* Pseudo action: lxc2_stop_0
* Resource action: lxc2 monitor on rhel7-5
* Resource action: lxc2 monitor on rhel7-4
* Resource action: lxc2 monitor on rhel7-2
* Fencing rhel7-1 (reboot)
* Pseudo action: rsc_rhel7-1_stop_0
* Pseudo action: container1_stop_0
* Pseudo action: container2_stop_0
* Pseudo action: stonith-lxc2-reboot on lxc2
* Pseudo action: stonith-lxc1-reboot on lxc1
* Resource action: Fencing start on rhel7-4
* Resource action: Fencing monitor=120000 on rhel7-4
* Resource action: rsc_rhel7-1 start on rhel7-5
* Resource action: container1 start on rhel7-2
* Resource action: container2 start on rhel7-3
* Pseudo action: lxc-ms_demote_0
* Pseudo action: lxc-ms-master_demoted_0
* Pseudo action: lxc-ms-master_stop_0
* Resource action: lxc1 start on rhel7-2
* Resource action: lxc2 start on rhel7-3
* Resource action: rsc_rhel7-1 monitor=5000 on rhel7-5
* Pseudo action: lxc-ms_stop_0
* Pseudo action: lxc-ms_stop_0
* Pseudo action: lxc-ms-master_stopped_0
* Pseudo action: lxc-ms-master_start_0
* Resource action: lxc1 monitor=30000 on rhel7-2
* Resource action: lxc2 monitor=30000 on rhel7-3
* Resource action: lxc-ms start on lxc1
* Resource action: lxc-ms start on lxc2
* Pseudo action: lxc-ms-master_running_0
* Resource action: lxc-ms monitor=10000 on lxc2
* Pseudo action: lxc-ms-master_promote_0
* Resource action: lxc-ms promote on lxc1
* Pseudo action: lxc-ms-master_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* OFFLINE: [ rhel7-1 ]
* GuestOnline: [ lxc1@rhel7-2 lxc2@rhel7-3 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-4
* rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-5
* container1 (ocf:heartbeat:VirtualDomain): Started rhel7-2
* container2 (ocf:heartbeat:VirtualDomain): Started rhel7-3
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Promoted: [ lxc1 ]
* Unpromoted: [ lxc2 ]
diff --git a/cts/scheduler/summary/migrate-fencing.summary b/cts/scheduler/summary/migrate-fencing.summary
index fd4fffa1d3..955bb0f434 100644
--- a/cts/scheduler/summary/migrate-fencing.summary
+++ b/cts/scheduler/summary/migrate-fencing.summary
@@ -1,108 +1,108 @@
Current cluster status:
* Node List:
* Node pcmk-4: UNCLEAN (online)
* Online: [ pcmk-1 pcmk-2 pcmk-3 ]
* Full List of Resources:
* Clone Set: Fencing [FencingChild]:
* Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Resource Group: group-1:
* r192.168.101.181 (ocf:heartbeat:IPaddr): Started pcmk-4
* r192.168.101.182 (ocf:heartbeat:IPaddr): Started pcmk-4
* r192.168.101.183 (ocf:heartbeat:IPaddr): Started pcmk-4
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-4
* migrator (ocf:pacemaker:Dummy): Started pcmk-1
* Clone Set: Connectivity [ping-1]:
* Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ pcmk-4 ]
* Unpromoted: [ pcmk-1 pcmk-2 pcmk-3 ]
Transition Summary:
* Fence (reboot) pcmk-4 'termination was requested'
* Stop FencingChild:0 ( pcmk-4 ) due to node availability
* Move r192.168.101.181 ( pcmk-4 -> pcmk-1 )
* Move r192.168.101.182 ( pcmk-4 -> pcmk-1 )
* Move r192.168.101.183 ( pcmk-4 -> pcmk-1 )
* Move rsc_pcmk-4 ( pcmk-4 -> pcmk-2 )
* Move lsb-dummy ( pcmk-4 -> pcmk-1 )
* Migrate migrator ( pcmk-1 -> pcmk-3 )
* Stop ping-1:0 ( pcmk-4 ) due to node availability
- * Stop stateful-1:0 ( Promoted pcmk-4 ) due to node availability
+ * Stop stateful-1:0 ( Promoted pcmk-4 ) due to node availability
* Promote stateful-1:1 ( Unpromoted -> Promoted pcmk-1 )
Executing Cluster Transition:
* Pseudo action: Fencing_stop_0
* Resource action: stateful-1:3 monitor=15000 on pcmk-3
* Resource action: stateful-1:2 monitor=15000 on pcmk-2
* Fencing pcmk-4 (reboot)
* Pseudo action: FencingChild:0_stop_0
* Pseudo action: Fencing_stopped_0
* Pseudo action: rsc_pcmk-4_stop_0
* Pseudo action: lsb-dummy_stop_0
* Resource action: migrator migrate_to on pcmk-1
* Pseudo action: Connectivity_stop_0
* Pseudo action: group-1_stop_0
* Pseudo action: r192.168.101.183_stop_0
* Resource action: rsc_pcmk-4 start on pcmk-2
* Resource action: migrator migrate_from on pcmk-3
* Resource action: migrator stop on pcmk-1
* Pseudo action: ping-1:0_stop_0
* Pseudo action: Connectivity_stopped_0
* Pseudo action: r192.168.101.182_stop_0
* Resource action: rsc_pcmk-4 monitor=5000 on pcmk-2
* Pseudo action: migrator_start_0
* Pseudo action: r192.168.101.181_stop_0
* Resource action: migrator monitor=10000 on pcmk-3
* Pseudo action: group-1_stopped_0
* Pseudo action: master-1_demote_0
* Pseudo action: stateful-1:0_demote_0
* Pseudo action: master-1_demoted_0
* Pseudo action: master-1_stop_0
* Pseudo action: stateful-1:0_stop_0
* Pseudo action: master-1_stopped_0
* Pseudo action: master-1_promote_0
* Resource action: stateful-1:1 promote on pcmk-1
* Pseudo action: master-1_promoted_0
* Pseudo action: group-1_start_0
* Resource action: r192.168.101.181 start on pcmk-1
* Resource action: r192.168.101.182 start on pcmk-1
* Resource action: r192.168.101.183 start on pcmk-1
* Resource action: stateful-1:1 monitor=16000 on pcmk-1
* Pseudo action: group-1_running_0
* Resource action: r192.168.101.181 monitor=5000 on pcmk-1
* Resource action: r192.168.101.182 monitor=5000 on pcmk-1
* Resource action: r192.168.101.183 monitor=5000 on pcmk-1
* Resource action: lsb-dummy start on pcmk-1
* Resource action: lsb-dummy monitor=5000 on pcmk-1
Revised Cluster Status:
* Node List:
* Online: [ pcmk-1 pcmk-2 pcmk-3 ]
* OFFLINE: [ pcmk-4 ]
* Full List of Resources:
* Clone Set: Fencing [FencingChild]:
* Started: [ pcmk-1 pcmk-2 pcmk-3 ]
* Stopped: [ pcmk-4 ]
* Resource Group: group-1:
* r192.168.101.181 (ocf:heartbeat:IPaddr): Started pcmk-1
* r192.168.101.182 (ocf:heartbeat:IPaddr): Started pcmk-1
* r192.168.101.183 (ocf:heartbeat:IPaddr): Started pcmk-1
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-2
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-1
* migrator (ocf:pacemaker:Dummy): Started pcmk-3
* Clone Set: Connectivity [ping-1]:
* Started: [ pcmk-1 pcmk-2 pcmk-3 ]
* Stopped: [ pcmk-4 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ pcmk-1 ]
* Unpromoted: [ pcmk-2 pcmk-3 ]
* Stopped: [ pcmk-4 ]
diff --git a/cts/scheduler/summary/migrate-shutdown.summary b/cts/scheduler/summary/migrate-shutdown.summary
index 551a41a175..1da9db21e8 100644
--- a/cts/scheduler/summary/migrate-shutdown.summary
+++ b/cts/scheduler/summary/migrate-shutdown.summary
@@ -1,92 +1,92 @@
Current cluster status:
* Node List:
* Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started pcmk-1
* Resource Group: group-1:
* r192.168.122.105 (ocf:heartbeat:IPaddr): Started pcmk-2
* r192.168.122.106 (ocf:heartbeat:IPaddr): Started pcmk-2
* r192.168.122.107 (ocf:heartbeat:IPaddr): Started pcmk-2
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Stopped
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-2
* migrator (ocf:pacemaker:Dummy): Started pcmk-1
* Clone Set: Connectivity [ping-1]:
* Started: [ pcmk-1 pcmk-2 pcmk-4 ]
* Stopped: [ pcmk-3 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ pcmk-2 ]
* Unpromoted: [ pcmk-1 pcmk-4 ]
* Stopped: [ pcmk-3 ]
Transition Summary:
* Stop Fencing ( pcmk-1 ) due to node availability
* Stop r192.168.122.105 ( pcmk-2 ) due to node availability
* Stop r192.168.122.106 ( pcmk-2 ) due to node availability
* Stop r192.168.122.107 ( pcmk-2 ) due to node availability
* Stop rsc_pcmk-1 ( pcmk-1 ) due to node availability
* Stop rsc_pcmk-2 ( pcmk-2 ) due to node availability
* Stop rsc_pcmk-4 ( pcmk-4 ) due to node availability
* Stop lsb-dummy ( pcmk-2 ) due to node availability
* Stop migrator ( pcmk-1 ) due to node availability
* Stop ping-1:0 ( pcmk-1 ) due to node availability
* Stop ping-1:1 ( pcmk-2 ) due to node availability
* Stop ping-1:2 ( pcmk-4 ) due to node availability
- * Stop stateful-1:0 ( Unpromoted pcmk-1 ) due to node availability
- * Stop stateful-1:1 ( Promoted pcmk-2 ) due to node availability
- * Stop stateful-1:2 ( Unpromoted pcmk-4 ) due to node availability
+ * Stop stateful-1:0 ( Unpromoted pcmk-1 ) due to node availability
+ * Stop stateful-1:1 ( Promoted pcmk-2 ) due to node availability
+ * Stop stateful-1:2 ( Unpromoted pcmk-4 ) due to node availability
Executing Cluster Transition:
* Resource action: Fencing stop on pcmk-1
* Resource action: rsc_pcmk-1 stop on pcmk-1
* Resource action: rsc_pcmk-2 stop on pcmk-2
* Resource action: rsc_pcmk-4 stop on pcmk-4
* Resource action: lsb-dummy stop on pcmk-2
* Resource action: migrator stop on pcmk-1
* Resource action: migrator stop on pcmk-3
* Pseudo action: Connectivity_stop_0
* Cluster action: do_shutdown on pcmk-3
* Pseudo action: group-1_stop_0
* Resource action: r192.168.122.107 stop on pcmk-2
* Resource action: ping-1:0 stop on pcmk-1
* Resource action: ping-1:1 stop on pcmk-2
* Resource action: ping-1:3 stop on pcmk-4
* Pseudo action: Connectivity_stopped_0
* Resource action: r192.168.122.106 stop on pcmk-2
* Resource action: r192.168.122.105 stop on pcmk-2
* Pseudo action: group-1_stopped_0
* Pseudo action: master-1_demote_0
* Resource action: stateful-1:0 demote on pcmk-2
* Pseudo action: master-1_demoted_0
* Pseudo action: master-1_stop_0
* Resource action: stateful-1:2 stop on pcmk-1
* Resource action: stateful-1:0 stop on pcmk-2
* Resource action: stateful-1:3 stop on pcmk-4
* Pseudo action: master-1_stopped_0
* Cluster action: do_shutdown on pcmk-4
* Cluster action: do_shutdown on pcmk-2
* Cluster action: do_shutdown on pcmk-1
Revised Cluster Status:
* Node List:
* Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Stopped
* Resource Group: group-1:
* r192.168.122.105 (ocf:heartbeat:IPaddr): Stopped
* r192.168.122.106 (ocf:heartbeat:IPaddr): Stopped
* r192.168.122.107 (ocf:heartbeat:IPaddr): Stopped
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Stopped
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Stopped
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Stopped
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Stopped
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Stopped
* migrator (ocf:pacemaker:Dummy): Stopped
* Clone Set: Connectivity [ping-1]:
* Stopped: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Stopped: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
diff --git a/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary b/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary
index 532f731235..8eb68a4cb9 100644
--- a/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary
+++ b/cts/scheduler/summary/no-promote-on-unrunnable-guest.summary
@@ -1,103 +1,103 @@
Using the original execution date of: 2020-05-14 10:49:31Z
Current cluster status:
* Node List:
* Online: [ controller-0 controller-1 controller-2 ]
* GuestOnline: [ galera-bundle-0@controller-0 galera-bundle-1@controller-1 galera-bundle-2@controller-2 ovn-dbs-bundle-0@controller-0 ovn-dbs-bundle-1@controller-1 ovn-dbs-bundle-2@controller-2 rabbitmq-bundle-0@controller-0 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ]
* Full List of Resources:
* Container bundle set: galera-bundle [cluster.common.tag/rhosp16-openstack-mariadb:pcmklatest]:
* galera-bundle-0 (ocf:heartbeat:galera): Promoted controller-0
* galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1
* galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2
* Container bundle set: rabbitmq-bundle [cluster.common.tag/rhosp16-openstack-rabbitmq:pcmklatest]:
* rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started controller-0
* rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1
* rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2
* Container bundle set: redis-bundle [cluster.common.tag/rhosp16-openstack-redis:pcmklatest]:
* redis-bundle-0 (ocf:heartbeat:redis): Promoted controller-0
* redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1
* redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2
* Container bundle set: ovn-dbs-bundle [cluster.common.tag/rhosp16-openstack-ovn-northd:pcmklatest]:
* ovn-dbs-bundle-0 (ocf:ovn:ovndb-servers): Unpromoted controller-0
* ovn-dbs-bundle-1 (ocf:ovn:ovndb-servers): Unpromoted controller-1
* ovn-dbs-bundle-2 (ocf:ovn:ovndb-servers): Unpromoted controller-2
* stonith-fence_ipmilan-5254005e097a (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400afe30e (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400985679 (stonith:fence_ipmilan): Started controller-1
* Container bundle: openstack-cinder-volume [cluster.common.tag/rhosp16-openstack-cinder-volume:pcmklatest]:
* openstack-cinder-volume-podman-0 (ocf:heartbeat:podman): Started controller-0
Transition Summary:
* Stop ovn-dbs-bundle-podman-0 ( controller-0 ) due to node availability
* Stop ovn-dbs-bundle-0 ( controller-0 ) due to unrunnable ovn-dbs-bundle-podman-0 start
- * Stop ovndb_servers:0 ( Unpromoted ovn-dbs-bundle-0 ) due to unrunnable ovn-dbs-bundle-podman-0 start
+ * Stop ovndb_servers:0 ( Unpromoted ovn-dbs-bundle-0 ) due to unrunnable ovn-dbs-bundle-podman-0 start
* Promote ovndb_servers:1 ( Unpromoted -> Promoted ovn-dbs-bundle-1 )
Executing Cluster Transition:
* Resource action: ovndb_servers cancel=30000 on ovn-dbs-bundle-1
* Pseudo action: ovn-dbs-bundle-master_pre_notify_stop_0
* Pseudo action: ovn-dbs-bundle_stop_0
* Resource action: ovndb_servers notify on ovn-dbs-bundle-0
* Resource action: ovndb_servers notify on ovn-dbs-bundle-1
* Resource action: ovndb_servers notify on ovn-dbs-bundle-2
* Pseudo action: ovn-dbs-bundle-master_confirmed-pre_notify_stop_0
* Pseudo action: ovn-dbs-bundle-master_stop_0
* Resource action: ovndb_servers stop on ovn-dbs-bundle-0
* Pseudo action: ovn-dbs-bundle-master_stopped_0
* Resource action: ovn-dbs-bundle-0 stop on controller-0
* Pseudo action: ovn-dbs-bundle-master_post_notify_stopped_0
* Resource action: ovn-dbs-bundle-podman-0 stop on controller-0
* Resource action: ovndb_servers notify on ovn-dbs-bundle-1
* Resource action: ovndb_servers notify on ovn-dbs-bundle-2
* Pseudo action: ovn-dbs-bundle-master_confirmed-post_notify_stopped_0
* Pseudo action: ovn-dbs-bundle-master_pre_notify_start_0
* Pseudo action: ovn-dbs-bundle_stopped_0
* Pseudo action: ovn-dbs-bundle-master_confirmed-pre_notify_start_0
* Pseudo action: ovn-dbs-bundle-master_start_0
* Pseudo action: ovn-dbs-bundle-master_running_0
* Pseudo action: ovn-dbs-bundle-master_post_notify_running_0
* Pseudo action: ovn-dbs-bundle-master_confirmed-post_notify_running_0
* Pseudo action: ovn-dbs-bundle_running_0
* Pseudo action: ovn-dbs-bundle-master_pre_notify_promote_0
* Pseudo action: ovn-dbs-bundle_promote_0
* Resource action: ovndb_servers notify on ovn-dbs-bundle-1
* Resource action: ovndb_servers notify on ovn-dbs-bundle-2
* Pseudo action: ovn-dbs-bundle-master_confirmed-pre_notify_promote_0
* Pseudo action: ovn-dbs-bundle-master_promote_0
* Resource action: ovndb_servers promote on ovn-dbs-bundle-1
* Pseudo action: ovn-dbs-bundle-master_promoted_0
* Pseudo action: ovn-dbs-bundle-master_post_notify_promoted_0
* Resource action: ovndb_servers notify on ovn-dbs-bundle-1
* Resource action: ovndb_servers notify on ovn-dbs-bundle-2
* Pseudo action: ovn-dbs-bundle-master_confirmed-post_notify_promoted_0
* Pseudo action: ovn-dbs-bundle_promoted_0
* Resource action: ovndb_servers monitor=10000 on ovn-dbs-bundle-1
Using the original execution date of: 2020-05-14 10:49:31Z
Revised Cluster Status:
* Node List:
* Online: [ controller-0 controller-1 controller-2 ]
* GuestOnline: [ galera-bundle-0@controller-0 galera-bundle-1@controller-1 galera-bundle-2@controller-2 ovn-dbs-bundle-1@controller-1 ovn-dbs-bundle-2@controller-2 rabbitmq-bundle-0@controller-0 rabbitmq-bundle-1@controller-1 rabbitmq-bundle-2@controller-2 redis-bundle-0@controller-0 redis-bundle-1@controller-1 redis-bundle-2@controller-2 ]
* Full List of Resources:
* Container bundle set: galera-bundle [cluster.common.tag/rhosp16-openstack-mariadb:pcmklatest]:
* galera-bundle-0 (ocf:heartbeat:galera): Promoted controller-0
* galera-bundle-1 (ocf:heartbeat:galera): Promoted controller-1
* galera-bundle-2 (ocf:heartbeat:galera): Promoted controller-2
* Container bundle set: rabbitmq-bundle [cluster.common.tag/rhosp16-openstack-rabbitmq:pcmklatest]:
* rabbitmq-bundle-0 (ocf:heartbeat:rabbitmq-cluster): Started controller-0
* rabbitmq-bundle-1 (ocf:heartbeat:rabbitmq-cluster): Started controller-1
* rabbitmq-bundle-2 (ocf:heartbeat:rabbitmq-cluster): Started controller-2
* Container bundle set: redis-bundle [cluster.common.tag/rhosp16-openstack-redis:pcmklatest]:
* redis-bundle-0 (ocf:heartbeat:redis): Promoted controller-0
* redis-bundle-1 (ocf:heartbeat:redis): Unpromoted controller-1
* redis-bundle-2 (ocf:heartbeat:redis): Unpromoted controller-2
* Container bundle set: ovn-dbs-bundle [cluster.common.tag/rhosp16-openstack-ovn-northd:pcmklatest]:
* ovn-dbs-bundle-0 (ocf:ovn:ovndb-servers): Stopped
* ovn-dbs-bundle-1 (ocf:ovn:ovndb-servers): Promoted controller-1
* ovn-dbs-bundle-2 (ocf:ovn:ovndb-servers): Unpromoted controller-2
* stonith-fence_ipmilan-5254005e097a (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400afe30e (stonith:fence_ipmilan): Started controller-2
* stonith-fence_ipmilan-525400985679 (stonith:fence_ipmilan): Started controller-1
* Container bundle: openstack-cinder-volume [cluster.common.tag/rhosp16-openstack-cinder-volume:pcmklatest]:
* openstack-cinder-volume-podman-0 (ocf:heartbeat:podman): Started controller-0
diff --git a/cts/scheduler/summary/novell-239082.summary b/cts/scheduler/summary/novell-239082.summary
index 431b6ddc63..01af7656e9 100644
--- a/cts/scheduler/summary/novell-239082.summary
+++ b/cts/scheduler/summary/novell-239082.summary
@@ -1,59 +1,59 @@
Current cluster status:
* Node List:
* Online: [ xen-1 xen-2 ]
* Full List of Resources:
* fs_1 (ocf:heartbeat:Filesystem): Started xen-1
* Clone Set: ms-drbd0 [drbd0] (promotable):
* Promoted: [ xen-1 ]
* Unpromoted: [ xen-2 ]
Transition Summary:
* Move fs_1 ( xen-1 -> xen-2 )
* Promote drbd0:0 ( Unpromoted -> Promoted xen-2 )
- * Stop drbd0:1 ( Promoted xen-1 ) due to node availability
+ * Stop drbd0:1 ( Promoted xen-1 ) due to node availability
Executing Cluster Transition:
* Resource action: fs_1 stop on xen-1
* Pseudo action: ms-drbd0_pre_notify_demote_0
* Resource action: drbd0:0 notify on xen-2
* Resource action: drbd0:1 notify on xen-1
* Pseudo action: ms-drbd0_confirmed-pre_notify_demote_0
* Pseudo action: ms-drbd0_demote_0
* Resource action: drbd0:1 demote on xen-1
* Pseudo action: ms-drbd0_demoted_0
* Pseudo action: ms-drbd0_post_notify_demoted_0
* Resource action: drbd0:0 notify on xen-2
* Resource action: drbd0:1 notify on xen-1
* Pseudo action: ms-drbd0_confirmed-post_notify_demoted_0
* Pseudo action: ms-drbd0_pre_notify_stop_0
* Resource action: drbd0:0 notify on xen-2
* Resource action: drbd0:1 notify on xen-1
* Pseudo action: ms-drbd0_confirmed-pre_notify_stop_0
* Pseudo action: ms-drbd0_stop_0
* Resource action: drbd0:1 stop on xen-1
* Pseudo action: ms-drbd0_stopped_0
* Cluster action: do_shutdown on xen-1
* Pseudo action: ms-drbd0_post_notify_stopped_0
* Resource action: drbd0:0 notify on xen-2
* Pseudo action: ms-drbd0_confirmed-post_notify_stopped_0
* Pseudo action: ms-drbd0_pre_notify_promote_0
* Resource action: drbd0:0 notify on xen-2
* Pseudo action: ms-drbd0_confirmed-pre_notify_promote_0
* Pseudo action: ms-drbd0_promote_0
* Resource action: drbd0:0 promote on xen-2
* Pseudo action: ms-drbd0_promoted_0
* Pseudo action: ms-drbd0_post_notify_promoted_0
* Resource action: drbd0:0 notify on xen-2
* Pseudo action: ms-drbd0_confirmed-post_notify_promoted_0
* Resource action: fs_1 start on xen-2
Revised Cluster Status:
* Node List:
* Online: [ xen-1 xen-2 ]
* Full List of Resources:
* fs_1 (ocf:heartbeat:Filesystem): Started xen-2
* Clone Set: ms-drbd0 [drbd0] (promotable):
* Promoted: [ xen-2 ]
* Stopped: [ xen-1 ]
diff --git a/cts/scheduler/summary/on_fail_demote4.summary b/cts/scheduler/summary/on_fail_demote4.summary
index 781f5488bb..b7b1388e58 100644
--- a/cts/scheduler/summary/on_fail_demote4.summary
+++ b/cts/scheduler/summary/on_fail_demote4.summary
@@ -1,189 +1,189 @@
Using the original execution date of: 2020-06-16 19:23:21Z
Current cluster status:
* Node List:
* RemoteNode remote-rhel7-2: UNCLEAN (offline)
* Node rhel7-4: UNCLEAN (offline)
* Online: [ rhel7-1 rhel7-3 rhel7-5 ]
* GuestOnline: [ lxc1@rhel7-3 stateful-bundle-1@rhel7-1 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-4 (UNCLEAN)
* Clone Set: rsc1-clone [rsc1] (promotable):
* rsc1 (ocf:pacemaker:Stateful): Promoted rhel7-4 (UNCLEAN)
* rsc1 (ocf:pacemaker:Stateful): Unpromoted remote-rhel7-2 (UNCLEAN)
* Unpromoted: [ lxc1 rhel7-1 rhel7-3 rhel7-5 ]
* Clone Set: rsc2-master [rsc2] (promotable):
* rsc2 (ocf:pacemaker:Stateful): Unpromoted rhel7-4 (UNCLEAN)
* rsc2 (ocf:pacemaker:Stateful): Promoted remote-rhel7-2 (UNCLEAN)
* Unpromoted: [ lxc1 rhel7-1 rhel7-3 rhel7-5 ]
* remote-rhel7-2 (ocf:pacemaker:remote): FAILED rhel7-1
* container1 (ocf:heartbeat:VirtualDomain): Started rhel7-3
* container2 (ocf:heartbeat:VirtualDomain): FAILED rhel7-3
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Unpromoted: [ lxc1 ]
* Stopped: [ remote-rhel7-2 rhel7-1 rhel7-3 rhel7-4 rhel7-5 ]
* Container bundle set: stateful-bundle [pcmktest:http]:
* stateful-bundle-0 (192.168.122.131) (ocf:pacemaker:Stateful): FAILED Promoted rhel7-5
* stateful-bundle-1 (192.168.122.132) (ocf:pacemaker:Stateful): Unpromoted rhel7-1
* stateful-bundle-2 (192.168.122.133) (ocf:pacemaker:Stateful): FAILED rhel7-4 (UNCLEAN)
Transition Summary:
* Fence (reboot) stateful-bundle-2 (resource: stateful-bundle-docker-2) 'guest is unclean'
* Fence (reboot) stateful-bundle-0 (resource: stateful-bundle-docker-0) 'guest is unclean'
* Fence (reboot) lxc2 (resource: container2) 'guest is unclean'
* Fence (reboot) remote-rhel7-2 'remote connection is unrecoverable'
* Fence (reboot) rhel7-4 'peer is no longer part of the cluster'
* Move Fencing ( rhel7-4 -> rhel7-5 )
- * Stop rsc1:0 ( Promoted rhel7-4 ) due to node availability
- * Promote rsc1:1 ( Unpromoted -> Promoted rhel7-3 )
- * Stop rsc1:4 ( Unpromoted remote-rhel7-2 ) due to node availability
- * Recover rsc1:5 ( Unpromoted lxc2 )
- * Stop rsc2:0 ( Unpromoted rhel7-4 ) due to node availability
- * Promote rsc2:1 ( Unpromoted -> Promoted rhel7-3 )
- * Stop rsc2:4 ( Promoted remote-rhel7-2 ) due to node availability
- * Recover rsc2:5 ( Unpromoted lxc2 )
+ * Stop rsc1:0 ( Promoted rhel7-4 ) due to node availability
+ * Promote rsc1:1 ( Unpromoted -> Promoted rhel7-3 )
+ * Stop rsc1:4 ( Unpromoted remote-rhel7-2 ) due to node availability
+ * Recover rsc1:5 ( Unpromoted lxc2 )
+ * Stop rsc2:0 ( Unpromoted rhel7-4 ) due to node availability
+ * Promote rsc2:1 ( Unpromoted -> Promoted rhel7-3 )
+ * Stop rsc2:4 ( Promoted remote-rhel7-2 ) due to node availability
+ * Recover rsc2:5 ( Unpromoted lxc2 )
* Recover remote-rhel7-2 ( rhel7-1 )
* Recover container2 ( rhel7-3 )
- * Recover lxc-ms:0 ( Promoted lxc2 )
+ * Recover lxc-ms:0 ( Promoted lxc2 )
* Recover stateful-bundle-docker-0 ( rhel7-5 )
* Restart stateful-bundle-0 ( rhel7-5 ) due to required stateful-bundle-docker-0 start
- * Recover bundled:0 ( Promoted stateful-bundle-0 )
+ * Recover bundled:0 ( Promoted stateful-bundle-0 )
* Move stateful-bundle-ip-192.168.122.133 ( rhel7-4 -> rhel7-3 )
* Recover stateful-bundle-docker-2 ( rhel7-4 -> rhel7-3 )
* Move stateful-bundle-2 ( rhel7-4 -> rhel7-3 )
- * Recover bundled:2 ( Unpromoted stateful-bundle-2 )
+ * Recover bundled:2 ( Unpromoted stateful-bundle-2 )
* Restart lxc2 ( rhel7-3 ) due to required container2 start
Executing Cluster Transition:
* Pseudo action: Fencing_stop_0
* Resource action: rsc1 cancel=11000 on rhel7-3
* Pseudo action: rsc1-clone_demote_0
* Resource action: rsc2 cancel=11000 on rhel7-3
* Pseudo action: rsc2-master_demote_0
* Pseudo action: lxc-ms-master_demote_0
* Resource action: stateful-bundle-0 stop on rhel7-5
* Pseudo action: stateful-bundle-2_stop_0
* Resource action: lxc2 stop on rhel7-3
* Pseudo action: stateful-bundle_demote_0
* Fencing remote-rhel7-2 (reboot)
* Fencing rhel7-4 (reboot)
* Pseudo action: rsc1_demote_0
* Pseudo action: rsc1-clone_demoted_0
* Pseudo action: rsc2_demote_0
* Pseudo action: rsc2-master_demoted_0
* Resource action: container2 stop on rhel7-3
* Pseudo action: stateful-bundle-master_demote_0
* Pseudo action: stonith-stateful-bundle-2-reboot on stateful-bundle-2
* Pseudo action: stonith-lxc2-reboot on lxc2
* Resource action: Fencing start on rhel7-5
* Pseudo action: rsc1-clone_stop_0
* Pseudo action: rsc2-master_stop_0
* Pseudo action: lxc-ms_demote_0
* Pseudo action: lxc-ms-master_demoted_0
* Pseudo action: lxc-ms-master_stop_0
* Pseudo action: bundled_demote_0
* Pseudo action: stateful-bundle-master_demoted_0
* Pseudo action: stateful-bundle_demoted_0
* Pseudo action: stateful-bundle_stop_0
* Resource action: Fencing monitor=120000 on rhel7-5
* Pseudo action: rsc1_stop_0
* Pseudo action: rsc1_stop_0
* Pseudo action: rsc1_stop_0
* Pseudo action: rsc1-clone_stopped_0
* Pseudo action: rsc1-clone_start_0
* Pseudo action: rsc2_stop_0
* Pseudo action: rsc2_stop_0
* Pseudo action: rsc2_stop_0
* Pseudo action: rsc2-master_stopped_0
* Pseudo action: rsc2-master_start_0
* Resource action: remote-rhel7-2 stop on rhel7-1
* Pseudo action: lxc-ms_stop_0
* Pseudo action: lxc-ms-master_stopped_0
* Pseudo action: lxc-ms-master_start_0
* Resource action: stateful-bundle-docker-0 stop on rhel7-5
* Pseudo action: stateful-bundle-docker-2_stop_0
* Pseudo action: stonith-stateful-bundle-0-reboot on stateful-bundle-0
* Resource action: remote-rhel7-2 start on rhel7-1
* Resource action: remote-rhel7-2 monitor=60000 on rhel7-1
* Resource action: container2 start on rhel7-3
* Resource action: container2 monitor=20000 on rhel7-3
* Pseudo action: stateful-bundle-master_stop_0
* Pseudo action: stateful-bundle-ip-192.168.122.133_stop_0
* Resource action: lxc2 start on rhel7-3
* Resource action: lxc2 monitor=30000 on rhel7-3
* Resource action: rsc1 start on lxc2
* Pseudo action: rsc1-clone_running_0
* Resource action: rsc2 start on lxc2
* Pseudo action: rsc2-master_running_0
* Resource action: lxc-ms start on lxc2
* Pseudo action: lxc-ms-master_running_0
* Pseudo action: bundled_stop_0
* Resource action: stateful-bundle-ip-192.168.122.133 start on rhel7-3
* Resource action: rsc1 monitor=11000 on lxc2
* Pseudo action: rsc1-clone_promote_0
* Resource action: rsc2 monitor=11000 on lxc2
* Pseudo action: rsc2-master_promote_0
* Pseudo action: lxc-ms-master_promote_0
* Pseudo action: bundled_stop_0
* Pseudo action: stateful-bundle-master_stopped_0
* Resource action: stateful-bundle-ip-192.168.122.133 monitor=60000 on rhel7-3
* Pseudo action: stateful-bundle_stopped_0
* Pseudo action: stateful-bundle_start_0
* Resource action: rsc1 promote on rhel7-3
* Pseudo action: rsc1-clone_promoted_0
* Resource action: rsc2 promote on rhel7-3
* Pseudo action: rsc2-master_promoted_0
* Resource action: lxc-ms promote on lxc2
* Pseudo action: lxc-ms-master_promoted_0
* Pseudo action: stateful-bundle-master_start_0
* Resource action: stateful-bundle-docker-0 start on rhel7-5
* Resource action: stateful-bundle-docker-0 monitor=60000 on rhel7-5
* Resource action: stateful-bundle-0 start on rhel7-5
* Resource action: stateful-bundle-0 monitor=30000 on rhel7-5
* Resource action: stateful-bundle-docker-2 start on rhel7-3
* Resource action: stateful-bundle-2 start on rhel7-3
* Resource action: rsc1 monitor=10000 on rhel7-3
* Resource action: rsc2 monitor=10000 on rhel7-3
* Resource action: lxc-ms monitor=10000 on lxc2
* Resource action: bundled start on stateful-bundle-0
* Resource action: bundled start on stateful-bundle-2
* Pseudo action: stateful-bundle-master_running_0
* Resource action: stateful-bundle-docker-2 monitor=60000 on rhel7-3
* Resource action: stateful-bundle-2 monitor=30000 on rhel7-3
* Pseudo action: stateful-bundle_running_0
* Resource action: bundled monitor=11000 on stateful-bundle-2
* Pseudo action: stateful-bundle_promote_0
* Pseudo action: stateful-bundle-master_promote_0
* Resource action: bundled promote on stateful-bundle-0
* Pseudo action: stateful-bundle-master_promoted_0
* Pseudo action: stateful-bundle_promoted_0
* Resource action: bundled monitor=10000 on stateful-bundle-0
Using the original execution date of: 2020-06-16 19:23:21Z
Revised Cluster Status:
* Node List:
* Online: [ rhel7-1 rhel7-3 rhel7-5 ]
* OFFLINE: [ rhel7-4 ]
* RemoteOnline: [ remote-rhel7-2 ]
* GuestOnline: [ lxc1@rhel7-3 lxc2@rhel7-3 stateful-bundle-0@rhel7-5 stateful-bundle-1@rhel7-1 stateful-bundle-2@rhel7-3 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-5
* Clone Set: rsc1-clone [rsc1] (promotable):
* Promoted: [ rhel7-3 ]
* Unpromoted: [ lxc1 lxc2 rhel7-1 rhel7-5 ]
* Stopped: [ remote-rhel7-2 rhel7-4 ]
* Clone Set: rsc2-master [rsc2] (promotable):
* Promoted: [ rhel7-3 ]
* Unpromoted: [ lxc1 lxc2 rhel7-1 rhel7-5 ]
* Stopped: [ remote-rhel7-2 rhel7-4 ]
* remote-rhel7-2 (ocf:pacemaker:remote): Started rhel7-1
* container1 (ocf:heartbeat:VirtualDomain): Started rhel7-3
* container2 (ocf:heartbeat:VirtualDomain): Started rhel7-3
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Promoted: [ lxc2 ]
* Unpromoted: [ lxc1 ]
* Container bundle set: stateful-bundle [pcmktest:http]:
* stateful-bundle-0 (192.168.122.131) (ocf:pacemaker:Stateful): Promoted rhel7-5
* stateful-bundle-1 (192.168.122.132) (ocf:pacemaker:Stateful): Unpromoted rhel7-1
* stateful-bundle-2 (192.168.122.133) (ocf:pacemaker:Stateful): Unpromoted rhel7-3
diff --git a/cts/scheduler/summary/probe-2.summary b/cts/scheduler/summary/probe-2.summary
index f2c60821ab..3523891d30 100644
--- a/cts/scheduler/summary/probe-2.summary
+++ b/cts/scheduler/summary/probe-2.summary
@@ -1,163 +1,163 @@
Current cluster status:
* Node List:
* Node wc02: standby (with active resources)
* Online: [ wc01 ]
* Full List of Resources:
* Resource Group: group_www_data:
* fs_www_data (ocf:heartbeat:Filesystem): Started wc01
* nfs-kernel-server (lsb:nfs-kernel-server): Started wc01
* intip_nfs (ocf:heartbeat:IPaddr2): Started wc01
* Clone Set: ms_drbd_mysql [drbd_mysql] (promotable):
* Promoted: [ wc02 ]
* Unpromoted: [ wc01 ]
* Resource Group: group_mysql:
* fs_mysql (ocf:heartbeat:Filesystem): Started wc02
* intip_sql (ocf:heartbeat:IPaddr2): Started wc02
* mysql-server (ocf:heartbeat:mysql): Started wc02
* Clone Set: ms_drbd_www [drbd_www] (promotable):
* Promoted: [ wc01 ]
* Unpromoted: [ wc02 ]
* Clone Set: clone_nfs-common [group_nfs-common]:
* Started: [ wc01 wc02 ]
* Clone Set: clone_mysql-proxy [group_mysql-proxy]:
* Started: [ wc01 wc02 ]
* Clone Set: clone_webservice [group_webservice]:
* Started: [ wc01 wc02 ]
* Resource Group: group_ftpd:
* extip_ftp (ocf:heartbeat:IPaddr2): Started wc01
* pure-ftpd (ocf:heartbeat:Pure-FTPd): Started wc01
* Clone Set: DoFencing [stonith_rackpdu] (unique):
* stonith_rackpdu:0 (stonith:external/rackpdu): Started wc01
* stonith_rackpdu:1 (stonith:external/rackpdu): Started wc02
Transition Summary:
* Promote drbd_mysql:0 ( Unpromoted -> Promoted wc01 )
- * Stop drbd_mysql:1 ( Promoted wc02 ) due to node availability
+ * Stop drbd_mysql:1 ( Promoted wc02 ) due to node availability
* Move fs_mysql ( wc02 -> wc01 )
* Move intip_sql ( wc02 -> wc01 )
* Move mysql-server ( wc02 -> wc01 )
- * Stop drbd_www:1 ( Unpromoted wc02 ) due to node availability
+ * Stop drbd_www:1 ( Unpromoted wc02 ) due to node availability
* Stop nfs-common:1 ( wc02 ) due to node availability
* Stop mysql-proxy:1 ( wc02 ) due to node availability
* Stop fs_www:1 ( wc02 ) due to node availability
* Stop apache2:1 ( wc02 ) due to node availability
* Restart stonith_rackpdu:0 ( wc01 )
* Stop stonith_rackpdu:1 ( wc02 ) due to node availability
Executing Cluster Transition:
* Resource action: drbd_mysql:0 cancel=10000 on wc01
* Pseudo action: ms_drbd_mysql_pre_notify_demote_0
* Pseudo action: group_mysql_stop_0
* Resource action: mysql-server stop on wc02
* Pseudo action: ms_drbd_www_pre_notify_stop_0
* Pseudo action: clone_mysql-proxy_stop_0
* Pseudo action: clone_webservice_stop_0
* Pseudo action: DoFencing_stop_0
* Resource action: drbd_mysql:0 notify on wc01
* Resource action: drbd_mysql:1 notify on wc02
* Pseudo action: ms_drbd_mysql_confirmed-pre_notify_demote_0
* Resource action: intip_sql stop on wc02
* Resource action: drbd_www:0 notify on wc01
* Resource action: drbd_www:1 notify on wc02
* Pseudo action: ms_drbd_www_confirmed-pre_notify_stop_0
* Pseudo action: ms_drbd_www_stop_0
* Pseudo action: group_mysql-proxy:1_stop_0
* Resource action: mysql-proxy:1 stop on wc02
* Pseudo action: group_webservice:1_stop_0
* Resource action: apache2:1 stop on wc02
* Resource action: stonith_rackpdu:0 stop on wc01
* Resource action: stonith_rackpdu:1 stop on wc02
* Pseudo action: DoFencing_stopped_0
* Pseudo action: DoFencing_start_0
* Resource action: fs_mysql stop on wc02
* Resource action: drbd_www:1 stop on wc02
* Pseudo action: ms_drbd_www_stopped_0
* Pseudo action: group_mysql-proxy:1_stopped_0
* Pseudo action: clone_mysql-proxy_stopped_0
* Resource action: fs_www:1 stop on wc02
* Resource action: stonith_rackpdu:0 start on wc01
* Pseudo action: DoFencing_running_0
* Pseudo action: group_mysql_stopped_0
* Pseudo action: ms_drbd_www_post_notify_stopped_0
* Pseudo action: group_webservice:1_stopped_0
* Pseudo action: clone_webservice_stopped_0
* Resource action: stonith_rackpdu:0 monitor=5000 on wc01
* Pseudo action: ms_drbd_mysql_demote_0
* Resource action: drbd_www:0 notify on wc01
* Pseudo action: ms_drbd_www_confirmed-post_notify_stopped_0
* Pseudo action: clone_nfs-common_stop_0
* Resource action: drbd_mysql:1 demote on wc02
* Pseudo action: ms_drbd_mysql_demoted_0
* Pseudo action: group_nfs-common:1_stop_0
* Resource action: nfs-common:1 stop on wc02
* Pseudo action: ms_drbd_mysql_post_notify_demoted_0
* Pseudo action: group_nfs-common:1_stopped_0
* Pseudo action: clone_nfs-common_stopped_0
* Resource action: drbd_mysql:0 notify on wc01
* Resource action: drbd_mysql:1 notify on wc02
* Pseudo action: ms_drbd_mysql_confirmed-post_notify_demoted_0
* Pseudo action: ms_drbd_mysql_pre_notify_stop_0
* Resource action: drbd_mysql:0 notify on wc01
* Resource action: drbd_mysql:1 notify on wc02
* Pseudo action: ms_drbd_mysql_confirmed-pre_notify_stop_0
* Pseudo action: ms_drbd_mysql_stop_0
* Resource action: drbd_mysql:1 stop on wc02
* Pseudo action: ms_drbd_mysql_stopped_0
* Pseudo action: ms_drbd_mysql_post_notify_stopped_0
* Resource action: drbd_mysql:0 notify on wc01
* Pseudo action: ms_drbd_mysql_confirmed-post_notify_stopped_0
* Pseudo action: ms_drbd_mysql_pre_notify_promote_0
* Resource action: drbd_mysql:0 notify on wc01
* Pseudo action: ms_drbd_mysql_confirmed-pre_notify_promote_0
* Pseudo action: ms_drbd_mysql_promote_0
* Resource action: drbd_mysql:0 promote on wc01
* Pseudo action: ms_drbd_mysql_promoted_0
* Pseudo action: ms_drbd_mysql_post_notify_promoted_0
* Resource action: drbd_mysql:0 notify on wc01
* Pseudo action: ms_drbd_mysql_confirmed-post_notify_promoted_0
* Pseudo action: group_mysql_start_0
* Resource action: fs_mysql start on wc01
* Resource action: intip_sql start on wc01
* Resource action: mysql-server start on wc01
* Resource action: drbd_mysql:0 monitor=5000 on wc01
* Pseudo action: group_mysql_running_0
* Resource action: fs_mysql monitor=30000 on wc01
* Resource action: intip_sql monitor=30000 on wc01
* Resource action: mysql-server monitor=30000 on wc01
Revised Cluster Status:
* Node List:
* Node wc02: standby
* Online: [ wc01 ]
* Full List of Resources:
* Resource Group: group_www_data:
* fs_www_data (ocf:heartbeat:Filesystem): Started wc01
* nfs-kernel-server (lsb:nfs-kernel-server): Started wc01
* intip_nfs (ocf:heartbeat:IPaddr2): Started wc01
* Clone Set: ms_drbd_mysql [drbd_mysql] (promotable):
* Promoted: [ wc01 ]
* Stopped: [ wc02 ]
* Resource Group: group_mysql:
* fs_mysql (ocf:heartbeat:Filesystem): Started wc01
* intip_sql (ocf:heartbeat:IPaddr2): Started wc01
* mysql-server (ocf:heartbeat:mysql): Started wc01
* Clone Set: ms_drbd_www [drbd_www] (promotable):
* Promoted: [ wc01 ]
* Stopped: [ wc02 ]
* Clone Set: clone_nfs-common [group_nfs-common]:
* Started: [ wc01 ]
* Stopped: [ wc02 ]
* Clone Set: clone_mysql-proxy [group_mysql-proxy]:
* Started: [ wc01 ]
* Stopped: [ wc02 ]
* Clone Set: clone_webservice [group_webservice]:
* Started: [ wc01 ]
* Stopped: [ wc02 ]
* Resource Group: group_ftpd:
* extip_ftp (ocf:heartbeat:IPaddr2): Started wc01
* pure-ftpd (ocf:heartbeat:Pure-FTPd): Started wc01
* Clone Set: DoFencing [stonith_rackpdu] (unique):
* stonith_rackpdu:0 (stonith:external/rackpdu): Started wc01
* stonith_rackpdu:1 (stonith:external/rackpdu): Stopped
diff --git a/cts/scheduler/summary/promoted-7.summary b/cts/scheduler/summary/promoted-7.summary
index 4fc3a85e9a..0602f95895 100644
--- a/cts/scheduler/summary/promoted-7.summary
+++ b/cts/scheduler/summary/promoted-7.summary
@@ -1,121 +1,121 @@
Current cluster status:
* Node List:
* Node c001n01: UNCLEAN (offline)
* Online: [ c001n02 c001n03 c001n08 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN)
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n03
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n03
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n03
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n02
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n01 (UNCLEAN)
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Started c001n01 (UNCLEAN)
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n02
* child_DoFencing:3 (stonith:ssh): Started c001n08
* Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Promoted c001n01 (UNCLEAN)
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n01 (UNCLEAN)
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
Transition Summary:
* Fence (reboot) c001n01 'peer is no longer part of the cluster'
* Move DcIPaddr ( c001n01 -> c001n03 )
* Move ocf_192.168.100.181 ( c001n03 -> c001n02 )
* Move heartbeat_192.168.100.182 ( c001n03 -> c001n02 )
* Move ocf_192.168.100.183 ( c001n03 -> c001n02 )
* Move lsb_dummy ( c001n02 -> c001n08 )
* Move rsc_c001n01 ( c001n01 -> c001n03 )
* Stop child_DoFencing:0 ( c001n01 ) due to node availability
- * Stop ocf_msdummy:0 ( Promoted c001n01 ) due to node availability
- * Stop ocf_msdummy:4 ( Unpromoted c001n01 ) due to node availability
+ * Stop ocf_msdummy:0 ( Promoted c001n01 ) due to node availability
+ * Stop ocf_msdummy:4 ( Unpromoted c001n01 ) due to node availability
Executing Cluster Transition:
* Pseudo action: group-1_stop_0
* Resource action: ocf_192.168.100.183 stop on c001n03
* Resource action: lsb_dummy stop on c001n02
* Resource action: child_DoFencing:2 monitor on c001n08
* Resource action: child_DoFencing:2 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n03
* Resource action: child_DoFencing:3 monitor on c001n02
* Pseudo action: DoFencing_stop_0
* Resource action: ocf_msdummy:4 monitor on c001n08
* Resource action: ocf_msdummy:4 monitor on c001n03
* Resource action: ocf_msdummy:4 monitor on c001n02
* Resource action: ocf_msdummy:5 monitor on c001n08
* Resource action: ocf_msdummy:5 monitor on c001n02
* Resource action: ocf_msdummy:6 monitor on c001n08
* Resource action: ocf_msdummy:6 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n03
* Resource action: ocf_msdummy:7 monitor on c001n02
* Pseudo action: master_rsc_1_demote_0
* Fencing c001n01 (reboot)
* Pseudo action: DcIPaddr_stop_0
* Resource action: heartbeat_192.168.100.182 stop on c001n03
* Resource action: lsb_dummy start on c001n08
* Pseudo action: rsc_c001n01_stop_0
* Pseudo action: child_DoFencing:0_stop_0
* Pseudo action: DoFencing_stopped_0
* Pseudo action: ocf_msdummy:0_demote_0
* Pseudo action: master_rsc_1_demoted_0
* Pseudo action: master_rsc_1_stop_0
* Resource action: DcIPaddr start on c001n03
* Resource action: ocf_192.168.100.181 stop on c001n03
* Resource action: lsb_dummy monitor=5000 on c001n08
* Resource action: rsc_c001n01 start on c001n03
* Pseudo action: ocf_msdummy:0_stop_0
* Pseudo action: ocf_msdummy:4_stop_0
* Pseudo action: master_rsc_1_stopped_0
* Resource action: DcIPaddr monitor=5000 on c001n03
* Pseudo action: group-1_stopped_0
* Pseudo action: group-1_start_0
* Resource action: ocf_192.168.100.181 start on c001n02
* Resource action: heartbeat_192.168.100.182 start on c001n02
* Resource action: ocf_192.168.100.183 start on c001n02
* Resource action: rsc_c001n01 monitor=5000 on c001n03
* Pseudo action: group-1_running_0
* Resource action: ocf_192.168.100.181 monitor=5000 on c001n02
* Resource action: heartbeat_192.168.100.182 monitor=5000 on c001n02
* Resource action: ocf_192.168.100.183 monitor=5000 on c001n02
Revised Cluster Status:
* Node List:
* Online: [ c001n02 c001n03 c001n08 ]
* OFFLINE: [ c001n01 ]
* Full List of Resources:
* DcIPaddr (ocf:heartbeat:IPaddr): Started c001n03
* Resource Group: group-1:
* ocf_192.168.100.181 (ocf:heartbeat:IPaddr): Started c001n02
* heartbeat_192.168.100.182 (ocf:heartbeat:IPaddr): Started c001n02
* ocf_192.168.100.183 (ocf:heartbeat:IPaddr): Started c001n02
* lsb_dummy (lsb:/usr/lib/heartbeat/cts/LSBDummy): Started c001n08
* rsc_c001n01 (ocf:heartbeat:IPaddr): Started c001n03
* rsc_c001n08 (ocf:heartbeat:IPaddr): Started c001n08
* rsc_c001n02 (ocf:heartbeat:IPaddr): Started c001n02
* rsc_c001n03 (ocf:heartbeat:IPaddr): Started c001n03
* Clone Set: DoFencing [child_DoFencing] (unique):
* child_DoFencing:0 (stonith:ssh): Stopped
* child_DoFencing:1 (stonith:ssh): Started c001n03
* child_DoFencing:2 (stonith:ssh): Started c001n02
* child_DoFencing:3 (stonith:ssh): Started c001n08
* Clone Set: master_rsc_1 [ocf_msdummy] (promotable) (unique):
* ocf_msdummy:0 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:1 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:2 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:3 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
* ocf_msdummy:4 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Stopped
* ocf_msdummy:5 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n03
* ocf_msdummy:6 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n02
* ocf_msdummy:7 (ocf:heartbeat:/usr/lib/heartbeat/cts/OCFMSDummy): Unpromoted c001n08
diff --git a/cts/scheduler/summary/promoted-asymmetrical-order.summary b/cts/scheduler/summary/promoted-asymmetrical-order.summary
index df6e00c9c2..e10568e898 100644
--- a/cts/scheduler/summary/promoted-asymmetrical-order.summary
+++ b/cts/scheduler/summary/promoted-asymmetrical-order.summary
@@ -1,37 +1,37 @@
2 of 4 resource instances DISABLED and 0 BLOCKED from further action due to failure
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* Clone Set: ms1 [rsc1] (promotable) (disabled):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
* Clone Set: ms2 [rsc2] (promotable):
* Promoted: [ node2 ]
* Unpromoted: [ node1 ]
Transition Summary:
- * Stop rsc1:0 ( Promoted node1 ) due to node availability
- * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
+ * Stop rsc1:0 ( Promoted node1 ) due to node availability
+ * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms1_demote_0
* Resource action: rsc1:0 demote on node1
* Pseudo action: ms1_demoted_0
* Pseudo action: ms1_stop_0
* Resource action: rsc1:0 stop on node1
* Resource action: rsc1:1 stop on node2
* Pseudo action: ms1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* Clone Set: ms1 [rsc1] (promotable) (disabled):
* Stopped (disabled): [ node1 node2 ]
* Clone Set: ms2 [rsc2] (promotable):
* Promoted: [ node2 ]
* Unpromoted: [ node1 ]
diff --git a/cts/scheduler/summary/promoted-demote-2.summary b/cts/scheduler/summary/promoted-demote-2.summary
index daea66ae8b..115da9aaaf 100644
--- a/cts/scheduler/summary/promoted-demote-2.summary
+++ b/cts/scheduler/summary/promoted-demote-2.summary
@@ -1,75 +1,75 @@
Current cluster status:
* Node List:
* Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started pcmk-1
* Resource Group: group-1:
* r192.168.122.105 (ocf:heartbeat:IPaddr): Stopped
* r192.168.122.106 (ocf:heartbeat:IPaddr): Stopped
* r192.168.122.107 (ocf:heartbeat:IPaddr): Stopped
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Stopped
* migrator (ocf:pacemaker:Dummy): Started pcmk-4
* Clone Set: Connectivity [ping-1]:
* Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Clone Set: master-1 [stateful-1] (promotable):
* stateful-1 (ocf:pacemaker:Stateful): FAILED pcmk-1
* Unpromoted: [ pcmk-2 pcmk-3 pcmk-4 ]
Transition Summary:
* Start r192.168.122.105 ( pcmk-2 )
* Start r192.168.122.106 ( pcmk-2 )
* Start r192.168.122.107 ( pcmk-2 )
* Start lsb-dummy ( pcmk-2 )
- * Recover stateful-1:0 ( Unpromoted pcmk-1 )
+ * Recover stateful-1:0 ( Unpromoted pcmk-1 )
* Promote stateful-1:1 ( Unpromoted -> Promoted pcmk-2 )
Executing Cluster Transition:
* Resource action: stateful-1:0 cancel=15000 on pcmk-2
* Pseudo action: master-1_stop_0
* Resource action: stateful-1:1 stop on pcmk-1
* Pseudo action: master-1_stopped_0
* Pseudo action: master-1_start_0
* Resource action: stateful-1:1 start on pcmk-1
* Pseudo action: master-1_running_0
* Resource action: stateful-1:1 monitor=15000 on pcmk-1
* Pseudo action: master-1_promote_0
* Resource action: stateful-1:0 promote on pcmk-2
* Pseudo action: master-1_promoted_0
* Pseudo action: group-1_start_0
* Resource action: r192.168.122.105 start on pcmk-2
* Resource action: r192.168.122.106 start on pcmk-2
* Resource action: r192.168.122.107 start on pcmk-2
* Resource action: stateful-1:0 monitor=16000 on pcmk-2
* Pseudo action: group-1_running_0
* Resource action: r192.168.122.105 monitor=5000 on pcmk-2
* Resource action: r192.168.122.106 monitor=5000 on pcmk-2
* Resource action: r192.168.122.107 monitor=5000 on pcmk-2
* Resource action: lsb-dummy start on pcmk-2
* Resource action: lsb-dummy monitor=5000 on pcmk-2
Revised Cluster Status:
* Node List:
* Online: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started pcmk-1
* Resource Group: group-1:
* r192.168.122.105 (ocf:heartbeat:IPaddr): Started pcmk-2
* r192.168.122.106 (ocf:heartbeat:IPaddr): Started pcmk-2
* r192.168.122.107 (ocf:heartbeat:IPaddr): Started pcmk-2
* rsc_pcmk-1 (ocf:heartbeat:IPaddr): Started pcmk-1
* rsc_pcmk-2 (ocf:heartbeat:IPaddr): Started pcmk-2
* rsc_pcmk-3 (ocf:heartbeat:IPaddr): Started pcmk-3
* rsc_pcmk-4 (ocf:heartbeat:IPaddr): Started pcmk-4
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started pcmk-2
* migrator (ocf:pacemaker:Dummy): Started pcmk-4
* Clone Set: Connectivity [ping-1]:
* Started: [ pcmk-1 pcmk-2 pcmk-3 pcmk-4 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ pcmk-2 ]
* Unpromoted: [ pcmk-1 pcmk-3 pcmk-4 ]
diff --git a/cts/scheduler/summary/promoted-failed-demote-2.summary b/cts/scheduler/summary/promoted-failed-demote-2.summary
index 198d9ad3ee..c8504e9e1d 100644
--- a/cts/scheduler/summary/promoted-failed-demote-2.summary
+++ b/cts/scheduler/summary/promoted-failed-demote-2.summary
@@ -1,47 +1,47 @@
Current cluster status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
* Clone Set: ms-sf [group] (promotable) (unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): FAILED dl380g5b
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
Transition Summary:
- * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability
+ * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability
* Promote stateful-1:1 ( Unpromoted -> Promoted dl380g5a )
* Promote stateful-2:1 ( Unpromoted -> Promoted dl380g5a )
Executing Cluster Transition:
* Resource action: stateful-1:1 cancel=20000 on dl380g5a
* Resource action: stateful-2:1 cancel=20000 on dl380g5a
* Pseudo action: ms-sf_stop_0
* Pseudo action: group:0_stop_0
* Resource action: stateful-1:0 stop on dl380g5b
* Pseudo action: group:0_stopped_0
* Pseudo action: ms-sf_stopped_0
* Pseudo action: ms-sf_promote_0
* Pseudo action: group:1_promote_0
* Resource action: stateful-1:1 promote on dl380g5a
* Resource action: stateful-2:1 promote on dl380g5a
* Pseudo action: group:1_promoted_0
* Resource action: stateful-1:1 monitor=10000 on dl380g5a
* Resource action: stateful-2:1 monitor=10000 on dl380g5a
* Pseudo action: ms-sf_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
* Clone Set: ms-sf [group] (promotable) (unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Stopped
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
diff --git a/cts/scheduler/summary/promoted-failed-demote.summary b/cts/scheduler/summary/promoted-failed-demote.summary
index 884a380063..f071025528 100644
--- a/cts/scheduler/summary/promoted-failed-demote.summary
+++ b/cts/scheduler/summary/promoted-failed-demote.summary
@@ -1,64 +1,64 @@
Current cluster status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
* Clone Set: ms-sf [group] (promotable) (unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): FAILED dl380g5b
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Unpromoted dl380g5a
Transition Summary:
- * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability
+ * Stop stateful-1:0 ( Unpromoted dl380g5b ) due to node availability
* Promote stateful-1:1 ( Unpromoted -> Promoted dl380g5a )
* Promote stateful-2:1 ( Unpromoted -> Promoted dl380g5a )
Executing Cluster Transition:
* Resource action: stateful-1:1 cancel=20000 on dl380g5a
* Resource action: stateful-2:1 cancel=20000 on dl380g5a
* Pseudo action: ms-sf_pre_notify_stop_0
* Resource action: stateful-1:0 notify on dl380g5b
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-pre_notify_stop_0
* Pseudo action: ms-sf_stop_0
* Pseudo action: group:0_stop_0
* Resource action: stateful-1:0 stop on dl380g5b
* Pseudo action: group:0_stopped_0
* Pseudo action: ms-sf_stopped_0
* Pseudo action: ms-sf_post_notify_stopped_0
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-post_notify_stopped_0
* Pseudo action: ms-sf_pre_notify_promote_0
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-pre_notify_promote_0
* Pseudo action: ms-sf_promote_0
* Pseudo action: group:1_promote_0
* Resource action: stateful-1:1 promote on dl380g5a
* Resource action: stateful-2:1 promote on dl380g5a
* Pseudo action: group:1_promoted_0
* Pseudo action: ms-sf_promoted_0
* Pseudo action: ms-sf_post_notify_promoted_0
* Resource action: stateful-1:1 notify on dl380g5a
* Resource action: stateful-2:1 notify on dl380g5a
* Pseudo action: ms-sf_confirmed-post_notify_promoted_0
* Resource action: stateful-1:1 monitor=10000 on dl380g5a
* Resource action: stateful-2:1 monitor=10000 on dl380g5a
Revised Cluster Status:
* Node List:
* Online: [ dl380g5a dl380g5b ]
* Full List of Resources:
* Clone Set: ms-sf [group] (promotable) (unique):
* Resource Group: group:0:
* stateful-1:0 (ocf:heartbeat:Stateful): Stopped
* stateful-2:0 (ocf:heartbeat:Stateful): Stopped
* Resource Group: group:1:
* stateful-1:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
* stateful-2:1 (ocf:heartbeat:Stateful): Promoted dl380g5a
diff --git a/cts/scheduler/summary/remote-connection-unrecoverable.summary b/cts/scheduler/summary/remote-connection-unrecoverable.summary
index bd1adfcfa4..3cfb64565a 100644
--- a/cts/scheduler/summary/remote-connection-unrecoverable.summary
+++ b/cts/scheduler/summary/remote-connection-unrecoverable.summary
@@ -1,54 +1,54 @@
Current cluster status:
* Node List:
* Node node1: UNCLEAN (offline)
* Online: [ node2 ]
* RemoteOnline: [ remote1 ]
* Full List of Resources:
* remote1 (ocf:pacemaker:remote): Started node1 (UNCLEAN)
* killer (stonith:fence_xvm): Started node2
* rsc1 (ocf:pacemaker:Dummy): Started remote1
* Clone Set: rsc2-master [rsc2] (promotable):
* rsc2 (ocf:pacemaker:Stateful): Promoted node1 (UNCLEAN)
* Promoted: [ node2 ]
* Stopped: [ remote1 ]
Transition Summary:
* Fence (reboot) remote1 'resources are active and the connection is unrecoverable'
* Fence (reboot) node1 'peer is no longer part of the cluster'
* Stop remote1 ( node1 ) due to node availability
* Restart killer ( node2 ) due to resource definition change
* Move rsc1 ( remote1 -> node2 )
- * Stop rsc2:0 ( Promoted node1 ) due to node availability
+ * Stop rsc2:0 ( Promoted node1 ) due to node availability
Executing Cluster Transition:
* Pseudo action: remote1_stop_0
* Resource action: killer stop on node2
* Resource action: rsc1 monitor on node2
* Fencing node1 (reboot)
* Fencing remote1 (reboot)
* Resource action: killer start on node2
* Resource action: killer monitor=60000 on node2
* Pseudo action: rsc1_stop_0
* Pseudo action: rsc2-master_demote_0
* Resource action: rsc1 start on node2
* Pseudo action: rsc2_demote_0
* Pseudo action: rsc2-master_demoted_0
* Pseudo action: rsc2-master_stop_0
* Resource action: rsc1 monitor=10000 on node2
* Pseudo action: rsc2_stop_0
* Pseudo action: rsc2-master_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node2 ]
* OFFLINE: [ node1 ]
* RemoteOFFLINE: [ remote1 ]
* Full List of Resources:
* remote1 (ocf:pacemaker:remote): Stopped
* killer (stonith:fence_xvm): Started node2
* rsc1 (ocf:pacemaker:Dummy): Started node2
* Clone Set: rsc2-master [rsc2] (promotable):
* Promoted: [ node2 ]
* Stopped: [ node1 remote1 ]
diff --git a/cts/scheduler/summary/remote-recover-all.summary b/cts/scheduler/summary/remote-recover-all.summary
index 176c1de8b3..18d10730bf 100644
--- a/cts/scheduler/summary/remote-recover-all.summary
+++ b/cts/scheduler/summary/remote-recover-all.summary
@@ -1,146 +1,146 @@
Using the original execution date of: 2017-05-03 13:33:24Z
Current cluster status:
* Node List:
* Node controller-1: UNCLEAN (offline)
* Online: [ controller-0 controller-2 ]
* RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-1 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 galera-2 ]
* Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN)
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* Clone Set: haproxy-clone [haproxy]:
* haproxy (systemd:haproxy): Started controller-1 (UNCLEAN)
* Started: [ controller-0 controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN)
Transition Summary:
* Fence (reboot) messaging-1 'resources are active and the connection is unrecoverable'
* Fence (reboot) galera-2 'resources are active and the connection is unrecoverable'
* Fence (reboot) controller-1 'peer is no longer part of the cluster'
* Stop messaging-1 ( controller-1 ) due to node availability
* Move galera-0 ( controller-1 -> controller-2 )
* Stop galera-2 ( controller-1 ) due to node availability
* Stop rabbitmq:2 ( messaging-1 ) due to node availability
- * Stop galera:1 ( Promoted galera-2 ) due to node availability
- * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
+ * Stop galera:1 ( Promoted galera-2 ) due to node availability
+ * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
* Move ip-172.17.1.14 ( controller-1 -> controller-2 )
* Move ip-172.17.1.17 ( controller-1 -> controller-2 )
* Move ip-172.17.4.11 ( controller-1 -> controller-2 )
* Stop haproxy:0 ( controller-1 ) due to node availability
* Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 )
Executing Cluster Transition:
* Pseudo action: messaging-1_stop_0
* Pseudo action: galera-0_stop_0
* Pseudo action: galera-2_stop_0
* Pseudo action: galera-master_demote_0
* Pseudo action: redis-master_pre_notify_stop_0
* Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0
* Fencing controller-1 (reboot)
* Pseudo action: redis_post_notify_stop_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-pre_notify_stop_0
* Pseudo action: redis-master_stop_0
* Pseudo action: haproxy-clone_stop_0
* Fencing galera-2 (reboot)
* Pseudo action: galera_demote_0
* Pseudo action: galera-master_demoted_0
* Pseudo action: galera-master_stop_0
* Pseudo action: redis_stop_0
* Pseudo action: redis-master_stopped_0
* Pseudo action: haproxy_stop_0
* Pseudo action: haproxy-clone_stopped_0
* Fencing messaging-1 (reboot)
* Resource action: galera-0 start on controller-2
* Pseudo action: rabbitmq_post_notify_stop_0
* Pseudo action: rabbitmq-clone_stop_0
* Pseudo action: galera_stop_0
* Resource action: galera monitor=10000 on galera-0
* Pseudo action: galera-master_stopped_0
* Pseudo action: redis-master_post_notify_stopped_0
* Pseudo action: ip-172.17.1.14_stop_0
* Pseudo action: ip-172.17.1.17_stop_0
* Pseudo action: ip-172.17.4.11_stop_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2
* Resource action: galera-0 monitor=20000 on controller-2
* Resource action: rabbitmq notify on messaging-2
* Resource action: rabbitmq notify on messaging-0
* Pseudo action: rabbitmq_notified_0
* Pseudo action: rabbitmq_stop_0
* Pseudo action: rabbitmq-clone_stopped_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-post_notify_stopped_0
* Resource action: ip-172.17.1.14 start on controller-2
* Resource action: ip-172.17.1.17 start on controller-2
* Resource action: ip-172.17.4.11 start on controller-2
* Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
* Pseudo action: redis_notified_0
* Resource action: ip-172.17.1.14 monitor=10000 on controller-2
* Resource action: ip-172.17.1.17 monitor=10000 on controller-2
* Resource action: ip-172.17.4.11 monitor=10000 on controller-2
Using the original execution date of: 2017-05-03 13:33:24Z
Revised Cluster Status:
* Node List:
* Online: [ controller-0 controller-2 ]
* OFFLINE: [ controller-1 ]
* RemoteOnline: [ galera-0 galera-1 messaging-0 messaging-2 ]
* RemoteOFFLINE: [ galera-2 messaging-1 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Stopped
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-2
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Stopped
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 messaging-1 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2
* Clone Set: haproxy-clone [haproxy]:
* Started: [ controller-0 controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2
diff --git a/cts/scheduler/summary/remote-recover-connection.summary b/cts/scheduler/summary/remote-recover-connection.summary
index fd6900dd96..a9723bc5e1 100644
--- a/cts/scheduler/summary/remote-recover-connection.summary
+++ b/cts/scheduler/summary/remote-recover-connection.summary
@@ -1,132 +1,132 @@
Using the original execution date of: 2017-05-03 13:33:24Z
Current cluster status:
* Node List:
* Node controller-1: UNCLEAN (offline)
* Online: [ controller-0 controller-2 ]
* RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-1 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 galera-2 ]
* Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN)
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* Clone Set: haproxy-clone [haproxy]:
* haproxy (systemd:haproxy): Started controller-1 (UNCLEAN)
* Started: [ controller-0 controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN)
Transition Summary:
* Fence (reboot) controller-1 'peer is no longer part of the cluster'
* Move messaging-1 ( controller-1 -> controller-2 )
* Move galera-0 ( controller-1 -> controller-2 )
* Move galera-2 ( controller-1 -> controller-2 )
- * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
+ * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
* Move ip-172.17.1.14 ( controller-1 -> controller-2 )
* Move ip-172.17.1.17 ( controller-1 -> controller-2 )
* Move ip-172.17.4.11 ( controller-1 -> controller-2 )
* Stop haproxy:0 ( controller-1 ) due to node availability
* Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 )
Executing Cluster Transition:
* Pseudo action: messaging-1_stop_0
* Pseudo action: galera-0_stop_0
* Pseudo action: galera-2_stop_0
* Pseudo action: redis-master_pre_notify_stop_0
* Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0
* Fencing controller-1 (reboot)
* Resource action: messaging-1 start on controller-2
* Resource action: galera-0 start on controller-2
* Resource action: galera-2 start on controller-2
* Resource action: rabbitmq monitor=10000 on messaging-1
* Resource action: galera monitor=10000 on galera-2
* Resource action: galera monitor=10000 on galera-0
* Pseudo action: redis_post_notify_stop_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-pre_notify_stop_0
* Pseudo action: redis-master_stop_0
* Pseudo action: haproxy-clone_stop_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2
* Resource action: messaging-1 monitor=20000 on controller-2
* Resource action: galera-0 monitor=20000 on controller-2
* Resource action: galera-2 monitor=20000 on controller-2
* Pseudo action: redis_stop_0
* Pseudo action: redis-master_stopped_0
* Pseudo action: haproxy_stop_0
* Pseudo action: haproxy-clone_stopped_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
* Pseudo action: redis-master_post_notify_stopped_0
* Pseudo action: ip-172.17.1.14_stop_0
* Pseudo action: ip-172.17.1.17_stop_0
* Pseudo action: ip-172.17.4.11_stop_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-post_notify_stopped_0
* Resource action: ip-172.17.1.14 start on controller-2
* Resource action: ip-172.17.1.17 start on controller-2
* Resource action: ip-172.17.4.11 start on controller-2
* Pseudo action: redis_notified_0
* Resource action: ip-172.17.1.14 monitor=10000 on controller-2
* Resource action: ip-172.17.1.17 monitor=10000 on controller-2
* Resource action: ip-172.17.4.11 monitor=10000 on controller-2
Using the original execution date of: 2017-05-03 13:33:24Z
Revised Cluster Status:
* Node List:
* Online: [ controller-0 controller-2 ]
* OFFLINE: [ controller-1 ]
* RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Started controller-2
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-2
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Started controller-2
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-1 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 galera-2 ]
* Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2
* Clone Set: haproxy-clone [haproxy]:
* Started: [ controller-0 controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2
diff --git a/cts/scheduler/summary/remote-recover-no-resources.summary b/cts/scheduler/summary/remote-recover-no-resources.summary
index 332d1c4123..d7d9ef942c 100644
--- a/cts/scheduler/summary/remote-recover-no-resources.summary
+++ b/cts/scheduler/summary/remote-recover-no-resources.summary
@@ -1,137 +1,137 @@
Using the original execution date of: 2017-05-03 13:33:24Z
Current cluster status:
* Node List:
* Node controller-1: UNCLEAN (offline)
* Online: [ controller-0 controller-2 ]
* RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-1 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN)
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* Clone Set: haproxy-clone [haproxy]:
* haproxy (systemd:haproxy): Started controller-1 (UNCLEAN)
* Started: [ controller-0 controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN)
Transition Summary:
* Fence (reboot) messaging-1 'resources are active and the connection is unrecoverable'
* Fence (reboot) controller-1 'peer is no longer part of the cluster'
* Stop messaging-1 ( controller-1 ) due to node availability
* Move galera-0 ( controller-1 -> controller-2 )
* Stop galera-2 ( controller-1 ) due to node availability
* Stop rabbitmq:2 ( messaging-1 ) due to node availability
- * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
+ * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
* Move ip-172.17.1.14 ( controller-1 -> controller-2 )
* Move ip-172.17.1.17 ( controller-1 -> controller-2 )
* Move ip-172.17.4.11 ( controller-1 -> controller-2 )
* Stop haproxy:0 ( controller-1 ) due to node availability
* Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 )
Executing Cluster Transition:
* Pseudo action: messaging-1_stop_0
* Pseudo action: galera-0_stop_0
* Pseudo action: galera-2_stop_0
* Pseudo action: redis-master_pre_notify_stop_0
* Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0
* Fencing controller-1 (reboot)
* Pseudo action: redis_post_notify_stop_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-pre_notify_stop_0
* Pseudo action: redis-master_stop_0
* Pseudo action: haproxy-clone_stop_0
* Fencing messaging-1 (reboot)
* Resource action: galera-0 start on controller-2
* Pseudo action: rabbitmq_post_notify_stop_0
* Pseudo action: rabbitmq-clone_stop_0
* Resource action: galera monitor=10000 on galera-0
* Pseudo action: redis_stop_0
* Pseudo action: redis-master_stopped_0
* Pseudo action: haproxy_stop_0
* Pseudo action: haproxy-clone_stopped_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2
* Resource action: galera-0 monitor=20000 on controller-2
* Resource action: rabbitmq notify on messaging-2
* Resource action: rabbitmq notify on messaging-0
* Pseudo action: rabbitmq_notified_0
* Pseudo action: rabbitmq_stop_0
* Pseudo action: rabbitmq-clone_stopped_0
* Pseudo action: redis-master_post_notify_stopped_0
* Pseudo action: ip-172.17.1.14_stop_0
* Pseudo action: ip-172.17.1.17_stop_0
* Pseudo action: ip-172.17.4.11_stop_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-post_notify_stopped_0
* Resource action: ip-172.17.1.14 start on controller-2
* Resource action: ip-172.17.1.17 start on controller-2
* Resource action: ip-172.17.4.11 start on controller-2
* Pseudo action: redis_notified_0
* Resource action: ip-172.17.1.14 monitor=10000 on controller-2
* Resource action: ip-172.17.1.17 monitor=10000 on controller-2
* Resource action: ip-172.17.4.11 monitor=10000 on controller-2
Using the original execution date of: 2017-05-03 13:33:24Z
Revised Cluster Status:
* Node List:
* Online: [ controller-0 controller-2 ]
* OFFLINE: [ controller-1 ]
* RemoteOnline: [ galera-0 galera-1 messaging-0 messaging-2 ]
* RemoteOFFLINE: [ galera-2 messaging-1 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Stopped
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-2
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Stopped
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 messaging-1 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2
* Clone Set: haproxy-clone [haproxy]:
* Started: [ controller-0 controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2
diff --git a/cts/scheduler/summary/remote-recover-unknown.summary b/cts/scheduler/summary/remote-recover-unknown.summary
index ac5143a16e..4f3d045284 100644
--- a/cts/scheduler/summary/remote-recover-unknown.summary
+++ b/cts/scheduler/summary/remote-recover-unknown.summary
@@ -1,139 +1,139 @@
Using the original execution date of: 2017-05-03 13:33:24Z
Current cluster status:
* Node List:
* Node controller-1: UNCLEAN (offline)
* Online: [ controller-0 controller-2 ]
* RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-1 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN)
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* Clone Set: haproxy-clone [haproxy]:
* haproxy (systemd:haproxy): Started controller-1 (UNCLEAN)
* Started: [ controller-0 controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN)
Transition Summary:
* Fence (reboot) galera-2 'resources are in an unknown state and the connection is unrecoverable'
* Fence (reboot) messaging-1 'resources are active and the connection is unrecoverable'
* Fence (reboot) controller-1 'peer is no longer part of the cluster'
* Stop messaging-1 ( controller-1 ) due to node availability
* Move galera-0 ( controller-1 -> controller-2 )
* Stop galera-2 ( controller-1 ) due to node availability
* Stop rabbitmq:2 ( messaging-1 ) due to node availability
- * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
+ * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
* Move ip-172.17.1.14 ( controller-1 -> controller-2 )
* Move ip-172.17.1.17 ( controller-1 -> controller-2 )
* Move ip-172.17.4.11 ( controller-1 -> controller-2 )
* Stop haproxy:0 ( controller-1 ) due to node availability
* Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 )
Executing Cluster Transition:
* Pseudo action: messaging-1_stop_0
* Pseudo action: galera-0_stop_0
* Pseudo action: galera-2_stop_0
* Pseudo action: redis-master_pre_notify_stop_0
* Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0
* Fencing controller-1 (reboot)
* Pseudo action: redis_post_notify_stop_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-pre_notify_stop_0
* Pseudo action: redis-master_stop_0
* Pseudo action: haproxy-clone_stop_0
* Fencing galera-2 (reboot)
* Fencing messaging-1 (reboot)
* Resource action: galera-0 start on controller-2
* Pseudo action: rabbitmq_post_notify_stop_0
* Pseudo action: rabbitmq-clone_stop_0
* Resource action: galera monitor=10000 on galera-0
* Pseudo action: redis_stop_0
* Pseudo action: redis-master_stopped_0
* Pseudo action: haproxy_stop_0
* Pseudo action: haproxy-clone_stopped_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2
* Resource action: galera-0 monitor=20000 on controller-2
* Resource action: rabbitmq notify on messaging-2
* Resource action: rabbitmq notify on messaging-0
* Pseudo action: rabbitmq_notified_0
* Pseudo action: rabbitmq_stop_0
* Pseudo action: rabbitmq-clone_stopped_0
* Pseudo action: redis-master_post_notify_stopped_0
* Pseudo action: ip-172.17.1.14_stop_0
* Pseudo action: ip-172.17.1.17_stop_0
* Pseudo action: ip-172.17.4.11_stop_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-post_notify_stopped_0
* Resource action: ip-172.17.1.14 start on controller-2
* Resource action: ip-172.17.1.17 start on controller-2
* Resource action: ip-172.17.4.11 start on controller-2
* Pseudo action: redis_notified_0
* Resource action: ip-172.17.1.14 monitor=10000 on controller-2
* Resource action: ip-172.17.1.17 monitor=10000 on controller-2
* Resource action: ip-172.17.4.11 monitor=10000 on controller-2
Using the original execution date of: 2017-05-03 13:33:24Z
Revised Cluster Status:
* Node List:
* Online: [ controller-0 controller-2 ]
* OFFLINE: [ controller-1 ]
* RemoteOnline: [ galera-0 galera-1 messaging-0 messaging-2 ]
* RemoteOFFLINE: [ galera-2 messaging-1 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Stopped
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-2
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Stopped
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 messaging-1 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2
* Clone Set: haproxy-clone [haproxy]:
* Started: [ controller-0 controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2
diff --git a/cts/scheduler/summary/remote-recovery.summary b/cts/scheduler/summary/remote-recovery.summary
index fd6900dd96..a9723bc5e1 100644
--- a/cts/scheduler/summary/remote-recovery.summary
+++ b/cts/scheduler/summary/remote-recovery.summary
@@ -1,132 +1,132 @@
Using the original execution date of: 2017-05-03 13:33:24Z
Current cluster status:
* Node List:
* Node controller-1: UNCLEAN (offline)
* Online: [ controller-0 controller-2 ]
* RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Started controller-1 (UNCLEAN)
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-1 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 galera-2 ]
* Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* redis (ocf:heartbeat:redis): Unpromoted controller-1 (UNCLEAN)
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-1 (UNCLEAN)
* Clone Set: haproxy-clone [haproxy]:
* haproxy (systemd:haproxy): Started controller-1 (UNCLEAN)
* Started: [ controller-0 controller-2 ]
* Stopped: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-1 (UNCLEAN)
Transition Summary:
* Fence (reboot) controller-1 'peer is no longer part of the cluster'
* Move messaging-1 ( controller-1 -> controller-2 )
* Move galera-0 ( controller-1 -> controller-2 )
* Move galera-2 ( controller-1 -> controller-2 )
- * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
+ * Stop redis:0 ( Unpromoted controller-1 ) due to node availability
* Move ip-172.17.1.14 ( controller-1 -> controller-2 )
* Move ip-172.17.1.17 ( controller-1 -> controller-2 )
* Move ip-172.17.4.11 ( controller-1 -> controller-2 )
* Stop haproxy:0 ( controller-1 ) due to node availability
* Move stonith-fence_ipmilan-5254005bdbb5 ( controller-1 -> controller-2 )
Executing Cluster Transition:
* Pseudo action: messaging-1_stop_0
* Pseudo action: galera-0_stop_0
* Pseudo action: galera-2_stop_0
* Pseudo action: redis-master_pre_notify_stop_0
* Pseudo action: stonith-fence_ipmilan-5254005bdbb5_stop_0
* Fencing controller-1 (reboot)
* Resource action: messaging-1 start on controller-2
* Resource action: galera-0 start on controller-2
* Resource action: galera-2 start on controller-2
* Resource action: rabbitmq monitor=10000 on messaging-1
* Resource action: galera monitor=10000 on galera-2
* Resource action: galera monitor=10000 on galera-0
* Pseudo action: redis_post_notify_stop_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-pre_notify_stop_0
* Pseudo action: redis-master_stop_0
* Pseudo action: haproxy-clone_stop_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2
* Resource action: messaging-1 monitor=20000 on controller-2
* Resource action: galera-0 monitor=20000 on controller-2
* Resource action: galera-2 monitor=20000 on controller-2
* Pseudo action: redis_stop_0
* Pseudo action: redis-master_stopped_0
* Pseudo action: haproxy_stop_0
* Pseudo action: haproxy-clone_stopped_0
* Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
* Pseudo action: redis-master_post_notify_stopped_0
* Pseudo action: ip-172.17.1.14_stop_0
* Pseudo action: ip-172.17.1.17_stop_0
* Pseudo action: ip-172.17.4.11_stop_0
* Resource action: redis notify on controller-0
* Resource action: redis notify on controller-2
* Pseudo action: redis-master_confirmed-post_notify_stopped_0
* Resource action: ip-172.17.1.14 start on controller-2
* Resource action: ip-172.17.1.17 start on controller-2
* Resource action: ip-172.17.4.11 start on controller-2
* Pseudo action: redis_notified_0
* Resource action: ip-172.17.1.14 monitor=10000 on controller-2
* Resource action: ip-172.17.1.17 monitor=10000 on controller-2
* Resource action: ip-172.17.4.11 monitor=10000 on controller-2
Using the original execution date of: 2017-05-03 13:33:24Z
Revised Cluster Status:
* Node List:
* Online: [ controller-0 controller-2 ]
* OFFLINE: [ controller-1 ]
* RemoteOnline: [ galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* Full List of Resources:
* messaging-0 (ocf:pacemaker:remote): Started controller-0
* messaging-1 (ocf:pacemaker:remote): Started controller-2
* messaging-2 (ocf:pacemaker:remote): Started controller-0
* galera-0 (ocf:pacemaker:remote): Started controller-2
* galera-1 (ocf:pacemaker:remote): Started controller-0
* galera-2 (ocf:pacemaker:remote): Started controller-2
* Clone Set: rabbitmq-clone [rabbitmq]:
* Started: [ messaging-0 messaging-1 messaging-2 ]
* Stopped: [ controller-0 controller-1 controller-2 galera-0 galera-1 galera-2 ]
* Clone Set: galera-master [galera] (promotable):
* Promoted: [ galera-0 galera-1 galera-2 ]
* Stopped: [ controller-0 controller-1 controller-2 messaging-0 messaging-1 messaging-2 ]
* Clone Set: redis-master [redis] (promotable):
* Promoted: [ controller-0 ]
* Unpromoted: [ controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* ip-192.168.24.6 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-10.0.0.102 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.1.14 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.1.17 (ocf:heartbeat:IPaddr2): Started controller-2
* ip-172.17.3.15 (ocf:heartbeat:IPaddr2): Started controller-0
* ip-172.17.4.11 (ocf:heartbeat:IPaddr2): Started controller-2
* Clone Set: haproxy-clone [haproxy]:
* Started: [ controller-0 controller-2 ]
* Stopped: [ controller-1 galera-0 galera-1 galera-2 messaging-0 messaging-1 messaging-2 ]
* openstack-cinder-volume (systemd:openstack-cinder-volume): Started controller-0
* stonith-fence_ipmilan-525400bbf613 (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-525400b4f6bd (stonith:fence_ipmilan): Started controller-0
* stonith-fence_ipmilan-5254005bdbb5 (stonith:fence_ipmilan): Started controller-2
diff --git a/cts/scheduler/summary/rsc-sets-promoted.summary b/cts/scheduler/summary/rsc-sets-promoted.summary
index a45e4b16e8..3db15881a0 100644
--- a/cts/scheduler/summary/rsc-sets-promoted.summary
+++ b/cts/scheduler/summary/rsc-sets-promoted.summary
@@ -1,49 +1,49 @@
Current cluster status:
* Node List:
* Node node1: standby (with active resources)
* Online: [ node2 ]
* Full List of Resources:
* Clone Set: ms-rsc [rsc] (promotable):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
* rsc1 (ocf:pacemaker:Dummy): Started node1
* rsc2 (ocf:pacemaker:Dummy): Started node1
* rsc3 (ocf:pacemaker:Dummy): Started node1
Transition Summary:
- * Stop rsc:0 ( Promoted node1 ) due to node availability
+ * Stop rsc:0 ( Promoted node1 ) due to node availability
* Promote rsc:1 ( Unpromoted -> Promoted node2 )
* Move rsc1 ( node1 -> node2 )
* Move rsc2 ( node1 -> node2 )
* Move rsc3 ( node1 -> node2 )
Executing Cluster Transition:
* Resource action: rsc1 stop on node1
* Resource action: rsc2 stop on node1
* Resource action: rsc3 stop on node1
* Pseudo action: ms-rsc_demote_0
* Resource action: rsc:0 demote on node1
* Pseudo action: ms-rsc_demoted_0
* Pseudo action: ms-rsc_stop_0
* Resource action: rsc:0 stop on node1
* Pseudo action: ms-rsc_stopped_0
* Pseudo action: ms-rsc_promote_0
* Resource action: rsc:1 promote on node2
* Pseudo action: ms-rsc_promoted_0
* Resource action: rsc1 start on node2
* Resource action: rsc2 start on node2
* Resource action: rsc3 start on node2
Revised Cluster Status:
* Node List:
* Node node1: standby
* Online: [ node2 ]
* Full List of Resources:
* Clone Set: ms-rsc [rsc] (promotable):
* Promoted: [ node2 ]
* Stopped: [ node1 ]
* rsc1 (ocf:pacemaker:Dummy): Started node2
* rsc2 (ocf:pacemaker:Dummy): Started node2
* rsc3 (ocf:pacemaker:Dummy): Started node2
diff --git a/cts/scheduler/summary/ticket-promoted-14.summary b/cts/scheduler/summary/ticket-promoted-14.summary
index ee8912b2e9..80ff84346b 100644
--- a/cts/scheduler/summary/ticket-promoted-14.summary
+++ b/cts/scheduler/summary/ticket-promoted-14.summary
@@ -1,31 +1,31 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
Transition Summary:
- * Stop rsc1:0 ( Promoted node1 ) due to node availability
- * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
+ * Stop rsc1:0 ( Promoted node1 ) due to node availability
+ * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms1_demote_0
* Resource action: rsc1:1 demote on node1
* Pseudo action: ms1_demoted_0
* Pseudo action: ms1_stop_0
* Resource action: rsc1:1 stop on node1
* Resource action: rsc1:0 stop on node2
* Pseudo action: ms1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Stopped: [ node1 node2 ]
diff --git a/cts/scheduler/summary/ticket-promoted-15.summary b/cts/scheduler/summary/ticket-promoted-15.summary
index ee8912b2e9..80ff84346b 100644
--- a/cts/scheduler/summary/ticket-promoted-15.summary
+++ b/cts/scheduler/summary/ticket-promoted-15.summary
@@ -1,31 +1,31 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
Transition Summary:
- * Stop rsc1:0 ( Promoted node1 ) due to node availability
- * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
+ * Stop rsc1:0 ( Promoted node1 ) due to node availability
+ * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms1_demote_0
* Resource action: rsc1:1 demote on node1
* Pseudo action: ms1_demoted_0
* Pseudo action: ms1_stop_0
* Resource action: rsc1:1 stop on node1
* Resource action: rsc1:0 stop on node2
* Pseudo action: ms1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Stopped: [ node1 node2 ]
diff --git a/cts/scheduler/summary/ticket-promoted-21.summary b/cts/scheduler/summary/ticket-promoted-21.summary
index f116a2eea0..788573facb 100644
--- a/cts/scheduler/summary/ticket-promoted-21.summary
+++ b/cts/scheduler/summary/ticket-promoted-21.summary
@@ -1,36 +1,36 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
Transition Summary:
* Fence (reboot) node1 'deadman ticket was lost'
* Move rsc_stonith ( node1 -> node2 )
- * Stop rsc1:0 ( Promoted node1 ) due to node availability
+ * Stop rsc1:0 ( Promoted node1 ) due to node availability
Executing Cluster Transition:
* Pseudo action: rsc_stonith_stop_0
* Pseudo action: ms1_demote_0
* Fencing node1 (reboot)
* Resource action: rsc_stonith start on node2
* Pseudo action: rsc1:1_demote_0
* Pseudo action: ms1_demoted_0
* Pseudo action: ms1_stop_0
* Pseudo action: rsc1:1_stop_0
* Pseudo action: ms1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node2 ]
* OFFLINE: [ node1 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node2
* Clone Set: ms1 [rsc1] (promotable):
* Unpromoted: [ node2 ]
* Stopped: [ node1 ]
diff --git a/cts/scheduler/summary/ticket-promoted-3.summary b/cts/scheduler/summary/ticket-promoted-3.summary
index ee8912b2e9..80ff84346b 100644
--- a/cts/scheduler/summary/ticket-promoted-3.summary
+++ b/cts/scheduler/summary/ticket-promoted-3.summary
@@ -1,31 +1,31 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
Transition Summary:
- * Stop rsc1:0 ( Promoted node1 ) due to node availability
- * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
+ * Stop rsc1:0 ( Promoted node1 ) due to node availability
+ * Stop rsc1:1 ( Unpromoted node2 ) due to node availability
Executing Cluster Transition:
* Pseudo action: ms1_demote_0
* Resource action: rsc1:1 demote on node1
* Pseudo action: ms1_demoted_0
* Pseudo action: ms1_stop_0
* Resource action: rsc1:1 stop on node1
* Resource action: rsc1:0 stop on node2
* Pseudo action: ms1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Stopped: [ node1 node2 ]
diff --git a/cts/scheduler/summary/ticket-promoted-9.summary b/cts/scheduler/summary/ticket-promoted-9.summary
index f116a2eea0..788573facb 100644
--- a/cts/scheduler/summary/ticket-promoted-9.summary
+++ b/cts/scheduler/summary/ticket-promoted-9.summary
@@ -1,36 +1,36 @@
Current cluster status:
* Node List:
* Online: [ node1 node2 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node1
* Clone Set: ms1 [rsc1] (promotable):
* Promoted: [ node1 ]
* Unpromoted: [ node2 ]
Transition Summary:
* Fence (reboot) node1 'deadman ticket was lost'
* Move rsc_stonith ( node1 -> node2 )
- * Stop rsc1:0 ( Promoted node1 ) due to node availability
+ * Stop rsc1:0 ( Promoted node1 ) due to node availability
Executing Cluster Transition:
* Pseudo action: rsc_stonith_stop_0
* Pseudo action: ms1_demote_0
* Fencing node1 (reboot)
* Resource action: rsc_stonith start on node2
* Pseudo action: rsc1:1_demote_0
* Pseudo action: ms1_demoted_0
* Pseudo action: ms1_stop_0
* Pseudo action: rsc1:1_stop_0
* Pseudo action: ms1_stopped_0
Revised Cluster Status:
* Node List:
* Online: [ node2 ]
* OFFLINE: [ node1 ]
* Full List of Resources:
* rsc_stonith (stonith:null): Started node2
* Clone Set: ms1 [rsc1] (promotable):
* Unpromoted: [ node2 ]
* Stopped: [ node1 ]
diff --git a/cts/scheduler/summary/whitebox-ms-ordering-move.summary b/cts/scheduler/summary/whitebox-ms-ordering-move.summary
index 6a5fb6eaeb..c9b13e032d 100644
--- a/cts/scheduler/summary/whitebox-ms-ordering-move.summary
+++ b/cts/scheduler/summary/whitebox-ms-ordering-move.summary
@@ -1,107 +1,107 @@
Current cluster status:
* Node List:
* Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* GuestOnline: [ lxc1@rhel7-1 lxc2@rhel7-1 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-3
* FencingPass (stonith:fence_dummy): Started rhel7-4
* FencingFail (stonith:fence_dummy): Started rhel7-5
* rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-1
* rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Started rhel7-2
* rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Started rhel7-3
* rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Started rhel7-4
* rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Started rhel7-5
* migrator (ocf:pacemaker:Dummy): Started rhel7-4
* Clone Set: Connectivity [ping-1]:
* Started: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* Stopped: [ lxc1 lxc2 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ rhel7-3 ]
* Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 rhel7-5 ]
* Resource Group: group-1:
* r192.168.122.207 (ocf:heartbeat:IPaddr2): Started rhel7-3
* petulant (service:DummySD): Started rhel7-3
* r192.168.122.208 (ocf:heartbeat:IPaddr2): Started rhel7-3
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started rhel7-3
* container1 (ocf:heartbeat:VirtualDomain): Started rhel7-1
* container2 (ocf:heartbeat:VirtualDomain): Started rhel7-1
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Promoted: [ lxc1 ]
* Unpromoted: [ lxc2 ]
Transition Summary:
* Move container1 ( rhel7-1 -> rhel7-2 )
- * Restart lxc-ms:0 ( Promoted lxc1 ) due to required container1 start
+ * Restart lxc-ms:0 ( Promoted lxc1 ) due to required container1 start
* Move lxc1 ( rhel7-1 -> rhel7-2 )
Executing Cluster Transition:
* Resource action: rsc_rhel7-1 monitor on lxc2
* Resource action: rsc_rhel7-2 monitor on lxc2
* Resource action: rsc_rhel7-3 monitor on lxc2
* Resource action: rsc_rhel7-4 monitor on lxc2
* Resource action: rsc_rhel7-5 monitor on lxc2
* Resource action: migrator monitor on lxc2
* Resource action: ping-1 monitor on lxc2
* Resource action: stateful-1 monitor on lxc2
* Resource action: r192.168.122.207 monitor on lxc2
* Resource action: petulant monitor on lxc2
* Resource action: r192.168.122.208 monitor on lxc2
* Resource action: lsb-dummy monitor on lxc2
* Pseudo action: lxc-ms-master_demote_0
* Resource action: lxc1 monitor on rhel7-5
* Resource action: lxc1 monitor on rhel7-4
* Resource action: lxc1 monitor on rhel7-3
* Resource action: lxc1 monitor on rhel7-2
* Resource action: lxc2 monitor on rhel7-5
* Resource action: lxc2 monitor on rhel7-4
* Resource action: lxc2 monitor on rhel7-3
* Resource action: lxc2 monitor on rhel7-2
* Resource action: lxc-ms demote on lxc1
* Pseudo action: lxc-ms-master_demoted_0
* Pseudo action: lxc-ms-master_stop_0
* Resource action: lxc-ms stop on lxc1
* Pseudo action: lxc-ms-master_stopped_0
* Pseudo action: lxc-ms-master_start_0
* Resource action: lxc1 stop on rhel7-1
* Resource action: container1 stop on rhel7-1
* Resource action: container1 start on rhel7-2
* Resource action: lxc1 start on rhel7-2
* Resource action: lxc-ms start on lxc1
* Pseudo action: lxc-ms-master_running_0
* Resource action: lxc1 monitor=30000 on rhel7-2
* Pseudo action: lxc-ms-master_promote_0
* Resource action: lxc-ms promote on lxc1
* Pseudo action: lxc-ms-master_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* GuestOnline: [ lxc1@rhel7-2 lxc2@rhel7-1 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started rhel7-3
* FencingPass (stonith:fence_dummy): Started rhel7-4
* FencingFail (stonith:fence_dummy): Started rhel7-5
* rsc_rhel7-1 (ocf:heartbeat:IPaddr2): Started rhel7-1
* rsc_rhel7-2 (ocf:heartbeat:IPaddr2): Started rhel7-2
* rsc_rhel7-3 (ocf:heartbeat:IPaddr2): Started rhel7-3
* rsc_rhel7-4 (ocf:heartbeat:IPaddr2): Started rhel7-4
* rsc_rhel7-5 (ocf:heartbeat:IPaddr2): Started rhel7-5
* migrator (ocf:pacemaker:Dummy): Started rhel7-4
* Clone Set: Connectivity [ping-1]:
* Started: [ rhel7-1 rhel7-2 rhel7-3 rhel7-4 rhel7-5 ]
* Stopped: [ lxc1 lxc2 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ rhel7-3 ]
* Unpromoted: [ rhel7-1 rhel7-2 rhel7-4 rhel7-5 ]
* Resource Group: group-1:
* r192.168.122.207 (ocf:heartbeat:IPaddr2): Started rhel7-3
* petulant (service:DummySD): Started rhel7-3
* r192.168.122.208 (ocf:heartbeat:IPaddr2): Started rhel7-3
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started rhel7-3
* container1 (ocf:heartbeat:VirtualDomain): Started rhel7-2
* container2 (ocf:heartbeat:VirtualDomain): Started rhel7-1
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Promoted: [ lxc1 ]
* Unpromoted: [ lxc2 ]
diff --git a/cts/scheduler/summary/whitebox-ms-ordering.summary b/cts/scheduler/summary/whitebox-ms-ordering.summary
index 921f6d068d..4d23221fa6 100644
--- a/cts/scheduler/summary/whitebox-ms-ordering.summary
+++ b/cts/scheduler/summary/whitebox-ms-ordering.summary
@@ -1,73 +1,73 @@
Current cluster status:
* Node List:
* Online: [ 18node1 18node2 18node3 ]
* Full List of Resources:
* shooter (stonith:fence_xvm): Started 18node2
* container1 (ocf:heartbeat:VirtualDomain): FAILED
* container2 (ocf:heartbeat:VirtualDomain): FAILED
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Stopped: [ 18node1 18node2 18node3 ]
Transition Summary:
* Fence (reboot) lxc2 (resource: container2) 'guest is unclean'
* Fence (reboot) lxc1 (resource: container1) 'guest is unclean'
* Start container1 ( 18node1 )
* Start container2 ( 18node1 )
- * Recover lxc-ms:0 ( Promoted lxc1 )
- * Recover lxc-ms:1 ( Unpromoted lxc2 )
+ * Recover lxc-ms:0 ( Promoted lxc1 )
+ * Recover lxc-ms:1 ( Unpromoted lxc2 )
* Start lxc1 ( 18node1 )
* Start lxc2 ( 18node1 )
Executing Cluster Transition:
* Resource action: container1 monitor on 18node3
* Resource action: container1 monitor on 18node2
* Resource action: container1 monitor on 18node1
* Resource action: container2 monitor on 18node3
* Resource action: container2 monitor on 18node2
* Resource action: container2 monitor on 18node1
* Resource action: lxc-ms monitor on 18node3
* Resource action: lxc-ms monitor on 18node2
* Resource action: lxc-ms monitor on 18node1
* Pseudo action: lxc-ms-master_demote_0
* Resource action: lxc1 monitor on 18node3
* Resource action: lxc1 monitor on 18node2
* Resource action: lxc1 monitor on 18node1
* Resource action: lxc2 monitor on 18node3
* Resource action: lxc2 monitor on 18node2
* Resource action: lxc2 monitor on 18node1
* Pseudo action: stonith-lxc2-reboot on lxc2
* Pseudo action: stonith-lxc1-reboot on lxc1
* Resource action: container1 start on 18node1
* Resource action: container2 start on 18node1
* Pseudo action: lxc-ms_demote_0
* Pseudo action: lxc-ms-master_demoted_0
* Pseudo action: lxc-ms-master_stop_0
* Resource action: lxc1 start on 18node1
* Resource action: lxc2 start on 18node1
* Pseudo action: lxc-ms_stop_0
* Pseudo action: lxc-ms_stop_0
* Pseudo action: lxc-ms-master_stopped_0
* Pseudo action: lxc-ms-master_start_0
* Resource action: lxc1 monitor=30000 on 18node1
* Resource action: lxc2 monitor=30000 on 18node1
* Resource action: lxc-ms start on lxc1
* Resource action: lxc-ms start on lxc2
* Pseudo action: lxc-ms-master_running_0
* Resource action: lxc-ms monitor=10000 on lxc2
* Pseudo action: lxc-ms-master_promote_0
* Resource action: lxc-ms promote on lxc1
* Pseudo action: lxc-ms-master_promoted_0
Revised Cluster Status:
* Node List:
* Online: [ 18node1 18node2 18node3 ]
* GuestOnline: [ lxc1@18node1 lxc2@18node1 ]
* Full List of Resources:
* shooter (stonith:fence_xvm): Started 18node2
* container1 (ocf:heartbeat:VirtualDomain): Started 18node1
* container2 (ocf:heartbeat:VirtualDomain): Started 18node1
* Clone Set: lxc-ms-master [lxc-ms] (promotable):
* Promoted: [ lxc1 ]
* Unpromoted: [ lxc2 ]
diff --git a/cts/scheduler/summary/whitebox-orphan-ms.summary b/cts/scheduler/summary/whitebox-orphan-ms.summary
index 0d0007dcc6..7e1b45b272 100644
--- a/cts/scheduler/summary/whitebox-orphan-ms.summary
+++ b/cts/scheduler/summary/whitebox-orphan-ms.summary
@@ -1,87 +1,87 @@
Current cluster status:
* Node List:
* Online: [ 18node1 18node2 18node3 ]
* GuestOnline: [ lxc1@18node1 lxc2@18node1 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started 18node2
* FencingPass (stonith:fence_dummy): Started 18node3
* FencingFail (stonith:fence_dummy): Started 18node3
* rsc_18node1 (ocf:heartbeat:IPaddr2): Started 18node1
* rsc_18node2 (ocf:heartbeat:IPaddr2): Started 18node2
* rsc_18node3 (ocf:heartbeat:IPaddr2): Started 18node3
* migrator (ocf:pacemaker:Dummy): Started 18node1
* Clone Set: Connectivity [ping-1]:
* Started: [ 18node1 18node2 18node3 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ 18node1 ]
* Unpromoted: [ 18node2 18node3 ]
* Resource Group: group-1:
* r192.168.122.87 (ocf:heartbeat:IPaddr2): Started 18node1
* r192.168.122.88 (ocf:heartbeat:IPaddr2): Started 18node1
* r192.168.122.89 (ocf:heartbeat:IPaddr2): Started 18node1
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started 18node1
* container2 (ocf:heartbeat:VirtualDomain): ORPHANED Started 18node1
* lxc1 (ocf:pacemaker:remote): ORPHANED Started 18node1
* lxc-ms (ocf:pacemaker:Stateful): ORPHANED Promoted [ lxc1 lxc2 ]
* lxc2 (ocf:pacemaker:remote): ORPHANED Started 18node1
* container1 (ocf:heartbeat:VirtualDomain): ORPHANED Started 18node1
Transition Summary:
* Move FencingFail ( 18node3 -> 18node1 )
* Stop container2 ( 18node1 ) due to node availability
* Stop lxc1 ( 18node1 ) due to node availability
- * Stop lxc-ms ( Promoted lxc1 ) due to node availability
- * Stop lxc-ms ( Promoted lxc2 ) due to node availability
+ * Stop lxc-ms ( Promoted lxc1 ) due to node availability
+ * Stop lxc-ms ( Promoted lxc2 ) due to node availability
* Stop lxc2 ( 18node1 ) due to node availability
* Stop container1 ( 18node1 ) due to node availability
Executing Cluster Transition:
* Resource action: FencingFail stop on 18node3
* Resource action: lxc-ms demote on lxc2
* Resource action: lxc-ms demote on lxc1
* Resource action: FencingFail start on 18node1
* Resource action: lxc-ms stop on lxc2
* Resource action: lxc-ms stop on lxc1
* Resource action: lxc-ms delete on 18node3
* Resource action: lxc-ms delete on 18node2
* Resource action: lxc-ms delete on 18node1
* Resource action: lxc2 stop on 18node1
* Resource action: lxc2 delete on 18node3
* Resource action: lxc2 delete on 18node2
* Resource action: lxc2 delete on 18node1
* Resource action: container2 stop on 18node1
* Resource action: container2 delete on 18node3
* Resource action: container2 delete on 18node2
* Resource action: container2 delete on 18node1
* Resource action: lxc1 stop on 18node1
* Resource action: lxc1 delete on 18node3
* Resource action: lxc1 delete on 18node2
* Resource action: lxc1 delete on 18node1
* Resource action: container1 stop on 18node1
* Resource action: container1 delete on 18node3
* Resource action: container1 delete on 18node2
* Resource action: container1 delete on 18node1
Revised Cluster Status:
* Node List:
* Online: [ 18node1 18node2 18node3 ]
* Full List of Resources:
* Fencing (stonith:fence_xvm): Started 18node2
* FencingPass (stonith:fence_dummy): Started 18node3
* FencingFail (stonith:fence_dummy): Started 18node1
* rsc_18node1 (ocf:heartbeat:IPaddr2): Started 18node1
* rsc_18node2 (ocf:heartbeat:IPaddr2): Started 18node2
* rsc_18node3 (ocf:heartbeat:IPaddr2): Started 18node3
* migrator (ocf:pacemaker:Dummy): Started 18node1
* Clone Set: Connectivity [ping-1]:
* Started: [ 18node1 18node2 18node3 ]
* Clone Set: master-1 [stateful-1] (promotable):
* Promoted: [ 18node1 ]
* Unpromoted: [ 18node2 18node3 ]
* Resource Group: group-1:
* r192.168.122.87 (ocf:heartbeat:IPaddr2): Started 18node1
* r192.168.122.88 (ocf:heartbeat:IPaddr2): Started 18node1
* r192.168.122.89 (ocf:heartbeat:IPaddr2): Started 18node1
* lsb-dummy (lsb:/usr/share/pacemaker/tests/cts/LSBDummy): Started 18node1
File Metadata
Details
Attached
Mime Type
text/x-diff
Expires
Wed, Jun 25, 6:25 AM (12 h, 27 m)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
1952419
Default Alt Text
(200 KB)
Attached To
Mode
rP Pacemaker
Attached
Detach File
Event Timeline
Log In to Comment